Dinesh Nirmal, IBM | IBM Think 2020
>> Announcer: From theCUBE Studios in Palo Alto and Boston, it's theCUBE, covering IBM Think, brought to you by IBM. >> Welcome back, I'm Stu Miniman, and this is theCUBE's coverage of IBM Think 2020, the digital experience. Welcome to the program, Dinesh Nirmal, who's the chief product officer for Cloud Paks inside IBM. Dinesh, nice to see you, thanks so much for joining us. >> Thank you Stu, really appreciate you taking the time. >> All right, so, I've been to many IBM shows, and of course, I'm an analyst in the cloud space, so I'm familiar with IBM Cloud Paks, but maybe just refresh our audience minds here, what they are, how long have they been around for, what clouds do they live on, and maybe what's new in 2020 that if somebody had looked at this in the past that they might not know about IBM Cloud Pak? >> Yeah, so thanks Stu. So to start with, let me say that Cloud Paks are cloud agnostic. So, the whole goal is that you build once and it can run anywhere. That is the basic mantra, or principle, that we want to build Cloud Paks with. So they are, look at them as a set of micro services containerized in a form that it can run on any public cloud or behind a firewall. So that's the whole premise of Cloud Paks. So, when you go back to Cloud Paks, it's an integrated set of services that solve a specific set of business problems and also accelerates building each set of applications and solutions. That's what Cloud Paks brings. So, especially in this environment Stu, think about it. If I'm an enterprise, my goal is how can I accelerate and how can I automate? Those are the two key things that comes to my mind if I am a C-level exec at an enterprise. So, Cloud Paks enables that, meaning you already have a set of stitched together services that accelerates the application development. It automates a lot of things for you. So you today have a lot of applications running on multiple clouds or behind the firewall. How do you manage those, right? Cloud Paks will help. So, let me give you one example since you asked specifically on Cloud Paks. Let's take Cloud Pak for Data. The set of services that is available in Cloud Pak for Data will make it easier for all the way from ingest to visualization. There's a set of services that you can use, so you don't have to go build a service or a product or user product for ingest, then use another product for ETL, use another product for building models, another product to manage those models. The Cloud Pak for Data will solve all the problems end to end. It's a rich set of services that will give you all the value that you need all the way from ingest to visualization. And with any personas, whether you are a data engineer, data scientist, or you are a business analyst, you all can collaborate through the Cloud Paks. So that's the two minute answer to your question what Cloud Paks is. >> Awesome, thanks Dinesh. Yeah, I guess you pointed out something right at the beginning there. I hear IBM Cloud Pak and I think IBM Cloud. But you said specifically this is really cloud agnostic. So this week is Think, last week I was covering Red Hat Summit, so I heard a lot about multicloud deployments, talked to the rail team, talked to the open chip team. So, help me understand where do Cloud Pak fit when we're talking about these multicloud employments? And is there some connection with the partnership that, of course, IBM has with Red Hat? >> Of course, so all Cloud Paks are optimized for OpenShift, meaning how do we use the set of services that OpenShift gives, the container management that OpenShift provides? So as we build containers or micro services, how do we make sure that we are optimizing or taking advantage of OpenShift? So, for example, the set of services like logging, monitoring, security, all those services metering that comes from OpenShift is what we are using as Cloud Pak. So Cloud Paks are optimized for OpenShift. From an automation perspective, how do we use Ansible, right? So, all the value that Red Hat and OpenShift brings is what Cloud Pak is built on. So if you look at as a layer as a Lego, the base Lego is OpenShift and rail. And then on top of it sit Cloud Paks, and applications and solutions on top of it. So it's, if I look at layer base, the base Lego layer is OpenShift and Red Hat rail. >> Well, great, that's super important because, one of the things we've been looking at for a while is, you talk about hybrid cloud, you talk about multicloud, and often it's not that platform, that infrastructure discussion, but the biggest challenge for companies today is how do I build new applications, how do I modernize what I have? So, sounds like this is exactly where you're targeting to help people through that transformation that they're going through. >> Yeah, exactly Stu, because if you look at it, in the past products were siloed. You build a product, you use a set of specs to build it. It was siloed. And customers becomes the software integrators, or system integrators, where they have to take the different products, put it together. So even if I am focused on the data space, or AI space, before I had to bring in three or four or five different products, make it all work together to build a model, deploy the model, manage the model, the lifecycle of the model, the lifecycle of the data. But the Cloud Paks bring it all in one box, where out of the box you are ready to go. So your time to value is much more higher with Cloud Paks because you already get a set of stitched together services that gets working right out of the box. >> So, I love the idea of out of the box. When I think of cloud native, modern application development, simplicity is not the first thing I think of, Dinesh. So, help me understand. So many customers, it's the tools, the skillsets, they don't necessarily have the experience. How is what your product set and your team's doing, help customers that deal with the ever-changing landscape and the complexity that they are faced with? >> Yeah, so the honest truth, Stu, is that enterprise applications are not an app that you create and put it on iPhone, right? I mean, it is much more complex, because it's dealing with hundreds of millions of people trying to transact with the system. You need to make sure there is a disaster recovery backup, scalability, elasticity, all those things, security, obviously, very critical piece, and multitenancy. All those things has to come together in an enterprise application. So, when people talk about simplicity, it comes at a price. So, what Cloud Paks has done, is that we have really focused on the user experience and design piece. So, you as an end-user has a great experience using the integrated set of services. The complexity piece will still be there, to some extent, because you're building a very complex multitenant enterprise application, but how do we make it easier for a developer or a data scientist to collaborate or reuse the assets, find the data much more easier, or trust the data much more easier than before? Use AI to predict a lot of the things, including bias detection, all those things. So, we are making a lot of the development, automation and acceleration easier. The complexity part will be there still, because enterprise applications tend to be complex by nature. But we are making it much more easier for you to develop, deploy, manage and govern what you are building. >> Yeah, so, how does Cloud Paks allow you to really work with the customers, focus on things like innovation, showing them the latest in the IBM software portfolio? >> Yeah, so the first piece is that we made it much more easier for the different personas to collaborate. So in the past, what is the biggest challenge, me as a data scientist had? Me as a data scientist, the biggest challenge was that getting access to the data, trusted data. Now we have put some governance around it, where by which you can get data, trusted data, much more easier using Cloud Pak for Data. Governance around the data, meaning if you have a CDO, you want to see who is using the data, how clean is the data, right? A lot of times he data might not be clean, so we want to make sure we can help with that. Now, let me move into the the line of business piece, not just the data. If I am an LOB, and I want to use, automate a lot of the process I have in today, in my enterprise, and not go through the every process automation, and go through your superior or supervisor to get approval, how do we use AI in the business process automation also? So those kind of things, you will get through Cloud Paks. Now, the other piece of Cloud Pak, if I am an IT space, right? The day-two operations, scalability, security, delivery of the software, backup and restore, how do we automate and help with that, the storage layer? Those are day-two operations. So, we are taking it all the way from day one, meaning the whole experience of setting it up, to day two, where enterprise is really worried about, making it seamless and easy using Cloud Paks, I go back to what I said in the beginning, which is out of the accelerate and automate, a lot of the work that enterprise have to do today, much more easier. >> Okay, we talked earlier in the discussion about that this can be used across multiple clouded environments. My understanding, you mentioned one of the IBM Cloud Paks, one for data. There's a number of different Cloud Paks out there. How does that work from a customer's standpoint? Do I have to choose a Cloud Pak or a specific cloud? Is it a license that goes across all of my environments? Help me understand how this deployment mechanism and its support and maintenance works. >> Right, so we have the base, obviously. I said look at it as a modular Lego model. The base is obviously open chipped and rail. On top of its cells sits a bedrock, we call, which is a common set of services and the logic to expand. On top of it sits Cloud Pak for Data, Cloud Pak for Security, Cloud Pak for Applications, there's Cloud Pak for Multicloud Management, there's Cloud Pak for Integration. So there is total of six Cloud Paks that's available, but you can pick and choose which Cloud Pak you want. So let's say you are a CDO, or you are an enterprise who want to focus on data and AI, you can just pick Cloud Pak for Data. Or let's say you are a Cloud Pak based on processes, BPM decision rules, you can go without platform automation, which gives you the set of tools. But the biggest benefits too, is that all these Cloud Paks are a set of integrated services that can all work together, sits optimized on top of open chipped. So, all of a sudden, you'll need Cloud Pak for Data, and now you want to do data, but now you want to expand it into your line of business, and you want Cloud Pak for Automation, you can bring that in. Now those two Cloud Paks works together well. Now you want to bring in Cloud Pak for Multicloud Management, because you have data, or applications running on multiple clouds, so now you can bring Cloud Pak for MCM, which is multicloud management, and those three work together. So it's all a set of integrated set of services that is optimized on top of OpenShift, which makes it much more easier for customers to bring the rich set of services together and accelerate and automate their lifecycle journey within the enterprise. >> Great, last question for you Dinesh. What new in 2020, what should customers be looking at today? Would love if you can give a little bit of guidance as to where customers should be looking at for things that might be coming a little bit down the line here, and if they want to learn more about IBM Cloud Paks, where should they be looking? >> Yeah, they want to learn more, there's www.ibm.com/cloudpaks. There's a place to go. There, all the details around Cloud Paks are there. You can also get in touch with me, and I can definitely take you to more detail. But what is coming is that, look, so we have a set of Cloud Paks, but we want to expand and make it extensible. So how do we, already it's built on an open platform, but how do we make sure our partners and ISPs can come and build on top of the base-cloud part? So that's the focus going to be, as each Cloud Pak innovate and add more value within those Cloud Paks. We also want to expand it so that our partners and our ISPs and GSIs can build on top of it. So this year, the focus is continuously innovate across the Cloud Paks, but also make it much more extensible for third parties to come and build more value on top of the Cloud Pak itself. That's one area we are focusing on. The other area's MCM, right? Multicloud management, because there is tremendous appetite for customers to move data or applications on cloud, and not only on one cloud, hybrid cloud. So how do you manage that, right? So multicloud management definitely helps on that perspective. So our focus this year is going to be one, make it extensible, make it more open, but at the same time continuously innovate on every single Cloud Pak to make that journey for customers on automating and accelerating application development easier. >> All right, well Dinesh, thank you so much. Yeah, the things that you talked about, that absolutely top of mind for customers that we talked to. Multicloud management, as you said, it was the ACM, the Advanced Cluster Management, that we heard about from the Red Hat team last week at Summit. So thank you so much for the updates. Definitely exciting to watch Cloud Pak, how you're helping customers deal with that huge, it's the opportunity but also the challenge of building their next applications, modernizing what they're doing without, still having to think about what they have from (faintly speaking), so thanks so much, great to talk with you. >> Well, thanks Stu, great talking. >> All right, lots more coverage from IBM Think 2020, the digital experience. I'm Stu Miniman, and as always, thank you for watching theCUBE. (upbeat music)
SUMMARY :
Think, brought to you by IBM. the digital experience. appreciate you taking the time. So, the whole goal is that you build once right at the beginning there. So, for example, the set but the biggest challenge the lifecycle of the model, and the complexity that lot of the development, for the different personas to collaborate. one of the IBM Cloud Paks, services and the logic to expand. a little bit down the line here, So that's the focus going to be, Yeah, the things that you talked about, the digital experience.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dinesh Nirmal | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Dinesh | PERSON | 0.99+ |
last week | DATE | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
four | QUANTITY | 0.99+ |
Boston | LOCATION | 0.99+ |
2020 | DATE | 0.99+ |
three | QUANTITY | 0.99+ |
iPhone | COMMERCIAL_ITEM | 0.99+ |
first piece | QUANTITY | 0.99+ |
Stu | PERSON | 0.99+ |
two minute | QUANTITY | 0.99+ |
one box | QUANTITY | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
Lego | ORGANIZATION | 0.99+ |
today | DATE | 0.99+ |
Cloud Pak for Data | TITLE | 0.98+ |
this year | DATE | 0.98+ |
two | QUANTITY | 0.98+ |
one example | QUANTITY | 0.98+ |
each set | QUANTITY | 0.98+ |
Cloud Paks | TITLE | 0.98+ |
Cloud Pak for Data | TITLE | 0.98+ |
six | QUANTITY | 0.98+ |
Red Hat Summit | EVENT | 0.98+ |
theCUBE | ORGANIZATION | 0.98+ |
www.ibm.com/cloudpaks | OTHER | 0.98+ |
this week | DATE | 0.98+ |
two key | QUANTITY | 0.97+ |
day two | QUANTITY | 0.97+ |
OpenShift | TITLE | 0.96+ |
Cloud Pak | TITLE | 0.96+ |
one | QUANTITY | 0.95+ |
Cloud Pak | TITLE | 0.95+ |
each | QUANTITY | 0.95+ |
hundreds of millions of people | QUANTITY | 0.93+ |
Cloud Paks | TITLE | 0.92+ |
theCUBE Studios | ORGANIZATION | 0.91+ |
day one | QUANTITY | 0.91+ |
Think 2020 | COMMERCIAL_ITEM | 0.9+ |
Cloud Pak for | TITLE | 0.89+ |
Cloud Paks | ORGANIZATION | 0.89+ |
Multicloud Management | TITLE | 0.88+ |
Paks | TITLE | 0.87+ |
Dinesh Nirmal, IBM | IBM Think 2020
>>Yeah, >>from the Cube Studios in Palo Alto and Boston. It's the Cube covering IBM think wrought by IBM. >>Welcome back. I'm Stew Middleman. And this is the Cube's coverage of IBM. Think 2020. The digital experience. Happy to welcome to the program. Dinesh Nirmal, Who's the chief product officer for cloud packs inside IBM Deneche. Nice to see you. Thanks so much for joining us. >>Thank you. Still really appreciate you taking the time. >>All right, So Ah, I've been to many IBM shows, and of course, I'm an analyst in the cloud space. So I'm familiar with IBM cloud packs. But maybe, you know, just refresh our audience minds here what they are. How long have they been around for? You know, what clouds do they live on? And maybe what's what's new in 2020? That if somebody had looked at this, you know, in the past that they might not know about IBM Cloud. >>Yeah, so thanks to start with, let me say that. Well, tax at 12. Agnostic. So the whole goal is that you build once and it can run anywhere. That is the basic mantra or principle that we want to build packs with, So they're looking at them as a set of micro services containerized in the form that it can run on any public or behind the firewall. So that's the whole premise of about Pass. So when you go back to cloud packs, it's an integrator set of services that solve a specific set of business problems and also accelerates building rich set of applications and solutions. That's what cloud back springs. So you know, especially in this and moments to think about it, you know, violent underprice. My goal is how can I accelerate and how can I automate? Those are the two key things you know that comes to my mind if I am a C level execs at an enterprise. So cloud practice enables that meaning you already have a set off stitched together services that accelerate the application development. It automates a lot of things for you. So you do. They have a lot of applications running on multiple clouds or behind the firewall. How do you manage those right about banks will help, so I'll let me give you one example. Since you are specifically on 12 packs, let's stay cloud back for data the set of services that is available in cloud back for data will make it easier for all the way from ingest to visualization. There's a set of services that you can use so you don't have to go build a service or a product or use a product for ingest. Then use another product for CTL. Use another product for building models, another product to manage those models. The cloud back for data will solve all the problems and to end. It's a rich set of services that will give you all the value that you need all the way from ingest to visualization and with any personas We know whether you are a data engineer, data scientist or you are, you know, business analyst. You all can cooperate through the part. So that's the you know, two minute answer your question. What about practice? >>Awesome. Thanks in. Actually, I I guess you pointed out something right at the beginning. There I hear IBM Cloud pack and I think IBM cloud. But you said specifically this is really cloud agnostic. So you know, this week is think Last week I was covering Red Hat Summit, so I heard a lot about multi cloud deployments. You know, talk to the well team, talk to the open shift team. Um, so help me understand. You know, where do cod packed bit when we're talking about, you know, these multi cloud employments, you know? And is there some connection with the partnership that, of course, IBM has with red hat >>off course. I mean, so all cloud packs are optimized for open shipped. Meaning, you know, how do we use the set of services that open ship gives that container management that open provides. So as we build containers or micro services, how do we make sure that we are optimizing or taking advantage of open ship? So, for example, the set of services like logging, monitoring, security, all those services meeting that comes from open shift is what we are using. Basketball packs of cloud packs are optimized for open shift. Um, you know, from an automation perspective, how do we use and simple Right? So all the value that red hat an open ship brings is what about back is built on. So if you look at it as a layer as a Lego, the based Lego is open shift and rail. And then on top of it sits cloud pass and applications and solutions on top of it. So if I look at layer bass, the bass Lego layer is open shift and red pepper. >>Well, great. That's that's super important because, you know, one of the things we've been looking at for a while is you talk about hybrid cloud, You talk about multi cloud, and often it's that platform that infrastructure discussion. But the biggest challenge for companies today is how do I build new applications? How do I modernize what I have? So >>it >>sounds like this is exactly, you know where you're targeting to help people, you know, through the through that transformation that they're going through. >>Yeah, exactly. Stew. Because if you look at it, you know, in the past, products for siloed I mean, you know, you build a product, you use a set of specs to build it. It was a silo, and customers becomes the software integrators, or system integrators, where they have to take the different products, put it together. So even if I am, you know, focused on the data space or AI space before I have to bring in three or four or five different products make it all work together to build a model, deploy the model, manage the model, the lifecycle of the model, the life cycle of the data. But the cloud packs bring it all in one box, were out of the box. You're ready to go, so your time to value is much more higher with cloud packs because you already get a several stitched together services that gets working right out of the box. >>So I love the idea of out of the box when I think of Cloud native modern modern application development. Simplicity is not the first thing I think of Danish. So help me understand. You know so many customers. It's, you know, the tools, the skill set. You know, they don't necessarily have the experience. How is what you know your products and your team's doing help customers deal with, You know, the ever changing landscape and the complexity that they're faced with. >>Yeah, so the that honest roots, too, is that enterprise applications are not an app that you create and put it on iPhone, right? I mean, it is much more complex because it's dealing with you know, hundreds of millions of people trying to transact with the system. You need to make sure there is a disaster recovery, backup scalability, the elasticity. I mean, all those things. Security, I mean, obviously very critical piece and multi tenancy. All those things has to come together in an enterprise application. So when people talk about, you know, simplicity, it comes at a price. So what cloud practice is done is that you know, we have really focused on the user experience and design piece. So you, as an end user, has a great Syrians using the integrated set of services. The complexity piece will still be there to some extent, because you're building a very complex, you know, multi tenant application, the price application. But how do we make it easier for a developer or a data scientist to collaborate or reuse the assets, find the data much more easier or trusted data much more easier than before? Use AI, you know, to predict a lot of the things including, you know, bias detection, all those things. So we're making a lot off the development, automation and acceleration easier. The complexity part will be there still, because You know, enterprise applications tend to be complex by nature, but we're making it much more easier for you to develop, deploy and manage and govern what you're building. >>Yeah, so? So how does cloud packs allow you to really, you know, work with customers focus on, you know, things like innovation showing them the latest in the IBM software portfolio. >>Yeah. So off the first pieces that we made it much more easier for the different personas to collaborate. So in the past, you know what is the biggest challenge? Me as a data scientist had me as a data scientist. The biggest challenge was that getting access to the data Trusted data. Now, you know, we have put some governance around it, or by average, you can get, you know, data trusted data much more easier using our back to data governments around the data. Meaning if you have a CDO, you want to see who is using the data? How clean is that data, right? I mean, a lot of times that data might not be clean, so we want to make sure we can. You can help with that. Now let me move into the you know the line of business. He's not just the data. If I am, you know, a l l o B. And I want to use order. Made a lot of the process I have in today in my enterprise and not go to the ah, the every process automation and go through your superior or supervisor to get approval. How do we use AI in the business process? Automation also. So those kind of things you will get to cloud parts now, the other piece of it. But if I'm an I t space, right, the day two operations scalability, security, delivery of the software, backup and restore how do we automate and help it at the storage layer? Those air day two operations. So we're taking it all the way from day one. The whole experience of setting it up today to where enterprises really wherever, making it seamless and easy. Using quote Thanks. I go back to what I said in the beginning, which is how do we accelerate and automate a lot of the work that enterprises today much more easier. >>Okay, Wei talked earlier in the discussion about that. This can be used across multiple cloud environments. My understanding you mentioned one of the IBM cloud packs, one for data. There's a number of different cloud tax out there. How does that >>work from >>a customer standpoint? Do I have to choose a cloud pack for a specific cloud? Does it is a license that goes across all of my environment, Help me understand on this deployment mechanism and support inmates works >>right? So we have the base. Obviously, I said, You know, look at us. A modular Lego model. The base is obviously open shift in Rome. On top of itself sits a bedroom because which is a common set of services on the logic to experience. On top of it sits lower back for data well back for security cloud. For applications, there's cloud back for multi cloud management. There's cloud platform integration, so there's total of six power packs that's available, but you can pick and choose which about back you want. So let's say you are. You're a CD or you are an enterprise. We want to focus on data and ai. You can just speak lower back for data or let's say you are a, you know, based on processes, BPM decision rules. You can go with our back for automation, which gives you the set of tools. But the biggest benefits do is that all these quarterbacks are set up in the greatest services that can all work together since optimized on top of open ship. So all of a sudden you lead bar Cloud back for data, and now you want to do data. But now you want to expand it in New York and line of business. And you want that for automation? You can bring that in. Now those two quarterbacks works together. Well, now you want to bring it back for multi cloud management because you have data or applications running on multiple clouds. So now you can will bring it back for EMC M, which is multi cloud management. And those three work together. So it's fall as set off integrated set of services that is optimized on top of open shift, which makes it much more easier for customers to bring the rich set of services together and accelerate and automate their lifecycle journey within the enterprise. >>Great last question for you, Dan Ashe. You know what? What new in 2020. What should customers be looking at today would love if you could give a little bit of guidance as to where customers should be looking at for things that might be coming a little bit down the line here. And if they want to learn more about IBM cloud backs, where should they be looking? >>Yeah, if they want to learn more, there's, you know, the VW IBM dot com slash power packs. That's a place to go there. All the details on power packs are there. You can also get in touch with me, and I can definitely much more detail. But what is coming is that look. So we have a set of cloud parts, but we want to expand on, Make it extensible. So how do we you know, already it's built on an open platform. But how do we make sure our partners and I s feeds can come and build on top of the space cloud? So that's the focus going to be as each quote back innovate and add more value in within those cloud grants. We also wanted bandit so that you know our partners and our eyes, fees and GS size and build on top of it. So this year the focus is continuously in a way across the cloud part, but also make it much more extensible for third parties to come and build more value. That's the you know, That's one area of focus in the various EMC em right multi cloud management, because there is tremendous appetite for customers to move data or applications on cloud, and not only on one cloud hybrid cloud. So how do you manage that? Right. So multi cloud management definitely helps from that perspective. So our focus this year is going to be one. Make it extensible, make it more open, but at the same time continuously innovate on every single cloud part to make that journey for customers on automation and accelerating off application element easier. >>All right, we'll do next. Thank you so much. Yeah, the things that you talked about, that absolutely, you know, top of mind for customers that we talk to multi cloud management. As you said, it was the ACM, the advanced cluster management that we heard about from the Red Hat team last week at Summit. So thank you so much for the updates. Definitely exciting to watch cloud packed. How you're helping customers, you know, deal with that. That huge. It's the opportunity. But also the challenge of building their next applications, modernizing what they're doing without, you know, still having to think about what they have from their existing. Thanks so much. Great to talk. >>Thanks to you. >>All right, lots more coverage from IBM. Think 2020 The digital experience. I'm stew minimum. And as always, Thank you for watching the Cube. >>Yeah, Yeah, yeah, yeah, yeah.
SUMMARY :
It's the Cube Dinesh Nirmal, Who's the chief product officer for Still really appreciate you taking the time. That if somebody had looked at this, you know, in the past that they might not know about So that's the you So you know, this week is think Last week I was covering Red Hat Summit, So if you look at it as a layer as a Lego, the based Lego is open shift and That's that's super important because, you know, one of the things we've been looking at you know, through the through that transformation that they're going through. So even if I am, you know, focused on the data space or AI space before It's, you know, the tools, the skill set. So what cloud practice is done is that you know, we have really focused on the user experience So how does cloud packs allow you to really, you know, So in the past, you know what My understanding you mentioned one of the IBM cloud packs, So all of a sudden you lead bar Cloud back for data, What should customers be looking at today would love if you could give a little bit of guidance as That's the you know, That's one area of focus in the various EMC em right multi cloud Yeah, the things that you talked about, that absolutely, you know, And as always, Thank you for watching the Cube. Yeah, Yeah, yeah, yeah,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dan Ashe | PERSON | 0.99+ |
Dinesh Nirmal | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
New York | LOCATION | 0.99+ |
Stew Middleman | PERSON | 0.99+ |
four | QUANTITY | 0.99+ |
2020 | DATE | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
12 packs | QUANTITY | 0.99+ |
five | QUANTITY | 0.99+ |
Boston | LOCATION | 0.99+ |
iPhone | COMMERCIAL_ITEM | 0.99+ |
Last week | DATE | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
Rome | LOCATION | 0.99+ |
last week | DATE | 0.99+ |
three | QUANTITY | 0.99+ |
Lego | ORGANIZATION | 0.99+ |
VW | ORGANIZATION | 0.99+ |
two minute | QUANTITY | 0.99+ |
first pieces | QUANTITY | 0.99+ |
this week | DATE | 0.99+ |
today | DATE | 0.98+ |
one example | QUANTITY | 0.98+ |
this year | DATE | 0.98+ |
Red Hat Summit | EVENT | 0.98+ |
two | QUANTITY | 0.98+ |
six power packs | QUANTITY | 0.98+ |
one | QUANTITY | 0.97+ |
two key | QUANTITY | 0.97+ |
each quote | QUANTITY | 0.97+ |
one box | QUANTITY | 0.97+ |
Cube Studios | ORGANIZATION | 0.96+ |
Wei | PERSON | 0.95+ |
Syrians | PERSON | 0.95+ |
first thing | QUANTITY | 0.94+ |
this year | DATE | 0.94+ |
EMC M | ORGANIZATION | 0.92+ |
hundreds of millions of people | QUANTITY | 0.89+ |
day one | QUANTITY | 0.89+ |
Stew | PERSON | 0.88+ |
Think 2020 | COMMERCIAL_ITEM | 0.85+ |
day two | QUANTITY | 0.79+ |
Cube | COMMERCIAL_ITEM | 0.77+ |
IBM Deneche | ORGANIZATION | 0.77+ |
Agnostic | PERSON | 0.71+ |
Cube | ORGANIZATION | 0.67+ |
single cloud part | QUANTITY | 0.63+ |
Danish | LOCATION | 0.6+ |
ACM | ORGANIZATION | 0.6+ |
IBM dot com | ORGANIZATION | 0.57+ |
bandit | ORGANIZATION | 0.52+ |
EMC | ORGANIZATION | 0.5+ |
12 | QUANTITY | 0.5+ |
Cloud | TITLE | 0.34+ |
Cloud | COMMERCIAL_ITEM | 0.29+ |
Dinesh Nirmal, IBM | IBM Think 2018
>> Voiceover: Live from Las Vegas it's the Cube. Covering IBM Think 2018. Brought to you by IBM. >> Welcome back to IBM Think 2018. This is the Cube, the leader in live tech coverage. My name is Dave Vellante and this is our third day of wall-to-wall coverage of IBM Think. Dinesh Nirmal is here, he's the Vice-President of Analytics Development at IBM. Dinesh, great to see you again. >> I know. >> We just say each other a couple of weeks ago. >> I know, in New York. >> Yeah and, of course, in Big Data SV >> Right. >> Over at the Strata Conference. So, great to see you again. >> Well, Thank you. >> A little different venue here. We had real intimate in New York City and in San Jose. >> I know, I know. >> Massive. What are your thoughts on bringing all the clients together like this? >> I mean, it's great because we have combined all the conferences into one, which obviously helps because the message is very clear to our clients on what we are doing end-to-end, but the feedback has been tremendous. I mean, you know, very positive. >> What has the feedback been like in terms of how you guys are making progress in the analytics group? What are they like? What are they asking you for more of? >> Right. So on the analytics side, the data is growing you know, by terabytes a day and the questions is how do they create insights into this massive amount of data that they have in their premise or on Cloud. So we have been working to make sure that how can we build the tools to enable our customers to create insights whether the data is on private cloud, public, or hybrid. And that's a very unique valid proposition that we bring to our customers. Regardless of where your data is, we can help you whether it's cloud, private, or hybrid. >> Well so, we're living in this multi-petabyte world now. Like overnight it became multi-petabyte. And one of the challenges of course people have is not only how do you deal with that volume of data, but how do I act on it and get insights quickly. How do I operationalize it? So maybe you can talk about some of the challenges of operationalizing data. >> Right. So, when I look at machine learning, there is three D's I always say and you know, the first D is the data, the development of the model, and the deployment of the model. When I talk about operationalization, especially the deployment piece, is the one that gets the most challenging for our enterprise customers. Once you clean the data and you build the model how do you take that model and you bring it your existing infrastructure. I mean, you know, look at your large enterprises. Right? I mean, you know, they've been around for decades. So they have third party software. They have existing infrastructure. They have legacy systems. >> Dave: A zillion data marks and data warehouses >> Data marks, so into all of that, how do you infuse machine learning, becomes very challenging. I met with the CTO of a major bank a few months ago, and his statement kind of stands out to me. Where he said, "Dinesh, it only took us three weeks to build the model. It's been 11 months, we still haven't deployed it". So that's the challenge our customers face and that's where we bring in the skillset. Not just the tools but we bring the skills to enable and bring that into production. >> So is that the challenge? It's the skillsets or is it the organizational inertia around well I don't have the time to do that now because I've got to get this report out or ... >> Dinesh: Right Maybe you can talk about that a little. Right. So that is always there. Right? I mean, because once a priority is set obviously the different challenges pull you in different directions, so every organization faces that to a large extent. But I think if you take from a pure technical perspective, I would say the challenge is two things. Getting the right tools, getting the right skills. So, with IBM, what we are focusing is how do we bring the right tools, regardless of the form factor you have, whether Cloud, Private Cloud, Hybrid Cloud, and then how do we bring the right skills into it. So this week we announce the data science lead team, who can come in and help you with building models. Looking at the use cases. Should we be using vanilla machine learning or should we be using deep learning. All those things and how do we bring that model into the production environment itself. So I would say tools and skills. >> So skills wise, in the skills there's at least two paths. It's like the multi-tool athlete. You've got the understanding of the tech. >> Dinesh: Right. >> You know, the tools, most technology people say hey, I'll figure that out. But then there's this data and digital >> Right. >> Skills. It's like this double deep skills that is challenging. So you're saying you can help. >> Right. Sort of kick-start that and how does that work? That sort of a services engagement? That's part of the ... >> So, once you identify a use case, the data science lead team can come in, because they have the some level of vertical knowledge of your industry. They are very trained data scientists. So they can come assess the use case. Help you pick the algorithms to build it. And then help you deploy, cleanse the data. I mean, you bring up a very, very good point. I mean, let's just look at the data, right. The personas that's involved in data; there is the data engineer, there's the data scientist, there's the data worker, there's the data steward, there's the CTO. So, that's just the data piece. Right? I mean, there's so many personas that have to come together. And that's why I said the skills a very critical piece of all it, but also, working together. The collaboration is important. >> Alright, tell us more about IBM Cloud Private for data. We've heard about IBM Cloud Private. >> Danish: Right. >> Cloud Private for Data is new. What's that all about? >> Right, so we announced IBM Cloud Private for Data this week and let me tell you, Dave, this has been the most significant announcement from an analytic perspective, probably in a while, that we are getting such a positive response. And, I will tell you why. So when you look at the platform, our customers want three things. One, they want to be able to build on top of the platform. They want it to be open and they want it to be extensible. And we have all three available. The platform is built on Kubernetes. So it's completely open, it's scalable, it's elastic. All those features comes with it. And then we put that end-to-end so you can ingest the data, you can cleanse it or transform it. You can build models or do deep analytics on it. You can visualize it. So you can do everything on the platform. So I'll take an example, like block chain, for example, I mean you have, if I were to simplify it, Right? You have the ledger, where you are, obviously, putting your transactions in and then you have a stay database where you are putting your latest transactions in. The ledger's unstructured. So, how do you, as that is getting filled, How do you ingest that, transform it on the fly, and be able to write into a persistent place and do analytics on it. Only a platform can do with that kind of volume of data. And that's where the data platform brings in, which is very unique especially on the modern applications that you want to do. >> Yes, because if you don't have the platform ... Let's unpack this a little bit. You've got a series of bespoke products and then you've got, just a lot of latency in terms of the elapsed times to get to the insights. >> Dinesh: Right. >> Along the way you've got data consistency issues, data quality >> Dinesh: Right >> maybe is variable. Things change. >> Right. I mean, think about it, right. If you don't have the platform then you have side-load products. So all of a sudden you've got to get a product for your governance, your integration catalog. You need to get a product for ingest. You got to get a product for persistence. You got to get a product for analytics. You got to get a product for visualization. And then you add the complexity of the different personas working together between the multitude of products. You have a mess in your hand at that point. The platform solves that problem because it brings you an integrated end-to-end solution that you can use to build, for example, block chain in this case. >> Okay, I've asked you this before, but I've got to again and get it on record with Think. So, a lot of people would hear that and say Okay but it's a bunch of bespoke products that IBM has taken they've put a UI layer on top and called it a platform. So, what defines a platform and how have you not done that? >> Right. >> And actually created the platform? >> Right. So, we are taking the functionality of the existing parts and that's what differentiates us. Right? If you look at our governance portfolio, I can sit here and very confidently say no one can match that, so >> Dave: Sure. We obviously have that strength >> Real Tap >> Right, Real Tap. That we can bring. So we are bringing the functionality. But what we have done is we are taking the existing products and disintegrated in to micro services so we can make it cloud native. So that is a huge step for us, right? And then once you make that containerized and micro services it fits into the open platform that we talked about before. And now you have an end-to-end, well orchestrated pipeline that's available in the platform that can scale and be elastic as needed. So, it's not that we are bringing the products, we are bringing the functionality of it. >> But I want to keep on this for a second, so the experience for the user is different if you microserviced what you say because if you just did what I said and put a layer a UI layer on top, you would be going into these stovepipes and then cul-de-sac and then coming back >> Dinesh: Right. And coming back. So, the development effort for that must have been >> Oh, yeah. >> Fairly massive. You could have done the UI layer in, you know, in months. >> Right, right, right, then it is not really cloud native way of doing it, right? I mean, if you're just changing the UI and the experience, that's completely different. What we have done is that we have completely re-architected the underlying product suite to meet the experience and the underlying platform layer. So, what can happen? How long did this take? What kind of resources did you have to throw at this from a development standpoint? >> So this has been in development for 12-18 months. >> Yeah. >> And we put, you know, a tremendous amount of resources to make this happen. I mean, fortunately in our case we have the depth, we have the functionality. So it was about translating that into the cloud native way of doing the app development. >> So did you approach this with sort of multiple small teams? Or was there a larger team? What was your philosophy here? >> It was multiple small teams, right. So if you look at our governance portfolio we got to take our governance catalog, rewrite that code. If we look at our master data management portfolio, we got to take, so it's multiple of small teams with very core focus. >> I mean, I ask you these questions because I think it adds credibility to the claims that you're making about we have a platform not a series of bespoke products. >> Right and we demoed it. Actually tomorrow at 11, I'm going to deep dive into the architecture of the whole platform itself. How we built it. What are the components we used and I'm going to demo it. So the code is up and running and we are going to put it out there into Cube for everybody to go us it. >> At Mandalay Bay, where is that demo? >> It's in Mandalay Bay, yeah. >> Okay. >> We have a session at 11:30. >> Talk more about machine learning and how you've infused machine learning into the portfolio. >> Right. So, every part of our product portfolio has machinery so, I'll take two examples. One is DB2. So today, DB2 Optimizer is a cost-based optimizer. We have taken the optimizer and infused machine learning into it to say, you know, based on the query that's coming in take the right access path, predict the right access path and take it. And that has been such a great experience because we are seeing 30-50 percent performance improvement in most of the queries that we run through the machinery. So that's one. The other one is the classification, so let's say, you have a business term and you want to classify. So, if you have a zip code, we can use in our catalog to say there's an 80% chance this particular number is a zip code and then it can learn over time, if you tell it, no that's not a zip code, that's a post code in Canada. So the next time you put that in there it has learned. So every product we have infused machine learning and that's our goal is to become completely a cognitive platform pretty soon. I mean, you know, so that has also been a tremendous piece of work that we're doing. >> So what can we expect? I mean, you guys are moving fast. >> Yeah. >> We've seen you go from sort of a bespoke product company division to this platform division. Injecting now machine learning into the equation. You're bringing in new technologies like block chain, which you're able to do because you have a platform. >> Right. >> What should we expect in terms of the pace and the types of innovations that we could see going forward? What could you share with us without divulging secrets? >> Right. So, from a product perspective we want to infuse cognitive machine learning into every aspect of the product. So, we don't want to, we don't want our customers calling us, telling there's a problem. We want to be able to tell our customer a day or two hours ahead that there is a problem. So that is predictability, Right? So we want not just in the product, even in the services side, we want to infuse total machine learning into the product. From a platform perspective we want to make it completely open, extensible. So our partners can come and build on top of it. So every customer can take advantage of vertical and other solutions that they build. >> You get a platform, you get this fly-wheel effect, inject machine learning everywhere open API so you can bring in new technologies like block chain as they evolve. Dinesh, thank you very much for coming on the Cube. >> Oh, thank you so much. >> Always great to have you. >> It's a pleasure, thank you. >> Alright, keep it right there everybody. We'll be right back with our next guest. This is the Cube live from IBM Think 2018. We'll be right back. (techno music)
SUMMARY :
Brought to you by IBM. Dinesh, great to see you again. So, great to see you again. in New York City and in San Jose. all the clients together like this? I mean, you know, very positive. So on the analytics side, the data is growing So maybe you can talk I mean, you know, Not just the tools but we bring the skills So is that the challenge? obviously the different challenges pull you You've got the understanding of the tech. You know, the tools, most technology people So you're saying you can help. That's part of the ... I mean, let's just look at the data, right. Alright, tell us more about IBM Cloud Private for data. What's that all about? You have the ledger, where you are, obviously, Yes, because if you don't have the platform ... maybe is variable. And then you add the complexity of the different personas and how have you not done that? of the existing parts and that's what differentiates us. We obviously have that strength bringing the products, we are bringing So, the development effort You could have done the UI layer in, What kind of resources did you have to throw And we put, you know, a tremendous amount of resources So if you look at our governance portfolio I mean, I ask you these questions because I think So the code is up and running and we are going infused machine learning into the portfolio. So the next time you put that in there it I mean, you guys are moving fast. Injecting now machine learning into the equation. even in the services side, we want to infuse total You get a platform, you get this fly-wheel effect, This is the Cube live from IBM Think 2018.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Canada | LOCATION | 0.99+ |
Dinesh Nirmal | PERSON | 0.99+ |
New York | LOCATION | 0.99+ |
San Jose | LOCATION | 0.99+ |
New York City | LOCATION | 0.99+ |
three weeks | QUANTITY | 0.99+ |
two hours | QUANTITY | 0.99+ |
80% | QUANTITY | 0.99+ |
Mandalay Bay | LOCATION | 0.99+ |
11 months | QUANTITY | 0.99+ |
Dinesh | PERSON | 0.99+ |
Las Vegas | LOCATION | 0.99+ |
11:30 | DATE | 0.99+ |
two things | QUANTITY | 0.99+ |
a day | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
third day | QUANTITY | 0.98+ |
this week | DATE | 0.98+ |
three things | QUANTITY | 0.98+ |
two examples | QUANTITY | 0.98+ |
three | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
DB2 | TITLE | 0.98+ |
one | QUANTITY | 0.98+ |
12-18 months | QUANTITY | 0.97+ |
first | QUANTITY | 0.97+ |
decades | QUANTITY | 0.93+ |
terabytes a day | QUANTITY | 0.93+ |
double | QUANTITY | 0.9+ |
Think | ORGANIZATION | 0.88+ |
Vice-President | PERSON | 0.87+ |
few months ago | DATE | 0.85+ |
petabyte | QUANTITY | 0.85+ |
tomorrow at | DATE | 0.83+ |
couple of weeks ago | DATE | 0.81+ |
30-50 percent | QUANTITY | 0.79+ |
Danish | OTHER | 0.76+ |
IBM Think 2018 | EVENT | 0.75+ |
two paths | QUANTITY | 0.73+ |
Think | COMMERCIAL_ITEM | 0.72+ |
11 | DATE | 0.68+ |
multi | QUANTITY | 0.68+ |
Kubernetes | TITLE | 0.68+ |
a second | QUANTITY | 0.66+ |
Cloud Private | TITLE | 0.62+ |
Cloud Private for | TITLE | 0.61+ |
Cloud Private for Data | TITLE | 0.61+ |
IBM | EVENT | 0.56+ |
CTO | PERSON | 0.56+ |
Strata Conference | EVENT | 0.55+ |
2018 | EVENT | 0.54+ |
Cube | TITLE | 0.5+ |
Cube | COMMERCIAL_ITEM | 0.45+ |
Cube | PERSON | 0.27+ |
Dinesh Nirmal, IBM | Machine Learning Everywhere 2018
>> Announcer: Live from New York, it's theCUBE, covering Machine Learning Everywhere: Build Your Ladder to AI. Brought to you by IBM. >> Welcome back to Midtown, New York. We are at Machine Learning Everywhere: Build Your Ladder to AI being put on by IBM here in late February in the Big Apple. Along with Dave Vellante, I'm John Walls. We're now joined by Dinesh Nirmal, who is the Vice President of Analytics Development and Site Executive at the IBM Silicon Valley lab, soon. Dinesh, good to see you, this morning, sir. >> Thank you, John. >> Fresh from California. You look great. >> Thanks. >> Alright, you've talked about this, and it's really your world: data, the new normal. Explain that. When you say it's the new normal, what exactly... How is it transforming, and what are people having to adjust to in terms of the new normal. >> So, if you look at data, I would say each and every one of us has become a living data set. Our age, our race, our salary. What our likes or dislikes, every business is collecting every second. I mean, every time you use your phone, that data is transmitted somewhere, stored somewhere. And, airlines for example, is looking, you know, what do you like? Do you like an aisle seat? Do you like to get home early? You know, all those data. >> All of the above, right? >> And petabytes and zettabytes of data is being generated. So now, businesses' challenge is that how do you take that data and make insights out of it to serve you as a better customer. That's where I've come to, but the biggest challenge is that, how do you deal with this tremendous amount of data? That is the challenge. And creating sites out of it. >> That's interesting. I mean, that means the definition of identity is really... For decades it's been the same, and what you just described is a whole new persona, identity of an individual. >> And now, you take the data, and you add some compliance or provisioning like GDPR on top of it, all of a sudden how do-- >> John: What is GDPR? For those who might not be familiar with it. >> Dinesh: That's the regulatory term that's used by EU to make sure that, >> In the EU. >> If me as a customer come to an enterprise, say, I don't want any of my data stored, it's up to you to go delete that data completely, right? That's the term that's being used. And that goes into effect in May. How do you make sure that that data gets completely deleted by that time the customer has... How do you get that consent from the customer to go do all those... So there's a whole lot of challenges, as data multiplies, how do you deal with the data, how do you create insights to the data, how do you create consent on the data, how do you be compliant on that data, how do you create the policies that's needed to generate that data? All those things need to be... Those are the challenges that enterprises face. >> You bring up GDPR, which, for those who are not familiar with it, actually went into effect last year but the fines go into effect this year, and the fines are onerous, like 4% of turnover, I mean it's just hideous, and the question I have for you is, this is really scary for companies because they've been trying to catch up to the big data world, and so they're just throwing big data projects all over the place, which is collecting data, oftentimes private information, and now the EU is coming down and saying, "Hey you have to be able to, if requested, delete that." A lot of times they don't even know where it is, so big challenge. Are you guys, can you help? >> Yeah, I mean, today if you look at it, the data exists all over the place. I mean, whether it's in your relational database or in your Hadoop, unstructured data, whereas you know, optics store, it exists everywhere. And you have to have a way to say where the data is and the customer has to consent given to go, for you to look at the data, for you to delete the data, all those things. We have tools that we have built and we have been in the business for a very long time for example our governance catalog where you can see all the data sources, the policies that's associated with it, the compliance, all those things. So for you to look through that catalog, and you can see which data is GDPR compliant, which data is not, which data you can delete, which data you cannot. >> We were just talking in the open, Dave made the point that many companies, you need all-stars, not just somebody who has a specialty in one particular area, but maybe somebody who's in a particular regiment and they've got to wear about five different hats. So how do you democratize data to the point that you can make these all-stars? Across all kinds of different business units or different focuses within a company, because all of a sudden people have access to great reams of information. I've never had to worry about this before. But now, you've got to spread that wealth out and make everybody valuable. >> Right, really good question. Like I said, the data is existing everywhere, and most enterprises don't want to move the data. Because it's a tremendous effort to move from an existing place to another one and make sure the applications work and all those things. We are building a data virtualization layer, a federation layer, whereby which if you are, let's say you're a business unit. You want to get access to that data. Now you can use that federational data virtualization layer without moving data, to go and grab that small piece of data, if you're a data scientist, let's say, you want only a very small piece of data that exists in your enterprise. You can go after, without moving the data, just pick that data, do your work, and build a model, for example, based on that data. So that data virtualization layer really helps because it's basically an SQL statement, if I were to simplify it. That can go after the data that exists, whether it's at relational or non-relational place, and then bring it back, have your work done, and then put that data back into work. >> I don't want to be a pessimist, because I am an optimist, but it's scary times for companies. If they're a 20th century organization, they're really built around human expertise. How to make something, how to transact something, or how to serve somebody, or consult, whatever it is. The 21st century organization, data is foundational. It's at the core, and if my data is all over the place, I wasn't born data-driven, born in the cloud, all those buzzwords, how do traditional organizations catch up? What's the starting point for them? >> Most, if not all, enterprises are moving into a data-driven economy, because it's all going to be driven by data. Now it's not just data, you have to change your applications also. Because your applications are the ones that's accessing the data. One, how do you make an application adaptable to the amount of data that's coming in? How do you make accuracy? I mean, if you're building a model, having an accurate model, generating accuracy, is key. How do you make it performant, or govern and self-secure? That's another challenge. How do you make it measurable, monitor all those things? If you take three or four core tenets, that's what the 21st century's going to be about, because as we augment our humans, or developers, with AI and machine learning, it becomes more and more critical how do you bring these three or four core tenets into the data so that, as the data grows, the applications can also scale. >> Big task. If you look at the industries that have been disrupted, taxis, hotels, books, advertising. >> Dinesh: Retail. >> Retail, thank you. Maybe less now, and you haven't seen that disruption yet in banks, insurance companies, certainly parts of government, defense, you haven't seen a big disruption yet, but it's coming. If you've got the data all over the place, you said earlier that virtually every company has to be data-driven, but a lot of companies that I talk to say, "Well, our industry is kind of insulated," or "Yeah, we're going to wait and see." That seems to me to be the recipe for disaster, what are your thoughts on that? >> I think the disruption will come from three angles. One, AI. Definitely that will change the way, blockchain, another one. When you say, we haven't seen in the financial side, blockchain is going to change that. Third is quantum computing. The way we do compute is completely going to change by quantum computing. So I think the disruption is coming. Those are the three, if I have to predict into the 21st century, that will change the way we work. I mean, AI is already doing a tremendous amount of work. Now a machine can basically look at an image and say what it is, right? We have Watson for cancer oncology, we have 400 to 500,000 patients being treated by Watson. So AI is changing, not just from an enterprise perspective, but from a socio-economic perspective and a from a human perspective, so Watson is a great example for that. But yeah, disruption is happening as we speak. >> And do you agree that foundational to AI is the data? >> Oh yeah. >> And so, with your clients, like you said, you described it, they've got data all over the place, it's all in silos, not all, but much of it is in silos. How does IBM help them be a silo-buster? >> Few things, right? One, data exists everywhere. How do you make sure you get access to the data without moving the data, that's one. But if you look at the whole lifecycle, it's about ingesting the data, bringing the data, cleaning the data, because like you said, data becomes the core. Garbage in, garbage out. You cannot get good models unless the data is clean. So there's that whole process, I would say if you're a data scientist, probably 70% of your time is spent on cleaning the data, making the data ready for building a model or for a model to consume. And then once you build that model, how do you make sure that the model gets retrained on a regular basis, how do you monitor the model, how do you govern the model, so that whole aspect goes in. And then the last piece is visualizational reporting. How do you make sure, once the model or the application is built, how do you create a report that you want to generate or you want to visualize that data. The data becomes the base layer, but then there's a whole lifecycle on top of it based on that data. >> So the formula for future innovation, then, starts with data. You add in AI, I would think that cloud economics, however we define that, is also a part of that. My sense is most companies aren't ready, what's your take? >> For the cloud, or? >> I'm talking about innovation. If we agree that innovation comes from the data plus AI plus you've got to have... By cloud economics I mean it's an API economy, you've got massive scale, those kinds of, to compete. If you look at the disruptions in taxis and retail, it's got cloud economics underneath it. So most customers don't really have... They haven't yet even mastered cloud economics, yet alone the data and the AI component. So there's a big gap. >> It's a huge challenge. How do we take the data and create insights out of the data? And not just existing data, right? The data is multiplying by the second. Every second, petabytes or zettabytes of data are being generated. So you're not thinking about the data that exists within your enterprise right now, but now the data is coming from several different places. Unstructured data, structured data, semi-structured data, how do you make sense of all of that? That is the challenge the customers face, and, if you have existing data, on top of the newcoming data, how do you predict what do you want to come out of that. >> It's really a pretty tough conundrum that some companies are in, because if you're behind the curve right now, you got a lot of catching up to do. So you think that we have to be in this space, but keeping up with this space, because the change happens so quickly, is really hard, so we have to pedal twice as fast just to get in the game. So talk about the challenge, how do you address it? How do you get somebody there to say, "Yep, here's your roadmap. "I know it's going to be hard, "but once you get there you're going to be okay, "or at least you're going to be on a level playing field." >> I look at the three D's. There's the data, there's the development of the models or the applications, and then the deployment of those models or applications into your existing enterprise infrastructure. Not only the data is changing, but that development of the models, the tools that you use to develop are also changing. If you look at just the predictive piece, I mean look from the 80's to now. You look at vanilla machine learning versus deep learning, I mean there's so many tools available. How do you bring it all together to make sense which one would you use? I think, Dave, you mentioned Hadoop was the term from a decade ago, now it's about object store and how do you make sure that data is there or JSON and all those things. Everything is changing, so how do you bring, as an enterprise, you keep up, afloat, on not only the data piece, but all the core infrastructure piece, the applications piece, the development of those models piece, and then the biggest challenge comes when you have to deploy it. Because now you have a model that you have to take and deploy in your current infrastructure, which is not easy. Because you're infusing machine learning into your legacy applications, your third-party software, software that was written in the 60's and 70's, it's not an easy task. I was at a major bank in Europe, and the CTO mentioned to me that, "Dinesh, we built our model in three weeks. "It has been 11 months, we still haven't deployed it." And that's the reality. >> There's a cultural aspect too, I think. I think it was Rob Thomas, I was reading a blog that he wrote, and he said that he was talking to a customer saying, "Thank god I'm not in the technology industry, "things change so fast I could never, "so glad I'm not a software company." And Rob's reaction was, "Uh, hang on. (laughs) "You are in the technology business, "you are a software company." And so there's that cultural mindset. And you saw it with GE, Jeffrey Immelt said, "I went to bed an industrial giant, "woke up a software company." But look at the challenges that industrial giant has had transforming, so... They need partners, they need people that have done this before, they need expertise and obviously technology, but it's people and process that always hold it up. >> I mean technology is one piece, and that's where I think companies like IBM make a huge difference. You understand enterprise. Because you bring that wealth of knowledge of working with them for decades and they understand your infrastructure, and that is a core element, like I said the last piece is the deployment piece, how do you bring that model into your existing infrastructure and make sure that it fits into that architecture. And that involves a tremendous amount of work, skills, and knowledge. >> Job security. (all laugh) >> Dinesh, thanks for being with us this morning, we appreciate that and good luck with the rest of the event, here in New York City. Back with more here on theCUBE, right after this. (calming techno music)
SUMMARY :
Brought to you by IBM. and Site Executive at the IBM Silicon Valley lab, soon. You look great. When you say it's the new normal, what exactly... I mean, every time you use your phone, how do you take that data and make insights out of it and what you just described is a whole new persona, For those who might not be familiar with it. How do you get that consent from the customer and the question I have for you is, given to go, for you to look at the data, So how do you democratize data to the point a federation layer, whereby which if you are, It's at the core, and if my data is all over the place, One, how do you make If you look at the industries that have been disrupted, Maybe less now, and you haven't seen that disruption yet When you say, we haven't seen in the financial side, like you said, you described it, how do you make sure that the model gets retrained So the formula for future innovation, If you look at the disruptions in taxis and retail, how do you predict what do you want to come out of that. So talk about the challenge, how do you address it? and how do you make sure that data is there And you saw it with GE, Jeffrey Immelt said, how do you bring that model the rest of the event, here in New York City.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
Tom | PERSON | 0.99+ |
Marta | PERSON | 0.99+ |
John | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
David | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Peter Burris | PERSON | 0.99+ |
Chris Keg | PERSON | 0.99+ |
Laura Ipsen | PERSON | 0.99+ |
Jeffrey Immelt | PERSON | 0.99+ |
Chris | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Chris O'Malley | PERSON | 0.99+ |
Andy Dalton | PERSON | 0.99+ |
Chris Berg | PERSON | 0.99+ |
Dave Velante | PERSON | 0.99+ |
Maureen Lonergan | PERSON | 0.99+ |
Jeff Frick | PERSON | 0.99+ |
Paul Forte | PERSON | 0.99+ |
Erik Brynjolfsson | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Andrew McCafee | PERSON | 0.99+ |
Yahoo | ORGANIZATION | 0.99+ |
Cheryl | PERSON | 0.99+ |
Mark | PERSON | 0.99+ |
Marta Federici | PERSON | 0.99+ |
Larry | PERSON | 0.99+ |
Matt Burr | PERSON | 0.99+ |
Sam | PERSON | 0.99+ |
Andy Jassy | PERSON | 0.99+ |
Dave Wright | PERSON | 0.99+ |
Maureen | PERSON | 0.99+ |
ORGANIZATION | 0.99+ | |
Cheryl Cook | PERSON | 0.99+ |
Netflix | ORGANIZATION | 0.99+ |
$8,000 | QUANTITY | 0.99+ |
Justin Warren | PERSON | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
2012 | DATE | 0.99+ |
Europe | LOCATION | 0.99+ |
Andy | PERSON | 0.99+ |
30,000 | QUANTITY | 0.99+ |
Mauricio | PERSON | 0.99+ |
Philips | ORGANIZATION | 0.99+ |
Robb | PERSON | 0.99+ |
Jassy | PERSON | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Mike Nygaard | PERSON | 0.99+ |
Dinesh Nirmal, IBM | CUBEConversation
(upbeat music) >> Hi everyone. We have a special program today. We are joined by Dinesh Nirmal, who is VP of Development and Analytics, for Analytics at the IBM company and Dinesh has an extremely broad perspective on what's going on in this part of the industry, and IBM has a very broad portfolio. So, between the two of us, I think we can cover a lot of ground today. So, Dinesh, welcome. >> Oh thank you George. Great to be here. >> So just to frame the discussion, I wanted to hit on sort of four key highlights. One is balancing the compatibility across cloud on-prem, and edge versus leveraging specialized services that might be on any one of those platforms. And then harmonizing and simplifying both the management and the development of services across these platforms. You have that trade-off between: do I do everything compatibly; or can I take advantage of platforms, specific stuff? And then, we've heard a huge amount of noise on Machine Learning. And everyone says they're democratizing it. We want to hear your perspective on how you think that's most effectively done. And then, if we have time, the how to manage Machine Learning feedback, data feedback loops to improve the models. So, having started with that. >> So you talked about the private cloud and the public cloud and then, how do you manage the data and the models, or the other analytical assets across the hybrid nature of today. So, if you look at our enterprises, it's a hybrid format that most customers adopt. I mean you have some data in the public side; but you have your mission critical data, that's very core to your transactions, exist in the private cloud. Now, how do you make sure that the data that you've pushed on the cloud, that you can go use to build models? And then you can take that model deployed on-prem or on public cloud. >> Is that the emerging sort of mainstream design pattern, where mission critical systems are less likely to move, for latency, for the fact that they're fused to their own hardware, but you take the data, and the researching for the models happens up in the cloud, and then that gets pushed down close to where the transaction decisions are. >> Right, so there's also the economics of data that comes into play, so if you are doing a, you know, large scale neural net, where you have GPUs, and you want to do deep learning, obviously, you know, it might make more sense for you to push it into the cloud and be able to do that or one of the other deep learning frameworks out there. But then you have your core transactional data that includes your customer data, you know, or your customer medical data, which I think some customers might be reluctant to push on a public cloud, and then, but you still want to build models and predict and all those things. So I think it's a hybrid nature, depending on the sensitivities of the data. Customers might decide to put it on cloud versus private cloud which is in their premises, right? So then how do you serve those customer needs, making sure that you can build a model on the cloud and that you can deploy that model on private cloud or vice versa. I mean, you can build that model on private cloud or only on private, and then deployed on your public cloud. Now the challenge, one last statement, is that people think, well, once I build a model, and I deploy it on public cloud, then it's easy, because it's just an API call at that time, just to call that model to execute the transactions. But that's not the case. You take support vector machine, for example, right, that still has vectors in there, that means your data is there, right, so even though you're saying you're deploying the model, you still have sensitive data there, so those are the kind of things customers need to think about before they go deploy those models. >> So I might, this is a topic for our Friday interview with a member of the Watson IT family, but it's not so black and white when you say we'll leave all your customer data with you, and we'll work on the models, because it, sort of like, teabags, you know, you can take the customer's teabag and squeeze some of the tea out, in your IBM or public cloud, and give them back the teabag, but you're getting some of the benefit of this data. >> Right, so like, it depends, depends on the algorithms you build. You could take a linear regression, and you don't have the challenges I mentioned, in support of retro machine, because none of the data is moving, it's just modeled. So it depends, I think that's where, you know, like Watson has done, will help tremendously because the data is secure in that sense. But if you're building on your own, it's a different challenge, you've got to make sure you pick the right algorithms to do that. >> Okay, so let's move on to the modern sort of what we call operational analytic pipeline, where the key steps are ingest, process, analyze, predict, serve, and you can drill down on those more. Today there's, those pipelines are pretty much built out of multi-vendor components. How do you see that evolving under pressures of, or tension between simplicity, coming from one vendor, and the pieces all designed together, and the specialization, where you want to have a, you know, unique tool in one component. >> Right, so you're exactly right. So you can take a two prong approach. One is, you can go to a cloud provider, and get each of the services, and you stitch it together. That's one approach. A challenging approach, but that has its benefits, right, I mean, you bring some core strengths from each vendor into it. The other one is the integrate approach, where you ingest the data, you shape or cleanse the data, you get it prepared for analytics, you build the model, you predict, you visualize. I mean, that all comes in one. The benefit there is you get the whole stack in one, you have one you have a whole pipeline that you can execute, you have one service provider that's giving them services, it's managed. So all those benefits come with it, and that's probably the preferred way for it integrated all together in one stack, I think that's the path most people go towards, because then you have the whole pipeline available to you, and also the services that comes with it. So any updates that comes with it, and how do you make sure, if you take the first round, one challenge you have is how do you make sure all these services are compatible with each other? How do you make sure they're compliant? So if you're an insurance company, you want it to be HIPAA compliant. Are you going to individually make sure that each of these services are HIPAA compliant? Or would you get from one integrated provider, you can make sure they are HIPAA compliant, tests are done, so all those benefits, to me, outweigh you going, putting unmanaged service all together, and then creating a data link to underlay all of it. >> Would it be fair to say, to use an analogy, where Hadoop, being sort of, originating in many different Apache products, is a quasi-multi vendor kind of pipeline, and the state of, the state of the machine learning analytic pipeline, is still kind of multi-vendor today. You see that moving toward single vendor pipeline, who do you see as the sort of last man standing? >> So, I mean, I can speak from an IBM perspective, I can say that the benefit that a company, a vendor like IBM brings forward, is like, so the different, public or private cloud or hybrid, you obviously have the choice of going to public cloud, you can get the same service on public cloud, so you get a hybrid experience, so that's one aspect of it. Then, if you get the integrated solution, all the way from ingest to visualization, you have one provider, it's tested, it's integrated, you know, it's combined, it works well together, so I would say, going forward, if you look at it purely from an enterprise perspective, I would say integrated solutions is the way to go, because that what will be the last man standing. I'll give you an example. I was with a major bank in Europe, about a month ago, and I took them through our data science experience, our machine learning project and all that, and you know, the CTO's take was that, Dinesh, I got it. Building the model itself, it only took us two days, but incorporating our model into our existing infrastructure, it has been 11 months, we haven't been able to do it. So that's the challenge our enterprises face, and they want an integrated solution to bring that model into their existing infrastructure. So that's, you know, that's my thought. >> Today though, let's talk about the IBM pipeline. Spark is core, Ingest is, off the-- >> Dinesh: Right, so you can do spark streaming, you can use Kafka, or you can use infostream which is our proprietary tool. >> Right, although, you wouldn't really use structured streaming for ingest, 'cause of the back pressure? >> Right, so they are-- >> The point that I'm trying to make is, it's still multi-vendor, and then the serving side, I don't know, where, once the analysis is done and predictions are made, some sort of sequel database has to take over, so it's, today, it's still pretty multi vendor. So how do you see any of those products broadening their footprints so that the number of pieces decreases. >> So good question, they are all going to get into end pipeline, because that's where the value is, unless you provide an integrated end to end solution for a customer, especially parts customer it's all about putting it all together, and putting these pieces together is not easy, even if you ingest the data, IOP kind of data, a lot of times, 99% of the time, data is not clean, unless you're in a competition where you get cleansed data, in real world, that never happens. So then, I would say 80% of a data scientists time is spent on cleaning the data, shaping the data, preparing the data to build that pipeline. So for most customers, it's critical that they get that end to end, well oiled, well connected solution integrated solution, than take it from each vendor, every isolated solution. To answer your question, yes, every vendor is going to move into the ingest, data cleansing phase, transformation, and the building the pipeline and then visualization, if you look at those five steps, has to be developed. >> But just building the data cleansing and transformation, having it in your, native to your own pipeline, that doesn't sound like it's going to solve the problem of messy data that needs, you know, human supervision to correct. >> I mean, so there is some level of human supervision to be sure, so I'll give you an example, right, so when data from an insurance company goes, a lot of times, the gender could be missing, how do you know if it's a male or female? Then you got to build another model to say, you know, this patient has gone for a prostate exam, you know, it's a male, gynecology is a female, so you have to do some intuitary work in there, to make sure that the data is clean, and then there's some human supervision to make sure that this is good to build models, because when you're executing that pipeline in real time, >> Yeah. >> It's all based on the past data, so you want to make sure that the data is as clean as possible to train and model, that you're going to execute on. >> So, let me ask you, turning to a slide we've got about complexity, and first, for developers, and then second, for admins, if we take the steps in the pipeline, as ingest, process, analyze, predict, serve, and sort of products or product categories as Kafka, Spark streaming and sequel, web service for predict, and MPP sequel, or no sequel for serve, even if they all came from IBM, would it be possible to unify the data model, the addressing and name space, and I'm just kicking off a few that I can think of, programming model, persistence, transaction model, workflow, testing integration, there's one thing to say it's all IBM, and then there's another thing, so that the developer working with it, sees as it as one suite. >> So it has to be validated, and that's the benefit that IBM brings already, because we obviously test each segment to make sure it works, but when you talk about complexity, building the model is one, you know, development of the model, but now the complexity also comes in the deployment of the model, now we talk about the management of the model, where, how you monitor it? When was the model deployed, was it deployed in tests, was it deployed in production, and who changed that model last, what was changed, and how is it scoring? Is it scoring high or low? You want to get notification when the model starts going low. So complexity is all the way across, all the way from getting the data, in cleaning the data, developing the model, it never ends. And the other benefit that IBM has added is the feedback loop, where when you talk about complexity, it reduces the complexity, so today, if the model scores low, you have to take it offline, retrain the model based on the new data, and then redeploy it. Usually for enterprises, there is slots where you can take it offline, put it back online, all these things, so it's a process. What we have done is created a feedback loop where we are training the model in real time, using real time data, so the model is continuously-- >> Online learning. >> Online learning. >> And challenger, champion, or AB testing to see which one is more robust. >> Right, so you can do that, I mean, you could have multiple models where you can do AB testing, but in this case, you can condition, train the model to say, okay, this model scores the best. And then, another benefit is that, if you look at the whole machine learning process, there's the data, there's development, there's deployment. On development side, more and more it's getting commoditized, meaning picking the right algorithm, there's a lot of tools, including IBM, where he can say, question what's the right one to use for this, so that piece is getting a little, less complex, I don't want to say easier, but less complex, but the data cleansing and the deployment, these are two enterprises, when you have thousands of models how do you make sure that you deploy the right model. >> So you might say that the pipeline for managing the model is sort of separate from the original data pipeline, maybe it includes the same technology, or as much of the same technology, but once your pipeline, your data pipeline is in production, the model pipeline has to keep cycling through. >> Exactly, so the data pipeline could be changing, so if you take a lone example, right, a lot of the data that goes in the model pipeline, is static, I mean, my age, it's not going to change every day, I mean, it is, but you know, the age that goes into my salary, my race, my gender, those are static data that you can take from data and put it in there, but then there's also real time data that's coming, my loan amount, my credit score, all those things, so how do you bring that data pipeline between real time and static data, into the model pipeline, so the model can predict accurately and based on the score dipping, you should be able to re-try the model using real time data. >> I want to take, Dinesh, to the issue of a multi-vendor stack again, and the administrative challenges, so here, we look at a slide that shows me just rattling off some of the admin challenges, governance, performance modeling, scheduling orchestration, availability, recovering authentication, authorization, resource isolation, elasticity, testing integration, so that's the Y-axis, and then for every different product in the pipeline, as the access, say Kafka, Spark, structured streaming MPP, sequel, no sequel, so you got a mess. >> Right. >> Most open source companies are trying to make life easier for companies by managing their software as a service for the customer, and that's typically how they monetize. But tell us what you see the problem is, or will be with that approach. >> So, great question. Let me take a very simple example. Probably most of our audience know about GDPR, which is the European law to write to forget. So if you're an enterprise, and say, George, I want my data deleted, you have to delete all of my data within a period of time. Now, that's where one of the aspects you talked about with governance comes in. How do you make sure you have governance across not just data but your individual assets? So if you're using a multi-vendor solution, in all of that, that state governance, how do I make sure that data get deleted by all these services that's all tied together. >> Let me maybe make an analogy. On CSI, when they pick up something at the crime scene, they got to make sure that it's bagged, and the chain of custody doesn't lose its integrity all the way back to the evidence room. I assume you're talking about something like that. >> Yeah, something similar. Where the data, as it moves between private cloud, public cloud, analytical assets, is using that data, all those things need to work seamlessly for you to execute that particular transaction to delete data from everywhere. >> So that's, it's not just administrative costs, but regulations that are pushing towards more homogenous platforms. >> Right, right, and even if you take some of the other things on the stack, monitoring, logging, metering, provides some of those capabilities, but you have to make sure when you put all these services together, how are they going to integrate all together? You have one monitoring stack, so if you're pulling you know, your IOT kind of data into a data center, or your whole stack evaluation, how do you make sure you're getting the right monitoring data across the board? Those are the kind of challenges that you will have. >> It's funny you mention that, because we were talking to an old Lotus colleague of mine, who was CTO of Microsoft's IT organization, and we were talking about how the cloud vendors can put machine learning application, machine learning management application across their properties, or their services, but he said one of the first problems he'll encounter is the telemetry, like it's really easy on hardware, CPUs, utilization, memory utilization, a noise enabler for iO, but as you get higher up in the application services, it becomes much more difficult to harmonize, so that a program can figure out what's going wrong. >> Right, and I mean, like anomaly detection, right? >> Yes. >> I mean, how do you make sure you're seeing patterns where you can predict something before it happens, right? >> Is that on the road map for...? >> Yeah, so we're already working with some big customers to say, if you have a data center, how do you look at outage to predict what can go wrong in the future, root cause analysis, I mean, that is a huge problem solved. So let's say customer hit a problem, you took an outage, what caused it? Because today, you have specialists who will come and try to figure out what the problem is, but can we use machine learning or deep learning to figure out, is it a fix that was missing, or an application got changed that caused a CPU spike, that caused the outage? So that whole cost analysis is the one that's the hardest to solve, because you are talking about people's decades worth of knowledge, now you are influencing a machine to do that prediction. >> And from my understanding, root cause analysis is most effective when you have a rich model of how your, in this case, data structure and apps are working, and there might be many little models, but they're held together by some sort of knowledge graph that says here is where all the pieces fit, these are the pieces below these, sort of as peers to these other things, how does that knowledge graph get built in, and is this the next generation of a configuration management database. >> Right, so I call it the self-healing, self-managing, self-fixing data center. It's easy for you to turn up the heat or A/C, the temperature goes down, I mean, those are good, but the real value for a customer is exactly what you mentioned, building up that knowledge graft from different models that all comes together, but the hardest part, is, how do you, predicting an anomaly is one thing, but getting to the root cause is a different thing, because at that point, now you're saying, I know exactly what's caused this problem, and I can prevent it from happening again. That's not easy. We are working with our customers to figure out how do we get to the root cause analysis, but it's all about building the knowledge graph with multiple models coming from different systems, today, I mean enterprises have different systems from multi-vendors. We have to bring all that monitoring data into one source, and that's where that knowledge comes in, and then different models will feed that data, and then you need to prime that data, using deep learning algorithms to say, what caused this? >> Okay, so this actually sounds extremely relevant, although we're probably, in the interest of time, going to have to dig down on that one another time, but, just at a high level, it sounds like the knowledge graph is sort of your web or directory, into how local components or local models work, and then, knowing that, if it sees problems coming up here, it can understand how it affects something else tangentially. >> So think of knowledge graph as a neural net, because it's just building new neural net based on the past data, and it has that built-in knowledge where it says, okay, these symptoms seem to be a problem that I have encountered in the past. Now I can predict the root cause because I know this happened in the past. So it's kind of like you putting that net to build new problem determinations as it goes along. So it's a complex task. It's not easy to get to root cause analysis. But that's something we are aggressively working on developing. >> Okay, so let me ask, let's talk about sort of democratizing machine learning and the different ways of doing that. You've actually talked about the big pain points, maybe not so sexy, but that are critical, which is operationalizing the models, and preparing the data. Let me bounce off you some of the other approaches. One that we have heard from Amazon is that they're saying, well, data expunging might be an issue, and operationalizing the models might be an issue, but the biggest issue in terms of making this developer ready, is we're going to take the machine learning we use to run our business, whether it's merchandising fashion, running recommendation engines, managing fulfillment or logistics, and just like I did with AWS, they're dog-fooding it internally, and then they're going to put it out on AWS as a new layer of a platform. Where do you see that being effective, and where less effective? >> Right, so let me answer the first part of your question, the democratization of learning. So that happens when for example, a real estate agent who has no idea about machine learning, be able to come and predict the house prices in this area. That's to me, is democratizing. Because at that time, you have made it available to everyone, everyone can use it. But that comes back to our first point, which is having that clean set of data. You can build all the pre-canned pipelines out there, but if you're not feeding the set of data into, none of this, you know. Garbage in, garbage out, that's what you're going to get. So when we talk about democratization, it's not that easy and simple because you can build all this pre-canned pipelines that you have used in-house for your own purposes, but every customer has many unique cases. So if I take you as a bank, your fraud detection methods is completely different than me as a bank, my limit for fraud detection could be completely different. So there is always customization that's involved, the data that's coming in is different, so while it's a buzzword, I think there's knowledge that people need to feed it, there's models that needs to be tuned and trained, and there's deployment that is completely different, so you know, there is work that has to be done. >> So then what I'm taking away from what you're saying is, you don't have to start from ground zero with your data, but you might want to add some of your data, which is specialized, or slightly different from what the pre-trained model is, you still have to worry about operationalizing it, so it's not a pure developer ready API, but it uplevels the skills requirement so that it's not quite as demanding as working with TensorFlow or something like that. >> Right, I mean, so you can always build pre-canned pipelines and make it available, so we have already done that. For example, fraud detection, we have pre-canned pipelines for IT analytics, we have pre-canned pipelines. So it's nothing new, you can always do what you have done in house, and make it available to the public or the customers, but then they have to take it and have to do customization to meet their demands, bring their data to re-train the model, all those things has to be done, it's not just about providing the model, but every customer use case is completely different, whether you are looking at fraud detection from that one bank's perspective, not all banks are going to do the same thing. Same thing for predicting, for example, the loan, I mean, your loan approval process is going to be completely different than me as a bank loan approval process. >> So let me ask you then, and we're getting low on time here, but what would you, if you had to characterize Microsoft, Azure, Google, Amazon, as each bringing to bear certain advantages and disadvantages, and you're now the ambassador, so you're not a representative of IBM, help us understand the sweet spot for each of those. Like you're trying to fix the two sides of the pipeline, I guess, thinking of it like a barbell, you know, where are the others based on their data assets and their tools, where do they need to work. >> So, there's two aspects of it, there's enterprise aspect, so as an enterprise, I would like to say, it's not just about the technology, but there's also the services aspect. If my model goes down in the middle of the night, and my banking app is down, who do I call? If I'm using a service that is available on the cloud provider which is open source, do I have the right amount of coverage to call somebody and fix it. So there's the enterprise capabilities, availabilities, reliability, that is different, than a developer comes in, has a CSV file that he or she wants to build a model to predict something, that's different, this is different, two different aspects. So if you talk about, you know, all these vendors, if I'm bearing an enterprise card, some of the things I would look is, can I get an integrated solution, end to end on the machine learning platform. >> And that means end to end in one location, >> Right. >> So you don't have network issues or latency and stuff like that. >> Right, it's an integrated solution, where I can bring in the data, there's no challenges to latency, those kinds of things, and then can I get the enterprise level service, SLA all those things, right? So, in there, the named vendors obviously have an upper hand, because they are preferred to enterprises than a brand new open source that will come along, but then there is, within enterprises, there are a line of businesses building models, using some of the open source vendors, which is okay, but eventually they'd have to get deployed and then how do you make sure you have that enterprise capabilities up there. So if you ask me, I think each vendor brings some capabilities. I think the benefit IBM brings in is, one, you have the choice or the freedom to bring in cloud or on-prem or hybrid, you have all the choices of languages, like we support R, Python Spar, Spark, I mean, SPS, so I mean, the choice, the freedom, the reliability, the availability, the enterprise nature, that's where IBM comes in and differentiates, and that's for our customers, a huge plus. >> One last question, and we're really out of time, in terms of thinking about a unified pipeline, when we were at Spark Summit, sitting down with Matei Zaharia and Reynold Shin, the question came up, the data breaks has an incomplete pipeline, no persistence, no ingest, not really much in the way of serving, but boy are they good at, you know, data transmigration, and munging and machine learning, but they said they consider it part of their ultimate responsibility to take control. And on the ingest side it's Kafka, the serving side, might be Redis or something else, or the Spark databases like Snappy Data and Splice Machine. Spark is so central to IBM's efforts. What might a unified Spark pipeline look like? Have you guys thought about that? >> It's not there, obviously they probably could be working on it, but for our purpose, Spark is critical for us, and the reason we invested in Spark so much is because of the executions, where you can take a tremendous amount of data, and, you know, crunch through it in a very short amount of time, that's the reason, we also invented Spark Sequel, because we have a good chunk of customers still use Sequel heavily, We put a lot of work into the Spark ML, so we are continuing to invest, and probably they will get to and integrated into a solution, but it's not there yet, but as it comes along, we'll adapt. If it meets our needs and demands, and enterprise can do it, then definitely, I mean, you know, we saw that Spark's core engine has the ability to crunch a tremendous amount of data, so we are using it, I mean, 45 of our internal products use Spark as our core engine. Our DSX, Data Science Experience, has Spark as our core engine. So, yeah, I mean, today it's not there, but I know they're probably working on it, and if there are elements of this whole pipeline that comes together, that is convenient for us to use, and at enterprise level, we will definitely consider using it. >> Okay, on that note, Dinesh, thanks for joining us, and taking time out of your busy schedule. My name is George Gilbert, I'm with Dinesh Nirmal from IBM, VP of Analytics Development, and we are at the Cube studio in Palo Alto, and we will be back in the not too distant future, with more interesting interviews with some of the gurus at IBM. (peppy music)
SUMMARY :
So, between the two of us, I think Oh thank you George. the how to manage Machine Learning feedback, that you can go use to build models? but you take the data, and the researching for and that you can deploy that model on private cloud but it's not so black and white when you say and you don't have the challenges I mentioned, and the specialization, where you want to have and get each of the services, and you stitch it together. who do you see as the sort of last man standing? So that's, you know, that's my thought. Spark is core, Ingest is, off the-- Dinesh: Right, so you can do spark streaming, so that the number of pieces decreases. and then visualization, if you look at those five steps, of messy data that needs, you know, human supervision so you want to make sure that the data is as clean as in the pipeline, as ingest, process, analyze, if the model scores low, you have to take it offline, to see which one is more robust. Right, so you can do that, I mean, you could have So you might say that the pipeline for managing I mean, it is, but you know, the age that goes MPP, sequel, no sequel, so you got a mess. But tell us what you see the problem is, Now, that's where one of the aspects you talked about and the chain of custody doesn't lose its integrity for you to execute that particular transaction to delete but regulations that are pushing towards more Those are the kind of challenges that you will have. It's funny you mention that, because we were to say, if you have a data center, how do you look at most effective when you have a rich model and then you need to prime that data, using deep learning but, just at a high level, it sounds like the knowledge So it's kind of like you putting that net Let me bounce off you some of the other approaches. pipelines that you have used in-house for your own purposes, the pre-trained model is, you still have to worry So it's nothing new, you can always do what you have So let me ask you then, and we're getting low on time So if you talk about, you know, all these vendors, So you don't have network issues or latency and then how do you make sure you have that but boy are they good at, you know, where you can take a tremendous amount of data, of the gurus at IBM.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Microsoft | ORGANIZATION | 0.99+ |
George Gilbert | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
George | PERSON | 0.99+ |
Europe | LOCATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Dinesh Nirmal | PERSON | 0.99+ |
99% | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
80% | QUANTITY | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
two | QUANTITY | 0.99+ |
HIPAA | TITLE | 0.99+ |
Dinesh | PERSON | 0.99+ |
Reynold Shin | PERSON | 0.99+ |
Friday | DATE | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Today | DATE | 0.99+ |
today | DATE | 0.99+ |
five steps | QUANTITY | 0.99+ |
45 | QUANTITY | 0.99+ |
two days | QUANTITY | 0.99+ |
11 months | QUANTITY | 0.99+ |
each segment | QUANTITY | 0.99+ |
first part | QUANTITY | 0.99+ |
two enterprises | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
first point | QUANTITY | 0.99+ |
first round | QUANTITY | 0.99+ |
each vendor | QUANTITY | 0.99+ |
Lotus | TITLE | 0.99+ |
each | QUANTITY | 0.99+ |
Azure | ORGANIZATION | 0.99+ |
two aspects | QUANTITY | 0.99+ |
one challenge | QUANTITY | 0.99+ |
one approach | QUANTITY | 0.99+ |
Spark | TITLE | 0.99+ |
two sides | QUANTITY | 0.99+ |
Cube | ORGANIZATION | 0.99+ |
one stack | QUANTITY | 0.98+ |
one component | QUANTITY | 0.98+ |
one source | QUANTITY | 0.98+ |
GDPR | TITLE | 0.98+ |
One last question | QUANTITY | 0.98+ |
one | QUANTITY | 0.98+ |
thousands of models | QUANTITY | 0.98+ |
one vendor | QUANTITY | 0.98+ |
both | QUANTITY | 0.98+ |
Kafka | TITLE | 0.97+ |
one thing | QUANTITY | 0.97+ |
Sequel | TITLE | 0.97+ |
one location | QUANTITY | 0.97+ |
second | QUANTITY | 0.96+ |
Roland Voelskow & Dinesh Nirmal - IBM Fast Track Your Data 2017
>> Narrator: Live from Munich, Germany, it's theCube, covering IBM, Fast Track Your Data. Brought to you by IBM. >> Welcome to Fast Track Your Data, everybody, welcome to Munich, Germany, this is theCube, the leader in live tech coverage, I'm Dave Vellante with my co-host Jim Kobielus. Dinesh Nirmal is here, he's the vice president of IBM Analytics Development, of course, at IBM, and he's joined by Roland Voelskow, who is the Portfolio Executive at T-Systems, which is a division of Deutche Telekom. Gentlemen, welcome to theCube, Dinesh, good to see you again. >> Thank you. Roland, let me start with you. So your role inside T-Systems, talk about that a little bit. >> Yeah, so thank you for being here, at T-Systems we serve our customers with all kinds of informal hosting services, from infrastructure up to application services, and we have recently, I'd say, about five years ago started to standardize our offerings as a product portfolio and are now focusing on coming from the infrastructure and infrastructure as a service offerings. We are now putting a strong effort in the virtualization container, virtualization to be able to move complete application landscapes from different platforms from, to T-Systems or between T-Systems platforms. The goal is to make, to enable customers to talk with us about their application needs, their business process needs, and have everything which is related to the right place to run the application will be managed automatically by our intelligent platform, which will decide in a multi-platform environment if an application, particularly a business application runs on high available private cloud or a test dev environment, for example, could run on a public cloud, so the customer should not need to deal with this kind of technology questions anymore, so we want to cover the application needs and have the rest automated. >> Yeah, we're seeing a massive trend in our community for organizations like yours to try to eliminate wherever possible undifferentiated infrastructure management, and provisioning of hardware, and Lund management and those things that really don't add value to the business trying to support their digital transformations and raise it up a little bit, and that's clearly what you just described, right? >> Roland: Exactly. >> Okay, and one of those areas that companies want to invest, of course, is data, you guys here in Munich, you chose this for a reason, but Dinesh, give us the update in what's going on in your world and what you're doing here, in Fast Track Your Data. >> Right, so actually myself and Roland was talking about this yesterday. One of the challenges our clients, customers have is the hybrid data management. So how do you make sure your data, whether it's on-premise or on the cloud, you have a seamless way to interact with that data, manage the data, govern the data, and that's the biggest challenge. I mean, lot of customers want to move to the cloud, but the critical, transactional data sits still on-prem. So that's one area that we are focusing in Munich here, is, especially with GDPR coming in 2018, how do we help our customers manage the data and govern the data all through that life cycle of the data? >> Okay, well, how do you do that? I mean, it's a multi-cloud world, most customers have, they might have some Bluemix, they might have some Amazon, they have a lot of on-prem, they got mainframe, they got all kinds of new things happening, like containers, and microservices, some are in the cloud, some are on-prem, but generally speaking, what I just described is a series of stovepipes, they each have their different lifecycle and data lifecycle and management frameworks. Is it your vision to bring all of those together in a single management framework and maybe share with us where you are on that journey and where you're going. >> Exactly, that's exactly our effort right now to bring every application service which we provide to our customers into containerized version which we can move across our platforms or which we can also transform from the external platforms from competition platforms, and onboard them into T-Systems when we acquire new customers. Is also a reality that customers work with different platforms, so we want to be the integrator, and so we would like to expand our product portfolio as an application portfolio and bring new applications, new, attractive applications into our application catalog, which is the containerized application catalog, and so here comes the part, the cooperation with IBM, so we are already a partner with IBM DB2, and we are now happy to talk about expanding the partnership into hosting the analytics portfolio of IBM, so we bring the strength of both companies together the marked excess credibility, security, in terms of European data law for T-Systems, from T-Systems, and the very attractive analytics portfolio of IBM so we can bring the best pieces together and have a very attractive offering to the market. >> So Dinesh, how does IBM fulfill that vision? Is it a product, is it a set of services, is it a framework, series of products, maybe you could describe in some more depth. >> Yeah, it all has to start with the platform. So you have the underlying platform, and then you build what you talked about, that container services on top of it, to meet the need of our enterprise customers, and then the biggest challenge is that how do you govern the data through the lifecycle of that data, right? Because that data could be sitting on-prem, data could be sitting on cloud, on a private cloud, how do you make sure that you can take that data, who touched the data, where that tech data went, and not just the data, but the analytical asset, right, so if your model's built, when was it deployed, where was it deployed? Was it deployed in QA, was it deployed in development? All those things have to be governed, so you have one governance policy, one governance console that you can go as a CDO to make sure that you can see where the data is moving and where the data is managed. So that's the biggest challenge, and that's what we are trying to make sure that, to our enterprise customers, we solve that problem. >> So IBM has announced at this show a unified governance catalog. Is that an enabler for this-- >> Dinesh: Oh, yeah. >> capability you're describing here? >> Oh yeah, I mean, that is the key piece of all of this would be the unified governance, >> Jim: Right. >> which is, you have one place to go govern that data as the CDO. >> And you've mentioned, as has Roland, the containerization of applications, now, I know that DB2 Developer Community Edition, the latest version, announced at this show, has the ability to orchestrate containerized applications, through Kubernetes, can you describe how that particular tool might be useful in this context? And how you might play DB2 Developer Community Edition in an environment where you're using the catalog to manage all the layers of data or metadata or so forth associated with these applications. >> Right, so it goes back to Dave's question, How do you manage the new products that's coming, so our goal is to make every product a container. A containerized way to deliver, so that way you have a doc or registry where you can go see what the updates are, you can update it when you're ready, all those things, but once you containerize the product and put it out there, then you can obviously have the governing infrastructures that sits on top of it to make sure all those containerized products are being managed. So that's one step towards that, but to go back to your DB2 Community Edition, our goal here is how do we simplify our product for our customers? So if you're a developer, how can we make it easy enough for you to assemble your application in matter of minutes, so that's our goal, simplify, be seamless, and be able to scale, so those are the three things we focused on the DB2 Community Edition. >> So in terms of the simplicity aspect of the tool, can you describe a few features or capabilities of the developer edition, the community edition, that are simpler than in the previous version, because I believe you've had a community edition for DB2 for developers for at least a year or two. Describe the simplifications that are introduced in this latest version. >> So one, I will give you is the JSON support. >> Okay. >> So today you want to combine the unstructured data with structured data? >> Yeah. >> I mean, it's simple, what we have a demo coming up in our main tent, where asset dialup, where you can easily go, get a JSON document put it in there, combined with your structured data, unstructured data, and you are ready to go, so that's a great example, where we are making it really easy, simple. The other example is download and go, where you can easily download in less than five clicks, less than 10 minutes, the product is up and running. So those are a couple of the things that we are doing to make sure that it is much more simpler, seamless and scalable for our customers. >> And what is Project Event Store, share with us whatever you can about that. >> Dinesh: Right. >> You're giving a demo here, I think, >> Dinesh: Yeah, yeah. >> So what is it, and why is it important? >> Yeah, so we are going to do a demo at the main tent on Project Event Store. It's about combining the strength of IBM Innovation with the power of open source. So it's about how do we do fast ingest, inserts into a object store, for example, and be able to do analytics on it. So now you have the strength of not only bringing data at very high speed or volume, but now you can do analytics on it. So for example, just to give you a very high level number we can do more than one million inserts per second. More than one million. And our closest competition is at 30,000 inserts per second. So that's huge for us. >> So use cases at the edge, obviously, could take advantage of something like this. Is that sort of where it's targeted? >> Well, yeah, so let's say, I'll give you a couple of examples. Let's say you're a hospital chain, you want the patient data coming in real time, streaming the data coming in, you want to do analytics on it, that's one example, or let's say you are a department store, you want to see all the traffic that goes into your stores and you want to do analytics on how well your campaign did on the traffic that came in. Or let's say you're an airline, right? You have IOT data that's streaming or coming in, millions of inserts per second, how do you do analytics, so this is, I would say this is a great innovation that will help all kinds of industries. >> Dinesh, I've had streaming price for quite awhile and fairly mature ones like IBM Streams, but also the structured streaming capability of Spark, and you've got a strong Spark portfolio. Is there any connection between Product Event Store and these other established IBM offerings? >> No, so what we have done is, like I said, took the power of open source, so Spark becomes obviously the execution engine, we're going to use something called the Parquet format where the data can be stored, and then we obviously have our own proprietary ingest Mechanism that brings in. So some similarity, but this is a brand new work that we have done between IBM research and it has been in the works for the last 12 to 18 months, now we are ready to bring it into the market. >> So we're about out of time, but Roland, I want to end with you and give us the perspective on Europe and European customers, particular, Rob Thomas was saying to us that part of the reason why IBM came here is because they noticed that 10 of the top companies that were out-performing the S&P 500 were US companies. And they were data-driven. And IBM kind of wanted to shake up Europe a little bit and say, "Hey guys, time to get on board." What do you see here in Europe? Obviously there are companies like Spotify which are European-based that are very data-driven, but from your perspective, what are you seeing in Europe, in terms of adoption of these data-driven technologies and to use that buzzword. >> Yes, so I think we are in an early stage of adoption of these data-driven applications and analytics, and the European companies are certainly very careful, cautious about, and sensitive about their data security. So whenever there's news about another data leakage, everyone is becoming more cautious and so here comes the unique, one of the unique positions of T-Systems, which has history and credibility in the market for data protection and uninterrupted service for our customers, so that's, we have achieved a number of cooperations, especially also with the American companies, where we do a giant approach to the European markets. So as I said, we bring the strength of T-Systems to the table, as the very competitive application portfolio, analytics portfolio, in this case, from our partner IBM, and the best worlds together for our customers. >> All right, we have to leave it there. Thank you, Roland, very much for coming on. Dinesh, great to see you again. >> Dinesh: Thank you. >> All right, you're welcome. Keep it right there, buddy. Jim and I will be back with our next guests on theCube. We're live from Munich, Germany, at Fast Track Your Data. Be right back.
SUMMARY :
Brought to you by IBM. Dinesh, good to see you again. So your role inside T-Systems, talk about that a little bit. so the customer should not need to deal is data, you guys here in Munich, So how do you make sure your data, where you are on that journey and where you're going. and so here comes the part, the cooperation with IBM, maybe you could describe in some more depth. to make sure that you can see where the data is moving So IBM has announced at this show which is, you have has the ability to orchestrate containerized applications, and be able to scale, So in terms of the simplicity aspect of the tool, So one, I will give you The other example is download and go, where you can easily whatever you can about that. So for example, just to give you a very high level number Is that sort of where it's targeted? and you want to do analytics but also the structured streaming capability of Spark, and then we obviously have our own proprietary I want to end with you and give us the perspective and so here comes the unique, one of the unique positions Dinesh, great to see you again. Jim and I will be back with our next guests on theCube.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Jim Kobielus | PERSON | 0.99+ |
Jim | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Roland Voelskow | PERSON | 0.99+ |
Roland | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Rob Thomas | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Europe | LOCATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Munich | LOCATION | 0.99+ |
Deutche Telekom | ORGANIZATION | 0.99+ |
T-Systems | ORGANIZATION | 0.99+ |
Dinesh Nirmal | PERSON | 0.99+ |
less than 10 minutes | QUANTITY | 0.99+ |
More than one million | QUANTITY | 0.99+ |
2018 | DATE | 0.99+ |
Spotify | ORGANIZATION | 0.99+ |
Dinesh | PERSON | 0.99+ |
less than five clicks | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
yesterday | DATE | 0.99+ |
DB2 Developer Community Edition | TITLE | 0.99+ |
one example | QUANTITY | 0.98+ |
DB2 | TITLE | 0.98+ |
Munich, Germany | LOCATION | 0.98+ |
both companies | QUANTITY | 0.98+ |
JSON | TITLE | 0.98+ |
2017 | DATE | 0.98+ |
Kubernetes | TITLE | 0.97+ |
Bluemix | ORGANIZATION | 0.97+ |
today | DATE | 0.97+ |
Project Event Store | ORGANIZATION | 0.96+ |
One | QUANTITY | 0.96+ |
one step | QUANTITY | 0.95+ |
single | QUANTITY | 0.95+ |
one | QUANTITY | 0.94+ |
one area | QUANTITY | 0.94+ |
IBM Analytics Development | ORGANIZATION | 0.93+ |
30,000 inserts per second | QUANTITY | 0.92+ |
three things | QUANTITY | 0.92+ |
Spark | TITLE | 0.91+ |
18 months | QUANTITY | 0.91+ |
American | OTHER | 0.91+ |
about five years ago | DATE | 0.91+ |
each | QUANTITY | 0.91+ |
IBM Innovation | ORGANIZATION | 0.9+ |
more than one million inserts per second | QUANTITY | 0.9+ |
DB2 Community Edition | TITLE | 0.9+ |
European | OTHER | 0.89+ |
a year | QUANTITY | 0.88+ |
US | LOCATION | 0.87+ |
theCube | COMMERCIAL_ITEM | 0.86+ |
Dinesh Nirmal, IBM - IBM Machine Learning Launch - #IBMML - #theCUBE
>> [Announcer] Live from New York, it's theCube, covering the IBM Machine Learning Launch Event brought to you by IBM. Now, here are your hosts, Dave Vellante and Stu Miniman. >> Welcome back to the Waldorf Astoria, everybody. This is theCube, the worldwide leader in live tech coverage. We're covering the IBM Machine Learning announcement. IBM bringing machine learning to its zMainframe, its private cloud. Dinesh Nirmel is here. He's the Vice President of Analytics at IBM and a Cube alum. Dinesh, good to see you again. >> Good to see you, Dave. >> So let's talk about ML. So we went through the big data, the data lake, the data swamp, all this stuff with the dupe. And now we're talking about machine learning and deep learning and AI and cognitive. Is it same wine, new bottle? Or is it an evolution of data and analytics? >> Good. So, Dave, let's talk about machine learning. Right. When I look at machine learning, there's three pillars. The first one is the product. I mean, you got to have a product, right. And you got to have a different shared set of functions and features available for customers to build models. For example, Canvas. I mean, those are table stakes. You got to have a set of algorithms available. So that's the product piece. >> [Dave] Uh huh. >> But then there's the process, the process of taking that model that you built in a notebook and being able to operationalize it. Meaning able to deploy it. That is, you know, I was talking to one of the customers today, and he was saying, "Machine learning is 20% fun and 80% elbow grease." Because that operationalizing of that model is not easy. Although they make it sound very simple, it's not. So if you take a banking, enterprise banking example, right? You build a model in the notebook. Some data sense build it. Now you have to take that and put it into your infrastructure or production environment, which has been there for decades. So you could have a third party software that you cannot change. You could have a set of rigid rules that already is there. You could have applications that was written in the 70's and 80's that nobody want to touch. How do you all of a sudden take the model and infuse in there? It's not easy. And so that is a tremendous amount of work. >> [Dave] Okay. >> The third pillar is the people or the expertise or the experience, the skills that needs to come through, right. So the product is one. The process of operationalizing and getting it into your production environment is another piece. And then the people is the third one. So when I look at machine learning, right. Those are three key pillars that you need to have to have a successful, you know, experience of machine learning. >> Okay, let's unpack that a little bit. Let's start with the differentiation. You mentioned Canvas, but talk about IBM specifically. >> [Dinesh] Right. What's so great about IBM? What's the differentiation? >> Right, exactly. Really good point. So we have been in the productive side for a very long time, right. I mean, it's not like we are coming into ML or AI or cognitive yesterday. We have been in that space for a very long time. We have SPSS predictive analytics available. So even if you look from all three pillars, what we are doing is we are, from a product perspective, we are bringing in the product where we are giving a choice or a flexibility to use the language you want. So there are customers who only want to use R. They are religious R users. They don't want to hear about anything else. There are customers who want to use Python, you know. They don't want to use anything else. So how do we give that choice of languages to our customers to say use any language you want. Or execution engines, right? Some folks want to use Park as execution engine. Some folks want to use R or Python, so we give that choice. Then you talked about Canvas. There are folks who want to use the GUI portion of the Canvas or a modeler to build models, or there are, you know, tekkie guys that we'll approach who want to use notebook. So how do you give that choice? So it becomes kind of like a freedom or a flexibility or a choice that we provide, so that's the product piece, right? We do that. Then the other piece is productivity. So one of the customers, the CTO of (mumbles) TV's going to come on stage with me during the main session, talk about how collaboration helped from an IBM machine learning perspective because their data scientists are sitting in New York City, our data scientists who are working with them are sitting in San Jose, California. And they were real time collaborating using notebooks in our ML projects where they can see the real time. What changes their data scientists are making. They can slack messages between each other. And that collaborative piece is what really helped us. So collaboration is one. Right from a productivity piece. We introduced something called Feedback Loop, whereby which your model can get trained. So today, you deploy a model. It could lose the score, and it could get degraded over time. Then you have to take it off-line and re-train, right? What we have done is like we introduced the Feedback Loops, so when you deploy your model, we give you two endpoints. The first endpoint is, basically, a URI, for you to plug-in your application when you, you know, run your application able call the scoring API. The second endpoint is this feedback endpoint, where you can choose to re-train the model. If you want three hours, if you want it to be six hours, you can do that. So we bring that flexibility, we bring that productivity into it. Then, the management of the models, right? How do we make sure that once you develop the model, you deploy the model. There's a life cycle involved there. How do you make sure that we enable, give you the tools to manage the model? So when you talk about differentiation, right? We are bringing differentiation on all three pillars. From a product perspective, with all the things I mentioned. From a deployment perspective. How do we make sure we have different choices of deployment, whether it's streaming, whether it's realtime, whether it's batch. You can do deployment, right? The Feedback Loop is another one. Once you deployed, how do we keep re-training it. And the last piece I talked about is the expertise or the people, right? So we are today announcing IBM Machine Learning Hub, which will become one place where our customers can go, ask questions, get education sessions, get training, right? Work together to build models. I'll give you an example, that although we are announcing hub, the IBM Machine Learning Hub today, we have been working with America First Credit Union for the last month or so. They approached us and said, you know, their underwriting takes a long time. All the knowledge is embedded in 15 to 20 human beings. And they want to make sure a machine should be able to absorb that knowledge and make that decision in minutes. So it takes hours or days. >> [Dave] So, Stu, before you jump in, so I got, put the portfolio. You know, you mentioned SPSS, expertise, choice. The collaboration, which I think you really stressed at the announcement last fall. The management of the models, so you can continuously improve it. >> Right. >> And then this knowledge base, what you're calling the hub. And I could argue, I guess, that if I take any one of those individual pieces, there, some of your competitors have them. Your argument would be it's all there. >> It all comes together, right? And you have to make sure that all three pillars come together. And customers see great value when you have that. >> Dinesh, customers today are used to kind of the deployment model on the public cloud, which is, "I want to activate a new service," you know. I just activate it, and it's there. When I think about private cloud environments, private clouds are operationally faster, but it's usually not miniature hours. It's usually more like months to deploy projects, which is still better than, you know, kind of, I think, before big data, it was, you know, oh, okay, 18 months to see if it works, and let's bring that down to, you know, a couple of months. Can you walk us through what does, you know, a customer today and says, "Great, I love this approach. "How long does it take?" You know, what's kind of the project life cycle of this? And how long will it take them to play around and pull some of these levers before they're, you know, getting productivity out of it? >> Right. So, really good questions, Stu. So let me back one step. So, in private cloud, we are going, we have new initiative called Download and Go, where our goal is to have our desktop products be able to install on your personal desktop in less than five clicks, in less than fifteen minutes. That's the goal. So the other day, you know, the team told me it's ready. That the first product is ready where you can go less than five clicks, fifteen minutes. I said the real test is I'm going to bring my son, who's five years old. Can he install it, and if he can install it, you know, we are good. And he did it. And I have a video to prove it, you know. So after the show, I will show you because and that's, when you talk about, you know, in the private cloud side, or the on-premise side, it has been a long project cycle. What we want is like you should be able to take our product, install it, and get the experience in minutes. That's the goal. And when you talk about private cloud and public cloud, another differentiating factor is that now you get the strength of IBM public cloud combined with the private cloud, so you could, you know, train your model in public cloud, and score on private cloud. You have the same experience. Not many folks, not many competitors can offer that, right? So that's another . .. >> [Stu] So if I get that right. If I as a customer have played around with the machine learning in Bluemix, I'm going to have a similar look, feel, API. >> Exactly the same, so what you have in Bluemix, right? I mean, so you have the Watson in Bluemix, which, you know, has deep learning, machine learning--all those capabilities. What we have done is we have done, is like, we have extracted the core capabilities of Watson on private cloud, and it's IBM Machine Learning. But the experience is the same. >> I want to talk about this notion of operationalizing analytics. And it ties, to me anyway, it ties into transformation. You mentioned going from Notebook to actually being able to embed analytics in workflow of the business. Can you double click on that a little bit, and maybe give some examples of how that has helped companies transform? >> Right. So when I talk about operationalizing, when you look at machine learning, right? You have all the way from data, which is the most critical piece, to building or deploying the model. A lot of times, data itself is not clean. I'll give you an example, right. So >> OSYX. >> Yeah. And when we are working with an insurance company, for example, the data that comes in. For example, if you just take gender, a lot of times the values are null. So we have to build another model to figure out if it's male or female, right? So in this case, for example, we have to say somebody has done a prostate exam. Obviously, he's a male. You know, we figured that. Or has a gynocology exam. It's a female. So we have to, you know, there's a lot of work just to get that data cleansed. So that's where I mentioned it's, you know, machine learning is 20% fun, 80% elbow grease because it's a lot of grease there that you need to make sure that you cleanse the data. Get that right. That's the shaping piece of it. Then, comes the building the model, right. And then, once you build the model on that data comes the operationalization of that model, which in itself is huge because how do you make sure that you infuse that model into your current infrastructure, which is where a lot of skill set, a lot of experience, and a lot of knowledge that comes in because you want to make sure, unless you are a start-up, right? You already have applications and programs and third-party vendors applications worth running for years, or decades, for that matter. So, yeah, so that's operationalization's a huge piece. Cleansing of the data is a huge piece. Getting the model right is another piece. >> And simplifying the whole process. I think about, I got to ingest the data. I've now got to, you know, play with it, explore. I've got to process it. And I've got to serve it to some, you know, some business need or application. And typically, those are separate processes, separate tools, maybe different personas that are doing that. Am I correct that your announcement in the Fall addressed that workflow. How is it being, you know, deployed and adopted in the field? How is it, again back to transformation, are you seeing that people are actually transforming their analytics processes and ultimately creating outcomes that they expect? >> Huge. So good point. We announced data science experience in the Fall. And the customers that who are going to speak with us today on stage, are the customers who have been using that. So, for example, if you take AFCU, America First Credit Union, they worked with us. In two weeks, you know, talk about transformation, we were able to absorb the knowledge of their underwriters. You know, what (mumbles) is in. Build that, get that features. And was able to build a model in two weeks. And the model is predicting 90%, with 90% accuracy. That's what early tests are showing. >> [Dave] And you say that was in a couple of weeks. You were, you developed that model. >> Yeah, yeah, right. So when we talk about transformation, right? We couldn't have done that a few years ago. We have transformed where the different personas can collaborate with each other, and that's a collaboration piece I talked about. Real time. Be able to build a model, and put it in the test to see what kind of benefits they're getting. >> And you've obviously got edge cases where people get really sophisticated, but, you know, we were sort of talking off camera, and you know like the 80/20 rule, or maybe it's the 90/10. You say most use cases can be, you know, solved with regression and classification. Can you talk about that a little more? >> So, so when we talk about machine learning, right? To me, I would say 90% of it is regression or classification. I mean there are edge case of our clustering and all those things. But linear regression or a classification can solve most of the, most of our customers problems, right? So whether it's fraud detection. Or whether it's underwriting the loan. Or whether you're trying to determine the sentiment analysis. I mean, you can kind of classify or do regression on it. So I would say that 90% of the cases can be covered, but like I said, most of the work is not about picking the right algorithm, but it's also about cleansing the data. Picking the algorithm, then comes building the model. Then comes deployment or operationalizing the model. So there's a step process that's involved, and each step involves some amount of work. So if I could make one more point on the technology and the transformation we have done. So even with picking the right algorithm, we automated, so you as a data scientist don't need to, you know, come in and figure out if I have 50 classifiers and each classifier has four parameters. That's 200 different combinations. Even if you take one hour on each combination, that's 200 hours or nine days that takes you to pick the right combination. What we have done is like in IBM Machine Learning we have something called cognitive assistance for data science, which will help you pick the right combination in minutes instead of days. >> So I can see how regression scales, and in the example you gave of classification, I can see how that scales. If you've got a, you know, fixed classification or maybe 200 parameters, or whatever it is, that scales, what happens, how are people dealing with, sort of automating that classification as things change, as they, some kind of new disease or pattern pops up. How do they address that at scale? >> Good point. So as the data changes, the model needs to change, right? Because everything that model knows is based on the training data. Now, if the data has changed, the symptoms of cancer or any disease has changed, obviously, you have to retrain that model. And that's where I talk about the, where the feedback loop comes in, where we will automatically retrain the model based on the new data that's coming in. So you, as an end user, for example, don't need to worry about it because we will take care of that piece also. We will automate that, also. >> Okay, good. And you've got a session this afternoon with you said two clients, right? AFCU and Kaden dot TV, and you're on, let's see, at 2:55. >> Right. >> So you folks watching the live stream, check that out. I'll give you the last word, you know, what shall we expect to hear there. Show a little leg on your discussion this afternoon. >> Right. So, obviously, I'm going to talk about the different shading factors, what we are delivering IBM Machine Learning, right? And I covered some of it. There's going to be much more. We are going to focus on how we are making freedom or flexibility available. How are we going to do productivity, right? Gains for our data scientists and developers. We are going to talk about trust, you know, the trust of data that we are bringing in. Then I'm going to bring the customers in and talk about their experience, right? We are delivering a product, but we already have customers using it, so I want them to come on stage and share the experiences of, you know, it's one thing you hear about that from us, but it's another thing that customers come and talk about it. So, and the last but not least is we are going to announce our first release of IBM Machine Learning on Z because if you look at 90% of the transactional data, today, it runs through Z, so they don't have to off-load the data to do analytics on it. We will make machine learning available, so you can do training and scoring right there on Z for your real time analytics, so. >> Right. Extending that theme that we talked about earlier, Stu, bringing analytics and transactions together, which is a big theme of the Z 13 announcement two years ago. Now you're seeing, you know, machine learning coming on Z. The live stream starts at 2 o'clock. Silicon Angle dot com had an article up on the site this morning from Maria Doucher on the IBM announcement, so check that out. Dinesh, thanks very much for coming back on theCube. Really appreciate it, and good luck today. >> Thank you. >> All right. Keep it right there, buddy. We'll be back with our next guest. This is theCube. We're live from the Waldorf Astoria for the IBM Machine Learning Event announcement. Right back.
SUMMARY :
brought to you by IBM. Dinesh, good to see you again. the data lake, the data swamp, And you got to have a different shared set So if you take a banking, to have a successful, you know, experience Let's start with the differentiation. What's the differentiation? the Feedback Loops, so when you deploy your model, The management of the models, so you can And I could argue, I guess, And customers see great value when you have that. and let's bring that down to, you know, So the other day, you know, the machine learning in Bluemix, I mean, so you have the Watson in Bluemix, Can you double click on that a little bit, when you look at machine learning, right? So we have to, you know, And I've got to serve it to some, you know, So, for example, if you take AFCU, [Dave] And you say that was in a couple of weeks. and put it in the test to see what kind You say most use cases can be, you know, we automated, so you as a data scientist and in the example you gave of classification, So as the data changes, with you said two clients, right? So you folks watching the live stream, you know, the trust of data that we are bringing in. on the IBM announcement, for the IBM Machine Learning Event announcement.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
20% | QUANTITY | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
AFCU | ORGANIZATION | 0.99+ |
15 | QUANTITY | 0.99+ |
one hour | QUANTITY | 0.99+ |
New York City | LOCATION | 0.99+ |
Dinesh Nirmal | PERSON | 0.99+ |
Dinesh Nirmel | PERSON | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
200 hours | QUANTITY | 0.99+ |
six hours | QUANTITY | 0.99+ |
90% | QUANTITY | 0.99+ |
Dave | PERSON | 0.99+ |
80% | QUANTITY | 0.99+ |
less than fifteen minutes | QUANTITY | 0.99+ |
New York | LOCATION | 0.99+ |
fifteen minutes | QUANTITY | 0.99+ |
Maria Doucher | PERSON | 0.99+ |
America First Credit Union | ORGANIZATION | 0.99+ |
50 classifiers | QUANTITY | 0.99+ |
nine days | QUANTITY | 0.99+ |
three hours | QUANTITY | 0.99+ |
two clients | QUANTITY | 0.99+ |
Kaden dot TV | ORGANIZATION | 0.99+ |
less than five clicks | QUANTITY | 0.99+ |
18 months | QUANTITY | 0.99+ |
San Jose, California | LOCATION | 0.99+ |
two weeks | QUANTITY | 0.99+ |
200 different combinations | QUANTITY | 0.99+ |
Dinesh | PERSON | 0.99+ |
each classifier | QUANTITY | 0.99+ |
200 parameters | QUANTITY | 0.99+ |
each combination | QUANTITY | 0.99+ |
Python | TITLE | 0.99+ |
today | DATE | 0.99+ |
each step | QUANTITY | 0.99+ |
two years ago | DATE | 0.99+ |
three key pillars | QUANTITY | 0.99+ |
one | QUANTITY | 0.98+ |
first product | QUANTITY | 0.98+ |
one step | QUANTITY | 0.98+ |
two endpoints | QUANTITY | 0.98+ |
third one | QUANTITY | 0.98+ |
first one | QUANTITY | 0.98+ |
Watson | TITLE | 0.98+ |
2 o'clock | DATE | 0.98+ |
last month | DATE | 0.98+ |
first endpoint | QUANTITY | 0.98+ |
three pillars | QUANTITY | 0.98+ |
Silicon Angle dot com | ORGANIZATION | 0.98+ |
70's | DATE | 0.97+ |
80's | DATE | 0.97+ |
this afternoon | DATE | 0.97+ |
Z 13 | TITLE | 0.97+ |
Z | TITLE | 0.97+ |
last fall | DATE | 0.96+ |
Bluemix | TITLE | 0.96+ |
yesterday | DATE | 0.95+ |
2:55 | DATE | 0.95+ |
Joel Horwitz, IBM | IBM CDO Summit Sping 2018
(techno music) >> Announcer: Live, from downtown San Francisco, it's theCUBE. Covering IBM Chief Data Officer Strategy Summit 2018. Brought to you by IBM. >> Welcome back to San Francisco everybody, this is theCUBE, the leader in live tech coverage. We're here at the Parc 55 in San Francisco covering the IBM CDO Strategy Summit. I'm here with Joel Horwitz who's the Vice President of Digital Partnerships & Offerings at IBM. Good to see you again Joel. >> Thanks, great to be here, thanks for having me. >> So I was just, you're very welcome- It was just, let's see, was it last month, at Think? >> Yeah, it's hard to keep track, right. >> And we were talking about your new role- >> It's been a busy year. >> the importance of partnerships. One of the things I want to, well let's talk about your role, but I really want to get into, it's innovation. And we talked about this at Think, because it's so critical, in my opinion anyway, that you can attract partnerships, innovation partnerships, startups, established companies, et cetera. >> Joel: Yeah. >> To really help drive that innovation, it takes a team of people, IBM can't do it on its own. >> Yeah, I mean look, IBM is the leader in innovation, as we all know. We're the market leader for patents, that we put out each year, and how you get that technology in the hands of the real innovators, the developers, the longtail ISVs, our partners out there, that's the challenging part at times, and so what we've been up to is really looking at how we make it easier for partners to partner with IBM. How we make it easier for developers to work with IBM. So we have a number of areas that we've been adding, so for example, we've added a whole IBM Code portal, so if you go to developer.ibm.com/code you can actually see hundreds of code patterns that we've created to help really any client, any partner, get started using IBM's technology, and to innovate. >> Yeah, and that's critical, I mean you're right, because to me innovation is a combination of invention, which is what you guys do really, and then it's adoption, which is what your customers are all about. You come from the data science world. We're here at the Chief Data Officer Summit, what's the intersection between data science and CDOs? What are you seeing there? >> Yeah, so when I was here last, it was about two years ago in 2015, actually, maybe three years ago, man, time flies when you're having fun. >> Dave: Yeah, the Spark Summit- >> Yeah Spark Technology Center and the Spark Summit, and we were here, I was here at the Chief Data Officer Summit. And it was great, and at that time, I think a lot of the conversation was really not that different than what I'm seeing today. Which is, how do you manage all of your data assets? I think a big part of doing good data science, which is my kind of background, is really having a good understanding of what your data governance is, what your data catalog is, so, you know we introduced the Watson Studio at Think, and actually, what's nice about that, is it brings a lot of this together. So if you look in the market, in the data market, today, you know we used to segment it by a few things, like data gravity, data movement, data science, and data governance. And those are kind of the four themes that I continue to see. And so outside of IBM, I would contend that those are relatively separate kind of tools that are disconnected, in fact Dinesh Nirmal, who's our engineer on the analytic side, Head of Development there, he wrote a great blog just recently, about how you can have some great machine learning, you have some great data, but if you can't operationalize that, then really you can't put it to use. And so it's funny to me because we've been focused on this challenge, and IBM is making the right steps, in my, I'm obviously biased, but we're making some great strides toward unifying the, this tool chain. Which is data management, to data science, to operationalizing, you know, machine learning. So that's what we're starting to see with Watson Studio. >> Well, I always push Dinesh on this and like okay, you've got a collection of tools, but are you bringing those together? And he flat-out says no, we developed this, a lot of this from scratch. Yes, we bring in the best of the knowledge that we have there, but we're not trying to just cobble together a bunch of disparate tools with a UI layer. >> Right, right. >> It's really a fundamental foundation that you're trying to build. >> Well, what's really interesting about that, that piece, is that yeah, I think a lot of folks have cobbled together a UI layer, so we formed a partnership, coming back to the partnership view, with a company called Lightbend, who's based here in San Francisco, as well as in Europe, and the reason why we did that, wasn't just because of the fact that Reactive development, if you're not familiar with Reactive, it's essentially Scala, Akka, Play, this whole framework, that basically allows developers to write once, and it kind of scales up with demand. In fact, Verizon actually used our platform with Lightbend to launch the iPhone 10. And they show dramatic improvements. Now what's exciting about Lightbend, is the fact that application developers are developing with Reactive, but if you turn around, you'll also now be able to operationalize models with Reactive as well. Because it's basically a single platform to move between these two worlds. So what we've continued to see is data science kind of separate from the application world. Really kind of, AI and cloud as different universes. The reality is that for any enterprise, or any company, to really innovate, you have to find a way to bring those two worlds together, to get the most use out of it. >> Fourier always says "Data is the new development kit". He said this I think five or six years ago, and it's barely becoming true. You guys have tried to make an attempt, and have done a pretty good job, of trying to bring those worlds together in a single platform, what do you call it? The Watson Data Platform? >> Yeah, Watson Data Platform, now Watson Studio, and I think the other, so one side of it is, us trying to, not really trying, but us actually bringing together these disparate systems. I mean we are kind of a systems company, we're IT. But not only that, but bringing our trained algorithms, and our trained models to the developers. So for example, we also did a partnership with Unity, at the end of last year, that's now just reaching some pretty good growth, in terms of bringing the Watson SDK to game developers on the Unity platform. So again, it's this idea of bringing the game developer, the application developer, in closer contact with these trained models, and these trained algorithms. And that's where you're seeing incredible things happen. So for example, Star Trek Bridge Crew, which I don't know how many Trekkies we have here at the CDO Summit. >> A few over here probably. >> Yeah, a couple? They're using our SDK in Unity, to basically allow a gamer to use voice commands through the headset, through a VR headset, to talk to other players in the virtual game. So we're going to see more, I can't really disclose too much what we're doing there, but there's some cool stuff coming out of that partnership. >> Real immersive experience driving a lot of data. Now you're part of the Digital Business Group. I like the term digital business, because we talk about it all the time. Digital business, what's the difference between a digital business and a business? What's the, how they use data. >> Joel: Yeah. >> You're a data person, what does that mean? That you're part of the Digital Business Group? Is that an internal facing thing? An external facing thing? Both? >> It's really both. So our Chief Digital Officer, Bob Lord, he has a presentation that he'll give, where he starts out, and he goes, when I tell people I'm the Chief Digital Officer they usually think I just manage the website. You know, if I tell people I'm a Chief Data Officer, it means I manage our data, in governance over here. The reality is that I think these Chief Digital Officer, Chief Data Officer, they're really responsible for business transformation. And so, if you actually look at what we're doing, I think on both sides is we're using data, we're using marketing technology, martech, like Optimizely, like Segment, like some of these great partners of ours, to really look at how we can quickly A/B test, get user feedback, to look at how we actually test different offerings and market. And so really what we're doing is we're setting up a testing platform, to bring not only our traditional offers to market, like DB2, Mainframe, et cetera, but also bring new offers to market, like blockchain, and quantum, and others, and actually figure out how we get better product-market fit. What actually, one thing, one story that comes to mind, is if you've seen the movie Hidden Figures- >> Oh yeah. >> There's this scene where Kevin Costner, I know this is going to look not great for IBM, but I'm going to say it anyways, which is Kevin Costner has like a sledgehammer, and he's like trying to break down the wall to get the mainframe in the room. That's what it feels like sometimes, 'cause we create the best technology, but we forget sometimes about the last mile. You know like, we got to break down the wall. >> Where am I going to put it? >> You know, to get it in the room! So, honestly I think that's a lot of what we're doing. We're bridging that last mile, between these different audiences. So between developers, between ISVs, between commercial buyers. Like how do we actually make this technology, not just accessible to large enterprise, which are our main clients, but also to the other ecosystems, and other audiences out there. >> Well so that's interesting Joel, because as a potential partner of IBM, they want, obviously your go-to-market, your massive company, and great distribution channel. But at the same time, you want more than that. You know you want to have a closer, IBM always focuses on partnerships that have intrinsic value. So you talked about offerings, you talked about quantum, blockchain, off-camera talking about cloud containers. >> Joel: Yeah. >> I'd say cloud and containers may be a little closer than those others, but those others are going to take a lot of market development. So what are the offerings that you guys are bringing? How do they get into the hands of your partners? >> I mean, the commonality with all of these, all the emerging offerings, if you ask me, is the distributed nature of the offering. So if you look at blockchain, it's a distributed ledger. It's a distributed transaction chain that's secure. If you look at data, really and we can hark back to say, Hadoop, right before object storage, it's distributed storage, so it's not just storing on your hard drive locally, it's storing on a distributed network of servers that are all over the world and data centers. If you look at cloud, and containers, what you're really doing is not running your application on an individual server that can go down. You're using containers because you want to distribute that application over a large network of servers, so that if one server goes down, you're not going to be hosed. And so I think the fundamental shift that you're seeing is this distributed nature, which in essence is cloud. So I think cloud is just kind of a synonym, in my opinion, for distributed nature of our business. >> That's interesting and that brings up, you're right, cloud and Big Data/Hadoop, we don't talk about Hadoop much anymore, but it kind of got it all started, with that notion of leave the data where it is. And it's the same thing with cloud. You can't just stuff your business into the public cloud. You got to bring the cloud to your data. >> Joel: That's right. >> But that brings up a whole new set of challenges, which obviously, you're in a position just to help solve. Performance, latency, physics come into play. >> Physics is a rough one. It's kind of hard to avoid that one. >> I hear your best people are working on it though. Some other partnerships that you want to sort of, elucidate. >> Yeah, no, I mean we have some really great, so I think the key kind of partnership, I would say area, that I would allude to is, one of the things, and you kind of referenced this, is a lot of our partners, big or small, want to work with our top clients. So they want to work with our top banking clients. They want, 'cause these are, if you look at for example, MaRisk and what we're doing with them around blockchain, and frankly, talk about innovation, they're innovating containers for real, not virtual containers- >> And that's a joint venture right? >> Yeah, it is, and so it's exciting because, what we're bringing to market is, I also lead our startup programs, called the Global Entrepreneurship Program, and so what I'm focused on doing, and you'll probably see more to come this quarter, is how do we actually bridge that end-to-end? How do you, if you're startup or a small business, ultimately reach that kind of global business partner level? And so kind of bridging that, that end-to-end. So we're starting to bring out a number of different incentives for partners, like co-marketing, so I'll help startups when they're early, figure out product-market fit. We'll give you free credits to use our innovative technology, and we'll also bring you into a number of clients, to basically help you not burn all of your cash on creating your own marketing channel. God knows I did that when I was at a start-up. So I think we're doing a lot to kind of bridge that end-to-end, and help any partner kind of come in, and then grow with IBM. I think that's where we're headed. >> I think that's a critical part of your job. Because I mean, obviously IBM is known for its Global 2000, big enterprise presence, but startups, again, fuel that innovation fire. So being able to attract them, which you're proving you can, providing whatever it is, access, early access to cloud services, or like you say, these other offerings that you're producing, in addition to that go-to-market, 'cause it's funny, we always talk about how efficient, capital efficient, software is, but then you have these companies raising hundreds of millions of dollars, why? Because they got to do promotion, marketing, sales, you know, go-to-market. >> Yeah, it's really expensive. I mean, you look at most startups, like their biggest ticket item is usually marketing and sales. And building channels, and so yeah, if you're, you know we're talking to a number of partners who want to work with us because of the fact that, it's not just like, the direct kind of channel, it's also, as you kind of mentioned, there's other challenges that you have to overcome when you're working with a larger company. for example, security is a big one, GDPR compliance now, is a big one, and just making sure that things don't fall over, is a big one. And so a lot of partners work with us because ultimately, a number of the decision makers in these larger enterprises are going, well, I trust IBM, and if IBM says you're good, then I believe you. And so that's where we're kind of starting to pull partners in, and pull an ecosystem towards us. Because of the fact that we can take them through that level of certification. So we have a number of free online courses. So if you go to partners, excuse me, ibm.com/partners/learn there's a number of blockchain courses that you can learn today, and will actually give you a digital certificate, that's actually certified on our own blockchain, which we're actually a first of a kind to do that, which I think is pretty slick, and it's accredited at some of the universities. So I think that's where people are looking to IBM, and other leaders in this industry, is to help them become experts in their, in this technology, and especially in this emerging technology. >> I love that blockchain actually, because it's such a growing, and interesting, and innovative field. But it needs players like IBM, that can bring credibility, enterprise-grade, whether it's security, or just, as I say, credibility. 'Cause you know, this is, so much of negative connotations associated with blockchain and crypto, but companies like IBM coming to the table, enterprise companies, and building that ecosystem out is in my view, crucial. >> Yeah, no, it takes a village. I mean, there's a lot of folks, I mean that's a big reason why I came to IBM, three, four years ago, was because when I was in start-up land, I used to work for H20, I worked for Alpine Data Labs, Datameer, back in the Hadoop days, and what I realized was that, it's an opportunity cost. So you can't really drive true global innovation, transformation, in some of these bigger companies because there's only so much that you can really kind of bite off. And so you know at IBM it's been a really rewarding experience because we have done things like for example, we partnered with Girls Who Code, Treehouse, Udacity. So there's a number of early educators that we've partnered with, to bring code to, to bring technology to, that frankly, would never have access to some of this stuff. Some of this technology, if we didn't form these alliances, and if we didn't join these partnerships. So I'm very excited about the future of IBM, and I'm very excited about the future of what our partners are doing with IBM, because, geez, you know the cloud, and everything that we're doing to make this accessible, is bar none, I mean, it's great. >> I can tell you're excited. You know, spring in your step. Always a lot of energy Joel, really appreciate you coming onto theCUBE. >> Joel: My pleasure. >> Great to see you again. >> Yeah, thanks Dave. >> You're welcome. Alright keep it right there, everybody. We'll be back. We're at the IBM CDO Strategy Summit in San Francisco. You're watching theCUBE. (techno music) (touch-tone phone beeps)
SUMMARY :
Brought to you by IBM. Good to see you again Joel. that you can attract partnerships, To really help drive that innovation, and how you get that technology Yeah, and that's critical, I mean you're right, Yeah, so when I was here last, to operationalizing, you know, machine learning. that we have there, but we're not trying that you're trying to build. to really innovate, you have to find a way in a single platform, what do you call it? So for example, we also did a partnership with Unity, to basically allow a gamer to use voice commands I like the term digital business, to look at how we actually test different I know this is going to look not great for IBM, but also to the other ecosystems, But at the same time, you want more than that. So what are the offerings that you guys are bringing? So if you look at blockchain, it's a distributed ledger. You got to bring the cloud to your data. But that brings up a whole new set of challenges, It's kind of hard to avoid that one. Some other partnerships that you want to sort of, elucidate. and you kind of referenced this, to basically help you not burn all of your cash early access to cloud services, or like you say, that you can learn today, but companies like IBM coming to the table, that you can really kind of bite off. really appreciate you coming onto theCUBE. We're at the IBM CDO Strategy Summit in San Francisco.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Joel | PERSON | 0.99+ |
Joel Horwitz | PERSON | 0.99+ |
Europe | LOCATION | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Kevin Costner | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Dinesh Nirmal | PERSON | 0.99+ |
Alpine Data Labs | ORGANIZATION | 0.99+ |
Lightbend | ORGANIZATION | 0.99+ |
Verizon | ORGANIZATION | 0.99+ |
San Francisco | LOCATION | 0.99+ |
Hidden Figures | TITLE | 0.99+ |
Bob Lord | PERSON | 0.99+ |
Both | QUANTITY | 0.99+ |
MaRisk | ORGANIZATION | 0.99+ |
both | QUANTITY | 0.99+ |
iPhone 10 | COMMERCIAL_ITEM | 0.99+ |
2015 | DATE | 0.99+ |
Datameer | ORGANIZATION | 0.99+ |
both sides | QUANTITY | 0.99+ |
one story | QUANTITY | 0.99+ |
Think | ORGANIZATION | 0.99+ |
five | DATE | 0.99+ |
hundreds | QUANTITY | 0.99+ |
Treehouse | ORGANIZATION | 0.99+ |
three years ago | DATE | 0.99+ |
developer.ibm.com/code | OTHER | 0.99+ |
Unity | ORGANIZATION | 0.98+ |
two worlds | QUANTITY | 0.98+ |
Reactive | ORGANIZATION | 0.98+ |
GDPR | TITLE | 0.98+ |
one side | QUANTITY | 0.98+ |
Digital Business Group | ORGANIZATION | 0.98+ |
today | DATE | 0.98+ |
Udacity | ORGANIZATION | 0.98+ |
ibm.com/partners/learn | OTHER | 0.98+ |
last month | DATE | 0.98+ |
Watson Studio | ORGANIZATION | 0.98+ |
each year | QUANTITY | 0.97+ |
three | DATE | 0.97+ |
single platform | QUANTITY | 0.97+ |
Girls Who Code | ORGANIZATION | 0.97+ |
Parc 55 | LOCATION | 0.97+ |
one thing | QUANTITY | 0.97+ |
four themes | QUANTITY | 0.97+ |
Spark Technology Center | ORGANIZATION | 0.97+ |
six years ago | DATE | 0.97+ |
H20 | ORGANIZATION | 0.97+ |
four years ago | DATE | 0.97+ |
martech | ORGANIZATION | 0.97+ |
Unity | TITLE | 0.96+ |
hundreds of millions of dollars | QUANTITY | 0.94+ |
Watson Studio | TITLE | 0.94+ |
Dinesh | PERSON | 0.93+ |
one server | QUANTITY | 0.93+ |
Rob Thomas, IBM | Machine Learning Everywhere 2018
>> Announcer: Live from New York, it's theCUBE, covering Machine Learning Everywhere: Build Your Ladder to AI, brought to you by IBM. >> Welcome back to New York City. theCUBE continue our coverage here at IBM's event, Machine Learning Everywhere: Build Your Ladder to AI. And with us now is Rob Thomas, who is the vice president of, or general manager, rather, of IBM analytics. Sorry about that, Rob. Good to have you with us this morning. Good to see you, sir. >> Great to see you John. Dave, great to see you as well. >> Great to see you. >> Well let's just talk about the event first. Great lineup of guests. We're looking forward to visiting with several of them here on theCUBE today. But let's talk about, first off, general theme with what you're trying to communicate and where you sit in terms of that ladder to success in the AI world. >> So, maybe start by stepping back to, we saw you guys a few times last year. Once in Munich, I recall, another one in New York, and the theme of both of those events was, data science renaissance. We started to see data science picking up steam in organizations. We also talked about machine learning. The great news is that, in that timeframe, machine learning has really become a real thing in terms of actually being implemented into organizations, and changing how companies run. And that's what today is about, is basically showcasing a bunch of examples, not only from our clients, but also from within IBM, how we're using machine learning to run our own business. And the thing I always remind clients when I talk to them is, machine learning is not going to replace managers, but I think machine learning, managers that use machine learning will replace managers that do not. And what you see today is a bunch of examples of how that's true because it gives you superpowers. If you've automated a lot of the insight, data collection, decision making, it makes you a more powerful manager, and that's going to change a lot of enterprises. >> It seems like a no-brainer, right? I mean, or a must-have. >> I think there's a, there's always that, sometimes there's a fear factor. There is a culture piece that holds people back. We're trying to make it really simple in terms of how we talk about the day, and the examples that we show, to get people comfortable, to kind of take a step onto that ladder back to the company. >> It's conceptually a no-brainer, but it's a challenge. You wrote a blog and it was really interesting. It was, one of the clients said to you, "I'm so glad I'm not in the technology industry." And you went, "Uh, hello?" (laughs) "I've got news for you, you are in the technology industry." So a lot of customers that I talk to feel like, meh, you know, in our industry, it's really not getting disrupted. That's kind of taxis and retail. We're in banking and, you know, but, digital is disrupting every industry and every industry is going to have to adopt ML, AI, whatever you want to call it. Can traditional companies close that gap? What's your take? >> I think they can, but, I'll go back to the word I used before, it starts with culture. Am I accepting that I'm a technology company, even if traditionally I've made tractors, as an example? Or if traditionally I've just been you know, selling shirts and shoes, have I embraced the role, my role as a technology company? Because if you set that culture from the top, everything else flows from there. It can't be, IT is something that we do on the side. It has to be a culture of, it's fundamental to what we do as a company. There was an MIT study that said, data-driven cultures drive productivity gains of six to 10 percent better than their competition. You can't, that stuff compounds, too. So if your competitors are doing that and you're not, not only do you fall behind in the short term but you fall woefully behind in the medium term. And so, I think companies are starting to get there but it takes a constant push to get them focused on that. >> So if you're a tractor company, you've got human expertise around making tractors and messaging and marketing tractors, and then, and data is kind of there, sort of a bolt-on, because everybody's got to be data-driven, but if you look at the top companies by market cap, you know, we were talking about it earlier. Data is foundational. It's at their core, so, that seems to me to be the hard part, Rob, I'd like you to comment in terms of that cultural shift. How do you go from sort of data in silos and, you know, not having cloud economics and, that are fundamental, to having that dynamic, and how does IBM help? >> You know, I think, to give companies credit, I think most organizations have developed some type of data practice or discipline over the last, call it five years. But most of that's historical, meaning, yeah, we'll take snapshots of history. We'll use that to guide decision making. You fast-forward to what we're talking about today, just so we're on the same page, machine learning is about, you build a model, you train a model with data, and then as new data flows in, your model is constantly updating. So your ability to make decisions improves over time. That's very different from, we're doing historical reporting on data. And so I think it's encouraging that companies have kind of embraced that data discipline in the last five years, but what we're talking about today is a big next step and what we're trying to break it down to what I call the building blocks, so, back to the point on an AI ladder, what I mean by an AI ladder is, you can't do AI without machine learning. You can't do machine learning without analytics. You can't do analytics without the right data architecture. So those become the building blocks of how you get towards a future of AI. And so what I encourage companies is, if you're not ready for that AI leading edge use case, that's okay, but you can be preparing for that future now. That's what the building blocks are about. >> You know, I think we're, I know we're ahead of, you know, Jeremiah Owyang on a little bit later, but I was reading something that he had written about gut and instinct, from the C-Suite, and how, that's how companies were run, right? You had your CEO, your president, they made decisions based on their guts or their instincts. And now, you've got this whole new objective tool out there that's gold, and it's kind of taking some of the gut and instinct out of it, in a way, and maybe there are people who still can't quite grasp that, that maybe their guts and their instincts, you know, what their gut tells them, you know, is one thing, but there's pretty objective data that might indicate something else. >> Moneyball for business. >> A little bit of a clash, I mean, is there a little bit of a clash in that respect? >> I think you'd be surprise by how much decision making is still pure opinion. I mean, I see that everywhere. But we're heading more towards what you described for sure. One of the clients talking here today, AMC Networks, think it's a great example of a company that you wouldn't think of as a technology company, primarily a content producer, they make great shows, but they've kind of gone that extra step to say, we can integrate data sources from third parties, our own data about viewer habits, we can do that to change our relationship with advertisers. Like, that's a company that's really embraced this idea of being a technology company, and you can see it in their results, and so, results are not coincidence in this world anymore. It's about a practice applied to data, leveraging machine learning, on a path towards AI. If companies are doing that, they're going to be successful. >> And we're going to have the tally from AMC on, but so there's a situation where they have embraced it, that they've dealt with that culture, and data has become foundational. Now, I'm interested as to what their journey look like. What are you seeing with clients? How they break this down, the silos of data that have been built up over decades. >> I think, so they get almost like a maturity curve. You've got, and the rule I talk about is 40-40-20, where 40% of organizations are really using data just to optimize costs right now. That's okay, but that's on the lower end of the maturity curve. 40% are saying, all right, I'm starting to get into data science. I'm starting to think about how I extend to new products, new services, using data. And then 20% are on the leading edge. And that's where I'd put AMC Networks, by the way, because they've done unique things with integrating data sets and building models so that they've automated a lot of what used to be painstakingly long processes, internal processes to do it. So you've got this 40-40-20 of organizations in terms of their maturity on this. If you're not on that curve right now, you have a problem. But I'd say most are somewhere on that curve. If you're in the first 40% and you're, right now data for you is just about optimizing cost, you're going to be behind. If you're not right now, you're going to be behind in the next year, that's a problem. So I'd kind of encourage people to think about what it takes to be in the next 40%. Ultimately you want to be in the 20% that's actually leading this transformation. >> So change it to 40-20-40. That's where you want it to go, right? You want to flip that paradigm. >> I want to ask you a question. You've done a lot of M and A in the past. You spent a lot of time in Silicon Valley and Silicon Valley obviously very, very disruptive, you know, cultures and organizations and it's always been a sort of technology disruption. It seems like there's a ... another disruption going on, not just horizontal technologies, you know, cloud or mobile or social, whatever it is, but within industries. Some industries, as we've been talking, radically disrupted. Retail, taxis, certainly advertising, et cetera et cetera. Some have not yet, the client that you talked to. Do you see, technology companies generally, Silicon Valley companies specifically, as being able to pull off a sort of disruption of not only technologies but also industries and where does IBM play there? You've made a sort of, Ginni in particular has made a deal about, hey, we're not going to compete with our customers. So talking about this sort of dual disruption agenda, one on the technology side, one within industries that Apple's getting into financial services and, you know, Amazon getting into grocery, what's your take on that and where does IBM fit in that world? >> So, I mean, IBM has been in Silicon Valley for a long time, I would say probably longer than 99.9% of the companies in Silicon Valley, so, we've got a big lab there. We do a lot of innovation out of there. So love it, I mean, the culture of the valley is great for the world because it's all about being the challenger, it's about innovation, and that's tremendous. >> No fear. >> Yeah, absolutely. So, look, we work with a lot of different partners, some who are, you know, purely based in the valley. I think they challenge us. We can learn from them, and that's great. I think the one, the one misnomer that I see right now, is there's a undertone that innovation is happening in Silicon Valley and only in Silicon Valley. And I think that's a myth. Give you an example, we just, in December, we released something called Event Store which is basically our stab at reinventing the database business that's been pretty much the same for the last 30 to 40 years. And we're now ingesting millions of rows of data a second. We're doing it in a Parquet format using a Spark engine. Like, this is an amazing innovation that will change how any type of IOT use case can manage data. Now ... people don't think of IBM when they think about innovations like that because it's not the only thing we talk about. We don't have, the IBM website isn't dedicated to that single product because IBM is a much bigger company than that. But we're innovating like crazy. A lot of that is out of what we're doing in Silicon Valley and our labs around the world and so, I'm very optimistic on what we're doing in terms of innovation. >> Yeah, in fact, I think, rephrase my question. I was, you know, you're right. I mean people think of IBM as getting disrupted. I wasn't posing it, I think of you as a disruptor. I know that may sound weird to some people but in the sense that you guys made some huge bets with things like Watson on solving some of the biggest, world's problems. And so I see you as disrupting sort of, maybe yourselves. Okay, frame that. But I don't see IBM as saying, okay, we are going to now disrupt healthcare, disrupt financial services, rather we are going to help our, like some of your comp... I don't know if you'd call them competitors. Amazon, as they say, getting into content and buying grocery, you know, food stores. You guys seems to have a different philosophy. That's what I'm trying to get to is, we're going to disrupt ourselves, okay, fine. But we're not going to go hard into healthcare, hard into financial services, other than selling technology and services to those organizations, does that make sense? >> Yeah, I mean, look, our mission is to make our clients ... better at what they do. That's our mission, we want to be essential in terms of their journey to be successful in their industry. So frankly, I love it every time I see an announcement about Amazon entering another vertical space, because all of those companies just became my clients. Because they're not going to work with Amazon when they're competing with them head to head, day in, day out, so I love that. So us working with these companies to make them better through things like Watson Health, what we're doing in healthcare, it's about making companies who have built their business in healthcare, more effective at how they perform, how they drive results, revenue, ROI for their investors. That's what we do, that's what IBM has always done. >> Yeah, so it's an interesting discussion. I mean, I tend to agree. I think Silicon Valley maybe should focus on those technology disruptions. I think that they'll have a hard time pulling off that dual disruption and maybe if you broadly define Silicon Valley as Seattle and so forth, but, but it seems like that formula has worked for decades, and will continue to work. Other thoughts on sort of the progression of ML, how it gets into organizations. You know, where you see this going, again, I was saying earlier, the parlance is changing. Big data is kind of, you know, mm. Okay, Hadoop, well, that's fine. We seem to be entering this new world that's pervasive, it's embedded, it's intelligent, it's autonomous, it's self-healing, it's all these things that, you know, we aspire to. We're now back in the early innings. We're late innings of big data, that's kind of ... But early innings of this new era, what are your thoughts on that? >> You know, I'd say the biggest restriction right now I see, we talked before about somehow, sometimes companies don't have the desire, so we have to help create the desire, create the culture to go do this. Even for the companies that have a burning desire, the issue quickly becomes a skill gap. And so we're doing a lot to try to help bridge that skill gap. Let's take data science as an example. There's two worlds of data science that I would describe. There's clickers, and there's coders. Clickers want to do drag and drop. They will use traditional tools like SPSS, which we're modernizing, that's great. We want to support them if that's how they want to work and build models and deploy models. There's also this world of coders. This is people that want to do all their data science in ML, and Python, and Scala, and R, like, that's what they want to do. And so we're supporting them through things like Data Science Experience, which is built on Apache Jupiter. It's all open source tooling, it'd designed for coders. The reason I think that's important, it goes back to the point on skill sets. There is a skill gap in most companies. So if you walk in and you say, this is the only way to do this thing, you kind of excluded half the companies because they say, I can't play in that world. So we are intentionally going after a strategy that says, there's a segmentation in skill types. In places there's a gap, we can help you fill that gap. That's how we're thinking about them. >> And who does that bode well for? If you say that you were trying to close a gap, does that bode well for, we talked about the Millennial crowd coming in and so they, you know, do they have a different approach or different mental outlook on this, or is it to the mid-range employee, you know, who is open minded, I mean, but, who is the net sweet spot, you think, that say, oh, this is a great opportunity right now? >> So just take data science as an example. The clicker coder comment I made, I would put the clicker audience as mostly people that are 20 years into their career. They've been around a while. The coder audience is all the Millennials. It's all the new audience. I think the greatest beneficiary is the people that find themselves kind of stuck in the middle, which is they're kind of interested in this ... >> That straddle both sides of the line yeah? >> But they've got the skill set and the desire to do some of the new tooling and new approaches. So I think this kind of creates an opportunity for that group in the middle to say, you know, what am I going to adopt as a platform for how I go forward and how I provide leadership in my company? >> So your advice, then, as you're talking to your clients, I mean you're also talking to their workforce. In a sense, then, your advice to them is, you know, join, jump in the wave, right? You've got your, you can't straddle, you've got to go. >> And you've got to experiment, you've got to try things. Ultimately, organizations are going to gravitate to things that they like using in terms of an approach or a methodology or a tool. But that comes with experimentation, so people need to get out there and try something. >> Maybe we could talk about developers a little bit. We were talking to Dinesh earlier and you guys of course have focused on data scientists, data engineers, obviously developers. And Dinesh was saying, look, many, if not most, of the 10 million Java developers out there, they're not, like, focused around the data. That's really the data scientist's job. But then, my colleague John Furrier says, hey, data is the new development kit. You know, somebody said recently, you know, Andreessen's comment, "software is eating the world." Well, data is eating software. So if Furrier is right and that comment is right, it seems like developers increasingly have to become more data aware, fundamentally. Blockchain developers clearly are more data focused. What's your take on the developer community, where they fit into this whole AI, machine learning space? >> I was just in Las Vegas yesterday and I did a session with a bunch of our business partners. ISVs, so software companies, mostly a developer audience, and the discussion I had with them was around, you're doing, you're building great products, you're building great applications. But your product is only as good as the data and the intelligence that you embed in your product. Because you're still putting too much of a burden on the user, as opposed to having everything happen magically, if you will. So that discussion was around, how do you embed data, embed AI, into your products and do that at the forefront versus, you deliver a product and the client has to say, all right, now I need to get my data out of this application and move it somewhere else so I can do the data science that I want to do. That's what I see happening with developers. It's kind of ... getting them to think about data as opposed to just thinking about the application development framework, because that's where most of them tend to focus. >> Mm, right. >> Well, we've talked about, well, earlier on about the governance, so just curious, with Madhu, which I'll, we'll have that interview in just a little bit here. I'm kind of curious about your take on that, is that it's a little kinder, gentler, friendlier than maybe some might look at it nowadays because of some organization that it causes, within your group and some value that's being derived from that, that more efficiency, more contextual information that's, you know, more relevant, whatever. When you talk to your clients about meeting rules, regs, GDPR, all these things, how do you get them to see that it's not a black veil of doom and gloom but it really is, really more of an opportunity for them to cash in? >> You know, my favorite question to ask when I go visit clients is I say, I say, just show of hands, how many people have all the data they need to do their job? To date, nobody has ever raised their hand. >> Not too many hands up. >> The reason I phrased it that way is, that's fundamentally a governance challenge. And so, when you think about governance, I think everybody immediately thinks about compliance, GDPR, types of things you mentioned, and that's great. But there's two use cases for governance. One is compliance, the other one is self service analytics. Because if you've done data governance, then you can make your data available to everybody in the organization because you know you've got the right rules, the right permissions set up. That will change how people do their jobs and I think sometimes governance gets painted into a compliance corner, when organizations need to think about it as, this is about making data accessible to my entire workforce. That's a big change. I don't think anybody has that today. Except for the clients that we're working with, where I think we've made good strides in that. >> What's your sort of number one, two, and three, or pick one, advice for those companies that as you blogged about, don't realize yet that they're in the software business and the technology business? For them to close the ... machine intelligence, machine learning, AI gap, where should they start? >> I do think it can be basic steps. And the reason I say that is, if you go to a company that hasn't really viewed themselves as a technology company, and you start talking about machine intelligence, AI, like, everybody like, runs away scared, like it's not interesting. So I bring it back to building blocks. For a client to be great in data, and to become a technology company, you really need three platforms for how you think about data. You need a platform for how you manage your data, so think of it as data management. You need a platform for unified governance and integration, and you need a platform for data science and business analytics. And to some extent, I don't care where you start, but you've got to start with one of those. And if you do that, you know, you'll start to create a flywheel of momentum where you'll get some small successes. Then you can go in the other area, and so I just encourage everybody, start down that path. Pick one of the three. Or you may already have something going in one of them, so then pick one where you don't have something going. Just start down the path, because, those building blocks, once you have those in place, you'll be able to scale AI and ML in the future in your organization. But without that, you're going to always be limited to kind of a use case at a time. >> Yeah, and I would add, this is, you talked about it a couple times today, is that cultural aspect, that realization that in order to be data driven, you know, buzzword, you have to embrace that and drive that through the culture. Right? >> That starts at the top, right? Which is, it's not, you know, it's not normal to have a culture of, we're going to experiment, we're going to try things, half of them may not work. And so, it starts at the top in terms of how you set the tone and set that culture. >> IBM Think, we're less than a month away. CUBE is going to be there, very excited about that. First time that you guys have done Think. You've consolidated all your big, big events. What can we expect from you guys? >> I think it's going to be an amazing show. To your point, we thought about this for a while, consolidating to a single IBM event. There's no question just based on the response and the enrollment we have so far, that was the right answer. We'll have people from all over the world. A bunch of clients, we've got some great announcements that will come out that week. And for clients that are thinking about coming, honestly the best thing about it is all the education and training. We basically build a curriculum, and think of it as a curriculum around, how do we make our clients more effective at competing with the Amazons of the world, back to the other point. And so I think we build a great curriculum and it will be a great week. >> Well, if I've heard anything today, it's about, don't be afraid to dive in at the deep end, just dive, right? Get after it and, looking forward to the rest of the day. Rob, thank you for joining us here and we'll see you in about a month! >> Sounds great. >> Right around the corner. >> All right, Rob Thomas joining us here from IBM Analytics, the GM at IBM Analytics. Back with more here on theCUBE. (upbeat music)
SUMMARY :
Build Your Ladder to AI, brought to you by IBM. Good to have you with us this morning. Dave, great to see you as well. and where you sit in terms of that ladder And what you see today is a bunch of examples I mean, or a must-have. onto that ladder back to the company. So a lot of customers that I talk to And so, I think companies are starting to get there to be the hard part, Rob, I'd like you to comment You fast-forward to what we're talking about today, and it's kind of taking some of the gut But we're heading more towards what you described for sure. Now, I'm interested as to what their journey look like. to think about what it takes to be in the next 40%. That's where you want it to go, right? I want to ask you a question. So love it, I mean, the culture of the valley for the last 30 to 40 years. but in the sense that you guys made some huge bets in terms of their journey to be successful Big data is kind of, you know, mm. create the culture to go do this. The coder audience is all the Millennials. for that group in the middle to say, you know, you know, join, jump in the wave, right? so people need to get out there and try something. and you guys of course have focused on data scientists, that you embed in your product. When you talk to your clients about have all the data they need to do their job? And so, when you think about governance, and the technology business? And to some extent, I don't care where you start, that in order to be data driven, you know, buzzword, Which is, it's not, you know, it's not normal CUBE is going to be there, very excited about that. I think it's going to be an amazing show. and we'll see you in about a month! from IBM Analytics, the GM at IBM Analytics.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Amazon | ORGANIZATION | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
December | DATE | 0.99+ |
Rob Thomas | PERSON | 0.99+ |
New York | LOCATION | 0.99+ |
Dinesh | PERSON | 0.99+ |
AMC Networks | ORGANIZATION | 0.99+ |
John | PERSON | 0.99+ |
Jeremiah Owyang | PERSON | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
Rob | PERSON | 0.99+ |
20 years | QUANTITY | 0.99+ |
Dave | PERSON | 0.99+ |
Munich | LOCATION | 0.99+ |
IBM Analytics | ORGANIZATION | 0.99+ |
Las Vegas | LOCATION | 0.99+ |
MIT | ORGANIZATION | 0.99+ |
10 million | QUANTITY | 0.99+ |
Apple | ORGANIZATION | 0.99+ |
20% | QUANTITY | 0.99+ |
last year | DATE | 0.99+ |
Furrier | PERSON | 0.99+ |
AMC | ORGANIZATION | 0.99+ |
One | QUANTITY | 0.99+ |
yesterday | DATE | 0.99+ |
six | QUANTITY | 0.99+ |
New York City | LOCATION | 0.99+ |
GDPR | TITLE | 0.99+ |
40% | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
three | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
Seattle | LOCATION | 0.99+ |
Scala | TITLE | 0.99+ |
two use cases | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
Python | TITLE | 0.98+ |
Andreessen | PERSON | 0.98+ |
both sides | QUANTITY | 0.98+ |
two | QUANTITY | 0.98+ |
Watson Health | ORGANIZATION | 0.98+ |
millions of rows | QUANTITY | 0.98+ |
five years | QUANTITY | 0.97+ |
next year | DATE | 0.97+ |
less than a month | QUANTITY | 0.97+ |
Madhu | PERSON | 0.97+ |
Amazons | ORGANIZATION | 0.96+ |
Steven Astorino, IBM - IBM Machine Learning Launch - #IBMML - #theCUBE
>> Announcer: Live from New York, it's the CUBE. Covering the IBM Machine Learning Launch Event. Brought to you by IBM. Now here are your hosts Dave Vellante and Stu Miniman. >> Welcome back to New York City everybody the is The CUBE the leader in live tech coverage. We're here at the IBM Machine Learning Launch Event, bringing machine learning to the Z platform. Steve Astorino is here, he's the VP for Development for the IBM Private Cloud Analytics Platform. Steve, good to see you, thanks for coming on. >> Hi how are you? >> Good thanks, how you doing? >> Good, good. >> Down from Toronto. So this is your baby. >> It is >> This product right? >> It is. So you developed this thing in the labs and now you point it at platforms. So talk about, sort of, what's new here today specifically. >> So today we're launching and announcing our machine learning, our IBM machine learning product. It's really a new solution that allows, obviously, machine learning to be automated and for data scientists and line of business, business analysts to work together and create models to be able to apply machine learning, do predictions and build new business models in the end. To provide better services for their customers. >> So how is it different than what we knew as Watson machine learning? Is it the same product pointed at Z or is it different? >> It's a great question. So Watson is our cloud solution, it's our cloud brand, so we're building something on private cloud for the private cloud customers and enterprises. Same product built for private cloud as opposed to public cloud. Think of it more as a branding and Watson is sort of a bigger solution set in the cloud. >> So it's your product, your baby, what's so great about it? How does it compare with what else is in the marketplace? Why should we get excited about this product? >> Actually, a bunch of things. It's great for many angles, what we're trying to do, obviously it's based on open source, it's an open platform just like what we've been talking about with the other products that we've been launching over the last six months to a year. It's based on Spark, you know we're bringing in all the open source technology, to your fingertips. As well as we're integrating with IBM's top-notch research and capabilities that we're driving in-house, integrating them together and being able to provide one experience to be able to do machine learning. That's at a very high level, also if you think about it there's three things that we're calling out, there's freedom, basically being able to choose what tools you want to use, what environments you want to use, what language you want to use, whether it's Python, Scala, R, right there's productivity. So we really enable and make it simple to be productive and build these machine learning models and then an application developer can leverage and use within their application. The other one is trust. IBM is very well known for its enterprise level capabilities, whether it's governance, whether its trust of the data, how to manage the data, but also more importantly, we're creating something called The Feedback Loop which allows the models to stay current and the data scientists, the administrators, know when these models, for example, is degrading. To make sure it's giving you the right outcome. >> OK, so you mention it's built on Spark. When I think about the efforts to build a data pipeline I think I've got to ingest the data, I've got to explore, I've got to process it and clean it up and then I've got to ultimately serve whomever, the business. >> Right, Right. >> What pieces of that does Spark unify and simplify? >> So we leveraged Spark to able to, obviously for the analytics. When you're building a model you one, have your choice of tooling that you want to use, whether it's programmatic or not. That's one of the value propositions we're bringing forward. But then we create these models, we train them, we evaluate them, we leverage Spark for that. Then obviously, we're trying to bring the models where the data is. So one of the key value proposition is we operationalize these models very simply and quickly. Just at a click of a button you can say hey deploy this model now and we deploy it right on where the data is in this case we're launching it on mainframe first. So Spark on the mainframe, we're deploying the model there and you can score the model directly in Spark on the mainframe. That's a huge value add, get better performance. >> Right, okay, just in terms of differentiates from the competition, you're the only company I think, providing machine learning on Z, so. >> Definitely, definitely. >> That's pretty easy, but in terms of the capabilities that you have, how are you different from the competition? When you talk to clients and they say well what about this vendor or that vendor, how do you respond? >> So let me talk about one of the research technologies that we're launching as part of this called CADS, Cognitive Assistant for Data Scientists. This is a feature where essentially, it takes the complexity out of building a model where you tell it, or you give it the algorithms you want to work with and the CADS assistant basically returns which one is the best which one performs the best. Now, all of a sudden you have the best model to use without having to go and spend, potentially weeks, on figuring out which one that is. So that's a huge value proposition. >> So automating the choice of the algorithm, an algorithm to choose the algorithm. what have you found in terms of it's level of accuracy in terms of the best fit? >> Actually it works really well. And in fact we have a live demo that we'll be doing today, where it shows CADS coming back with a 90% accurate model in terms of the data that we're feeding it and outcome it will give you in terms of what model to use. It works really well. >> Choosing an algorithm is not like choosing a programming language right, this bias if I like Scala or R or whatever, Java, Python okay fine, I've got skill sets associated with that. Algorithm choice is one that's more scientific, I guess? >> It is more scientific, it's based on the algorithm, the statistical algorithm and the selection of the algorithm or the model itself is a huge deal because that's where you're going to drive your business. If you're offering a new service that's where you're providing that solution from, so it has to be the right algorithm the right model so that you can build that more efficiently. >> What are you seeing as the big barriers to customer adopting machine learning? >> I think everybody, I mean it's the hottest thing around right now, everybody wants machine learning it's great, it's a huge buzz. The hardest thing is they know they want it, but don't really know how to apply it into their own environment, or they think they don't have the right skills. So, that actually one of the things that we're going after, to be able to enable them to do that. We're for example working on building different industry-based examples to showcase here's how you would use it in your environment. So last year when we did the Watson data platform we did a retail example, now today we're doing a finance example, a churn example with customers potentially churning and leaving a bank. So we're looking at all those different scenarios, and then also we're creating hubs, locations we're launching today also, announcing today, actually Dinesh will be doing that. There is a hub in Silicon Valley where it would allow customers to come in and work with us and we help them figure out how they can leverage machine learning. It is a great way to interact with our customers and be able to do that. >> So Steve nirvana is, and you gave that example, the retail example in September, when you launched Watson Data Platform, the nirvana in this world is you can use data, and maybe put in an offer, or save a patients life or effect an outcome in real time. So the retail example was just that. If I recall, you were making an offer real-time it was very fast, live demo it wasn't just a fakey. The example on churn, is the outcome is to effect that customer's decisions so that they don't leave? Is that? >> Yes, pretty much, Essentially what we are looking at is , we're using live data, we're using social media data bringing in Twitter sentiment about a particular individual for example, and try to predict if this customer, if this user is happy with the service that they are getting or not. So for example, people will go and socialize, oh I went to this bank and I hated this experience, or they really got me upset or whatever. Bringing that data from Twitter, so open data and merging it with the bank's data, banks have a lot of data they can leverage and monetize. And then making an assessment using machine learning to predict is this customer going to leave me or not? What probability do they have that they are going to leave me or not based on the machine learning model. The example or scenario we are using now, if we think they are going to leave us, we're going to make special offers to them. It's a way to enhance your service for those customers. So that they don't leave you. >> So operationalizing that would be a call center has some kind on dashboard that says red, green, yellow, boom heres an offer that you should make, and that's done in near real time. In fact, real time is before you lose the customer. That's as good a definition as anything else. >> But it's actually real-time, and when we call it the scoring of the data, so as the data transaction is coming in, you can actually make that assessment in real time, it's called in-transaction scoring where you can make that right on the fly and be able to determine is this customer at risk or not. And then be able to make smarter decisions to that service you are providing on whether you want to offer something better. >> So is the primary use case for this those streams those areas I'm getting you know, whether it be, you mentioned Twitter data, maybe IoT, you're getting can we point machine learning at just archives of data and things written historically or is it mostly the streams? >> It's both of course and machine learning is based on historical data right and that's hot the models are built. The more accurate or more data you have on historical data, the more accurate that you picked the right model and you'll get the better predictition of what's going to happen next time. So it's exactly, it's both. >> How are you helping customers with that initial fit? My understanding is how big of a data set do you need, Do I have enough to really model where I have, how do you help customers work through that? >> So my opinion is obvious to a certain extent, the more data you have as your sample set, the more accurate your model is going to be. So if we have one that's too small, your prediction is going to be inaccurate. It really depends on the scenario, it depends on how many features or the fields you have you're looking at within your dataset. It depends on many things, and it's variable depending on the scenario, but in general you want to have a good chunk of historical data that you can build expertise on right. >> So you've worked on both the Watson Services in the public cloud and now this private cloud, is there any differentiation or do you see significant use case different between those two or is it just kind of where the data lives and we're going to do similar activities there. >> So it is similar. At the end of the day, we're trying to provide similar products on both public cloud and private cloud. But for this specific case, we're launching it on mainframe that's a different angle at this. But we know that's where the biggest banks, the insurance companies, the biggest retailers in the world are, and that's where the biggest transactions are running and we really want to help them leverage machine learning and get their services to the next level. I think it's going to be a huge differentiator for them. >> Steve, you gave an example before of Twitter sentiment data. How would that fit in to this announcement. So I've got this ML on Z and I what API into the twitter data? How does that sort of all get adjusted and consolidated? >> So we allow hooks to be able to access data from different sources, bring in data. That is part of the ingest process. Then once you have that data there into data frames into the machine learning product, now you're feeding into a statistical algorithm to figure out what the best prediction is going to be, and the best model's going to be. >> I have a slide that you guys are sharing on the data scientist workflow. It starts with ingestion, selection, preparation, generation, transform, model. It's a complex set of tasks, and typically historically, at least in the last fIve or six years, different tools to de each of those. And not just different tools, multiples of different tools. That you had to cobble together. If I understand it correctly the Watson Data Platform was designed to really consolidate that and simplify that, provide collaboration tools for different personas, so my question is this. Because you were involved in that product as well. And I was excited about it when I saw it, I talked to people about it, sometimes I hear the criticism of well IBM just took a bunch of legacy products threw them together, threw and abstraction layer on top and is now going to wrap a bunch of services around it. Is that true? >> Absolutely not. Actually, you may have heard a while back IBM had made a big shift into design first design methodology. So we started with the Watson Data Platform, the Data Science Experience, they started with design first approach. We looked at this, we said what do we want the experience to be, for which persona do we want to target. Then we understood what we wanted the experience to be and then we leverage IBM analytics portfolio to be able to feed in and provide and integrate those services together to fit into that experience. So, its not a dumping ground for, I'll take this product, it's part of Watson Data Platform, not at all the case. It was the design first, and then integrate for that experience. >> OK, but there are some so-called legacy products in there, but you're saying you picked the ones that were relevant and then was there additional design done? >> There was a lot of work involved to take them from a traditional product, to be able to componentize, create a micro service architecture, I mean the whole works to be able to redesign it and fit into this new experience. >> So microservices architecture, runs on cloud, I think it only runs on cloud today right? >> Correct, correct. >> OK, maybe roadmap without getting too specific. What should we be paying attention to in the future? >> Right now we're doing our first release. Definitely we want to target any platform behind the firewall. So we don't have specific dates, but now we started with machine learning on a mainframe and we want to be able to target the other platforms behind the firewall and the private cloud environment. Definitely we should be looking at that. Our goal is to make, I talked about the feedback loop a little bit, so that is essentially once you deploy the model we actually look at that model you could schedule in a valuation, automatically, within the machine learning product. To be able to say, this model is still good enough. And if it's not we automatically flag it, and we look at the retraining process and redeployment process to make sure you always have the most up to date model. So this is truly machine learning where it requires very little to no intervention from a human. We're going to continue down that path and continue that automation in providing those capabilities so there's a bigger roadmap, there's a lot of things we're looking at. >> We've sort of looked at our big data analyst George Gilbert has talked about you had batch and you had interactive, not the sort of emergent workload is this continuous, streaming data. How do you see the adoption. First of all, is it a valid assertion? That there is a new class of workload, and then how do you see that adoption occurring? Is it going to be a dominant force over the next 10 years? >> Yeah, I think so. Like I said there is a huge buzz around machine learning in general and artificial intelligence, deep learning, all of these terms you hear about. I think as users and customers get more comfortable with understanding how they're going to leverage this in their enterprise. This real-time streaming of data and being able to do analytics on the fly and machine learning on the fly. It's a big deal and it will really helps them be more competitive in their own space with the services we're providing. >> OK Steve, thanks very much for coming on The CUBE. We'll give you the last word. The event, very intimate event a lot of customers coming in very shortly here in just a couple of hours. Give us the bumper sticker. >> All of that's very exciting, we're very excited, this is a big deal for us, that's why whenever IBM does a signature moment it's a big deal for us and we got something cool to talk about, we're very excited about that. Lot's of clients coming so there's an entire session this afternoon, which will be live streamed as well. So it's great, I think we have a differentiating product and we're already getting that feedback from our customers. >> Well congratulations, I love the cadence that you're on. We saw some announcements in September, we're here in February, I expect we're going to see more innovation coming out of your labs in Toronto, and cross IBM so thank you very much for coming on The CUBE. >> Thank you. >> You're welcome OK keep it right there everybody, we'll be back with our next guest right after this short break. This is The CUBE we're live from New York City. (energetic music)
SUMMARY :
Brought to you by IBM. for the IBM Private So this is your baby. and now you point it at platforms. and create models to be able for the private cloud the last six months to a year. the data, I've got to explore, So Spark on the mainframe, from the competition, you're the best model to use without So automating the of the data that we're feeding it Algorithm choice is one that's and the selection and be able to do that. the retail example in September, when you based on the machine learning model. boom heres an offer that you should make, and be able to determine on historical data, the more accurate the more data you have as your sample set, in the public cloud and and get their services to the next level. to this announcement. and the best model's going to be. and is now going to wrap a the experience to be, I mean the whole works attention to in the future? to make sure you always and then how do you see and machine learning on the fly. We'll give you the last word. So it's great, I think we and cross IBM so thank you very This is The CUBE we're
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Steve | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
George Gilbert | PERSON | 0.99+ |
Steve Astorino | PERSON | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
September | DATE | 0.99+ |
Toronto | LOCATION | 0.99+ |
90% | QUANTITY | 0.99+ |
February | DATE | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
New York City | LOCATION | 0.99+ |
Scala | TITLE | 0.99+ |
New York City | LOCATION | 0.99+ |
last year | DATE | 0.99+ |
New York | LOCATION | 0.99+ |
Python | TITLE | 0.99+ |
ORGANIZATION | 0.99+ | |
two | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
ORGANIZATION | 0.99+ | |
R | TITLE | 0.99+ |
both | QUANTITY | 0.99+ |
Java | TITLE | 0.99+ |
first release | QUANTITY | 0.98+ |
three things | QUANTITY | 0.98+ |
IBM Machine Learning Launch Event | EVENT | 0.97+ |
one experience | QUANTITY | 0.96+ |
one | QUANTITY | 0.96+ |
Watson Data Platform | TITLE | 0.96+ |
first approach | QUANTITY | 0.95+ |
Watson | TITLE | 0.95+ |
Steve nirvana | PERSON | 0.94+ |
Watson Data Platform | TITLE | 0.93+ |
Spark | TITLE | 0.93+ |
six years | QUANTITY | 0.92+ |
First | QUANTITY | 0.91+ |
Watson Services | ORGANIZATION | 0.91+ |
this afternoon | DATE | 0.9+ |
first | QUANTITY | 0.89+ |
last six months | DATE | 0.89+ |
each | QUANTITY | 0.86+ |
#IBMML | TITLE | 0.82+ |
Astorino | PERSON | 0.77+ |
Dinesh | ORGANIZATION | 0.76+ |
CUBE | ORGANIZATION | 0.74+ |
next 10 years | DATE | 0.72+ |
Private Cloud Analytics Platform | TITLE | 0.71+ |
a year | QUANTITY | 0.65+ |
first design methodology | QUANTITY | 0.65+ |
of clients | QUANTITY | 0.62+ |
Watson | ORGANIZATION | 0.55+ |
Loop | OTHER | 0.48+ |
James Kobielus, IBM - IBM Machine Learning Launch - #IBMML - #theCUBE
>> [Announcer] Live from New York, it's the Cube. Covering the IBM Machine Learning Launch Event. Brought to you by IBM. Now here are your hosts Dave Vellante and Stu Miniman. >> Welcome back to New York City everybody, this is the CUBE. We're here live at the IBM Machine Learning Launch Event. Bringing analytics and transactions together on Z, extending an announcement that IBM made a couple years ago, sort of laid out that vision, and now bringing machine learning to the mainframe platform. We're here with Jim Kobielus. Jim is the Director of IBM's Community Engagement for Data Science and a long time CUBE alum and friend. Great to see you again James. >> Great to always be back here with you. Wonderful folks from the CUBE. You ask really great questions and >> Well thank you. >> I'm prepared to answer. >> So we saw you last week at Spark Summit so back to back, you know, continuous streaming, machine learning, give us the lay of the land from your perspective of machine learning. >> Yeah well machine learning very much is at the heart of what modern application developers build and that's really the core secret sauce in many of the most disruptive applications. So machine learning has become the core of, of course, what data scientists do day in and day out or what they're asked to do which is to build, essentially artificial neural networks that can process big data and find patterns that couldn't normally be found using other approaches. And then as Dinesh and Rob indicated a lot of it's for regression analysis and classification and the other core things that data scientists have been doing for a long time, but machine learning has come into its own because of the potential for great automation of this function of finding patterns and correlations within data sets. So today at the IBM Machine Learning Launch Event, and we've already announced it, IBM Machine Learning for ZOS takes that automation promised to the next step. And so we're real excited and there'll be more details today in the main event. >> One of the most funs I had, most fun I had last year, most fun interviews I had last year was with you, when we interviewed, I think it was 10 data scientists, rock star data scientists, and Dinesh had a quote, he said, "Machine learning is 20% fun, 80% elbow grease." And data scientists sort of echoed that last year. We spent 80% of our time wrangling data. >> [Jim] Yeah. >> It gets kind of tedious. You guys have made announcements to address that, is the needle moving? >> To some degree the needle's moving. Greater automation of data sourcing and preparation and cleansing is ongoing. Machine learning is being used for that function as well. But nonetheless there is still a lot of need in the data science, sort of, pipeline for a lot of manual effort. So if you look at the core of what machine learning is all about, it's supervised learning involves humans, meaning data scientists, to train their algorithms with data and so that involves finding the right data and then of course doing the feature engineering which is a very human and creative process. And then to be training the data and iterating through models to improve the fit of the machine learning algorithms to the data. In many ways there's still a lot of manual functions that need expertise of data scientists to do it right. There's a lot of ways to do machine learning wrong you know there's a lot of, as it were, tricks of the trade you have to learn just through trial and error. A lot of things like the new generation of things like generative adversarial models ride on machine learning or deep learning in this case, a multilayered, and they're not easy to get going and get working effectively the first time around. I mean with the first run of your training data set, so that's just an example of how, the fact is there's a lot of functions that can't be fully automated yet in the whole machine learning process, but a great many can in fact, especially data preparation and transformation. It's being automated to a great degree, so that data scientists can focus on the more creative work that involves subject matter expertise and really also application development and working with larger teams of coders and subject matter experts and others, to be able to take the machine learning algorithms that have been proved out, have been trained, and to dry them to all manner of applications to deliver some disruptive business value. >> James, can you expand for us a little bit this democratization of before it was not just data but now the machine learning, the analytics, you know, when we put these massive capabilities in the broader hands of the business analysts the business people themselves, what are you seeing your customers, what can they do now that they couldn't do before? Why is this such an exciting period of time for the leveraging of data analytics? >> I don't know that it's really an issue of now versus before. Machine learning has been around for a number of years. It's artificial neural networks at the very heart, and that got going actually in many ways in the late 50s and it steadily improved in terms of sophistication and so forth. But what's going on now is that machine learning tools have become commercialized and refined to a greater degree and now they're in a form in the cloud, like with IBM machine learning for the private cloud on ZOS, or Watson machine learning for the blue mixed public cloud. They're at a level of consumability that they've never been at before. With software as a service offering you just, you pay for it, it's available to you. If you're a data scientist you being doing work right away to build applications, derive quick value. So in other words, the time to value on a machine learning project continues to shorten and shorten, due to the consumability, the packaging of these capabilities and to cloud offerings and into other tools that are prebuilt to deliver success. That's what's fundamentally different now and it's just an ongoing process. You sort of see the recent parallels with the business intelligence market. 10 years ago BI was reporting and OLEP and so forth, was only for the, what we now call data scientists or the technical experts and all that area. But in the last 10 years we've seen the business intelligence community and the industry including IBM's tools, move toward more self service, interactive visualization, visual design, BI and predictive analytics, you know, through our cognos and SPSS portfolios. A similar dynamic is coming in to the progress of machine learning, the democratization, to use your term, the more self service model wherein everybody potentially will be able to be, to do machine learning, to build machine learning and deep learning models without a whole of university training. That day is coming and it's coming fairly rapidly. It's just a matter of the maturation of this technology in the marketplace. >> So I want to ask you, you're right, 1950s it was artificial neural networks or AI, sort of was invented I guess, the concept, and then in the late 70s and early 80s it was heavily hyped. It kind of died in the late 80s or in the 90s, you never heard about it even the early 2000s. Why now, why is it here now? Is it because IBM's putting so much muscle behind it? Is it because we have Siri? What is it that has enabled that? >> Well I wish that IBM putting muscle behind a technology can launch anything to success. And we've done a lot of things in that regard. But the thing is, if you look back at the historical progress of AI, I mean, it's older than me and you in terms of when it got going in the middle 50s as a passion or a focus of computer scientists. What we had for the last, most of the last half century is AI or expert systems that were built on having to do essentially programming is right, declared a rule defining how AI systems could process data whatever under various scenarios. That didn't prove scalable. It didn't prove agile enough to learn on the fly from the statistical patterns within the data that you're trying to process. For face recognition and voice recognition, pattern recognition, you need statistical analysis, you need something along the lines of an artificial neural network that doesn't have to be pre-programmed. That's what's new now about in the last this is the turn of this century, is that AI has become predominantly now focused not so much on declarative rules, expert systems of old, but statistical analysis, artificial neural networks that learn from the data. See the, in the long historical sweep of computing, we have three eras of computing. The first era before the second world war was all electromechanical computing devices like IBM's start of course, like everybody's, was in that era. The business logic was burned into the hardware as it were. The second era from the second world war really to the present day, is all about software, programming, it's COBAL, 4trans, C, Java, where the business logic has to be developed, coded by a cadre of programmers. Since the turn of this millennium and really since the turn of this decade, it's all moved towards the third era, which is the cognitive era, where you're learning the business rules automatically from the data itself, and that involves machine learning at its very heart. So most of what has been commercialized and most of what is being deployed in the real world working, successful AI, is all built on artificial neural networks and cognitive computing in the way that I laid out. Where, you still need human beings in the equation, it can't be completely automated. There's things like unsupervised learning that take the automation of machine learning to a greater extent, but you still have the bulk of machine learning is supervised learning where you have training data sets and you need experts, data scientists, to manage that whole process, that over time supervised learning is evolving towards who's going to label the training data sets, especially when you have so much data flooding in from the internet of things and social media and so forth. A lot of that is being outsourced to crowd sourcing environments in terms of the ongoing labeling of data for machine learning projects of all sorts. That trend will continue a pace. So less and less of the actual labeling of the data for machine learning will need to be manually coded by data scientists or data engineers. >> So the more data the better. See I would argue in the enablement pie. You're going to disagree with that which is good. Let's have a discussion [Jim Laughs]. In the enablement pie, I would say the profundity of Hadup was two things. One is I can leave data where it is and bring code to data. >> [Jim] Yeah. >> 5 megabytes of code to petabyte of data, but the second was the dramatic reduction in the cost to store more data, hence my statement of the more data the better, but you're saying, meh maybe not. Certainly for compliance and other things you might not want to have data lying around. >> Well it's an open issue. How much data do you actually need to find the patterns of interest to you, the correlations of interest to you? Sampling of your data set, 10% sample or whatever, in most cases that might be sufficient to find the correlations you're looking for. But if you're looking for some highly deepened rare nuances in terms of anomalies or outliers or whatever within your data set, you may only find those if you have a petabyte of data of the population of interest. So but if you're just looking for broad historical trends and to do predictions against broad trends, you may not need anywhere near that amount. I mean, if it's a large data set, you may only need five to 10% sample. >> So I love this conversation because people have been on the CUBE, Abi Metter for example said, "Dave, sampling is dead." Now a statistician said that's BS, no way. Of course it's not dead. >> Storage isn't free first of all so you can't necessarily save and process all the data. Compute power isn't free yet, memory isn't free yet, so forth so there's lots... >> You're working on that though. >> Yeah sure, it's asymptotically all moving towards zero. But the bottom line is if the underlying resources, including the expertise of your data scientists that's not for free, these are human beings who need to make a living. So you've got to do a lot of things. A, automate functions on the data science side so that your, these experts can radically improve their productivity. Which is why the announcement today of IBM machine learning is so important, it enables greater automation in the creation and the training and deployment of machine learning models. It is a, as Rob Thomas indicated, it's very much a multiplier of productivity of your data science teams, the capability we offer. So that's the core value. Because our customers live and die increasingly by machine learning models. And the data science teams themselves are highly inelastic in the sense that you can't find highly skilled people that easily at an affordable price if you're a business. And you got to make the most of the team that you have and help them to develop their machine learning muscle. >> Okay, I want to ask you to weigh in on one of Stu's favorite topics which is man versus machine. >> Humans versus mechanisms. Actually humans versus bots, let's, okay go ahead. >> Okay so, you know a lot of discussions, about, machines have always replaced humans for jobs, but for the first time it's really beginning to replace cognitive functions. >> [Jim] Yeah. >> What does that mean for jobs, for skill sets? The greatest, I love the comment, the greatest chess player in the world is not a machine. It's humans and machines, but what do you see in terms of the skill set shift when you talk to your data science colleagues in these communities that you're building? Is that the right way to think about it, that it's the creativity of humans and machines that will drive innovation going forward. >> I think it's symbiotic. If you take Watson, of course, that's a star case of a cognitive AI driven machine in the cloud. We use a Watson all the time of course in IBM. I use it all the time in my job for example. Just to give an example of one knowledge worker and how he happens to use AI and machine learning. Watson is an awesome search engine. Through multi-structure data types and in real time enabling you to ask a sequence of very detailed questions and Watson is a relevance ranking engine, all that stuff. What I've found is it's helped me as a knowledge worker to be far more efficient in doing my upfront research for anything that I might be working on. You see I write blogs and I speak and I put together slide decks that I present and so forth. So if you look at knowledge workers in general, AI as driving far more powerful search capabilities in the cloud helps us to eliminate a lot of the grunt work that normally was attended upon doing deep research into like a knowledge corpus that may be preexisting. And that way we can then ask more questions and more intelligent questions and really work through our quest for answers far more rapidly and entertain and rule out more options when we're trying to develop a strategy. Because we have all the data at our fingertips and we've got this expert resource increasingly in a conversational back and forth that's working on our behalf predictively to find what we need. So if you look at that, everybody who's a knowledge worker which is really the bulk now of the economy, can be far more productive cause you have this high performance virtual assistant in the cloud. I don't know that it's really going, AI or deep learning or machine learning, is really going to eliminate a lot of those jobs. It'll just make us far smarter and more efficient doing what we do. That's, I don't want to belittle, I don't want to minimize the potential for some structural dislocation in some fields. >> Well it's interesting because as an example, you're like the, you're already productive, now you become this hyper-productive individual, but you're also very creative and can pick and choose different toolings and so I think people like you it's huge opportunities. If you're a person who used to put up billboards maybe it's time for retraining. >> Yeah well maybe you know a lot of the people like the research assistants and so forth who would support someone like me and most knowledge worker organizations, maybe those people might be displaced cause we would have less need for them. In the same way that one of my very first jobs out of college before I got into my career, I was a file clerk in a court in Detroit, it's like you know, a totally manual job, and there was no automation or anything. You know that most of those functions, I haven't revisited that court in recent years, I'm sure are automated because you have this thing called computers, especially PCs and LANs and so forth that came along since then. So a fair amount of those kinds of feather bedding jobs have gone away and in any number of bureaucracies due to automation and machine learning is all about automation. So who knows where we'll all end up. >> Alright well we got to go but I wanted to ask you about... >> [Jim] I love unions by the way. >> And you got to meet a lot of lawyers I'm sure. >> Okay cool. >> So I got to ask you about your community of data scientists that you're building. You've been early on in that. It's been a persona that you've really tried to cultivate and collaborate with. So give us an update there. What's your, what's the latest, what's your effort like these days? >> Yeah, well, what we're doing is, I'm on a team now that's managing and bringing together all of our program for community engagement programs for really for across portfolio not just data scientists. That involves meet ups and hack-a-thons and developer days and user groups and so forth. These are really important professional forums for our customers, our developers, our partners, to get together and share their expertise and provide guidance to each other. And these are very very important for these people to become very good at, to help them, get better at what they do, help them stay up to speed on the latest technologies. Like deep learning, machine learning and so forth. So we take it very seriously at IBM that communities are really where customers can realize value and grow their human capital ongoing so we're making significant investments in growing those efforts and bringing them together in a unified way and making it easier for like developers and IT administrators to find the right forums, the right events, the right content, within IBM channels and so forth, to help them do their jobs effectively and machine learning is at the heart, not just of data science, but other professions within the IT and business analytics universe, relying more heavily now on machine learning and understanding the tools of the trade to be effective in their jobs. So we're bringing, we're educating our communities on machine learning, why it's so critically important to the future of IT. >> Well your content machine is great content so congratulations on not only kicking that off but continuing it. Thanks Jim for coming on the CUBE. It's good to see you. >> Thanks for having me. >> You're welcome. Alright keep it right there everybody, we'll be back with our next guest. The CUBE, we're live from the Waldorf-Astoria in New York City at the IBM Machine Learning Launch Event right back. (techno music)
SUMMARY :
Brought to you by IBM. Great to see you again James. Wonderful folks from the CUBE. so back to back, you know, continuous streaming, and that's really the core secret sauce in many One of the most funs I had, most fun I had last year, is the needle moving? of the machine learning algorithms to the data. of machine learning, the democratization, to use your term, It kind of died in the late 80s or in the 90s, So less and less of the actual labeling of the data So the more data the better. but the second was the dramatic reduction in the cost the correlations of interest to you? because people have been on the CUBE, so you can't necessarily save and process all the data. and the training and deployment of machine learning models. Okay, I want to ask you to weigh in Actually humans versus bots, let's, okay go ahead. but for the first time it's really beginning that it's the creativity of humans and machines and in real time enabling you to ask now you become this hyper-productive individual, In the same way that one of my very first jobs So I got to ask you about your community and machine learning is at the heart, Thanks Jim for coming on the CUBE. in New York City at the IBM Machine Learning
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Jim Kobielus | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Jim | PERSON | 0.99+ |
Dinesh | PERSON | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
James | PERSON | 0.99+ |
80% | QUANTITY | 0.99+ |
James Kobielus | PERSON | 0.99+ |
20% | QUANTITY | 0.99+ |
Jim Laughs | PERSON | 0.99+ |
five | QUANTITY | 0.99+ |
Rob Thomas | PERSON | 0.99+ |
Detroit | LOCATION | 0.99+ |
1950s | DATE | 0.99+ |
last year | DATE | 0.99+ |
New York | LOCATION | 0.99+ |
New York City | LOCATION | 0.99+ |
10 data scientists | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
Siri | TITLE | 0.99+ |
Dave | PERSON | 0.99+ |
10% | QUANTITY | 0.99+ |
5 megabytes | QUANTITY | 0.99+ |
Abi Metter | PERSON | 0.99+ |
two things | QUANTITY | 0.99+ |
first time | QUANTITY | 0.99+ |
last week | DATE | 0.99+ |
second | QUANTITY | 0.99+ |
90s | DATE | 0.99+ |
ZOS | TITLE | 0.99+ |
Rob | PERSON | 0.99+ |
last half century | DATE | 0.99+ |
today | DATE | 0.99+ |
early 2000s | DATE | 0.98+ |
Java | TITLE | 0.98+ |
one | QUANTITY | 0.98+ |
C | TITLE | 0.98+ |
10 years ago | DATE | 0.98+ |
first run | QUANTITY | 0.98+ |
late 80s | DATE | 0.98+ |
Watson | TITLE | 0.97+ |
late 70s | DATE | 0.97+ |
late 50s | DATE | 0.97+ |
zero | QUANTITY | 0.97+ |
IBM Machine Learning Launch Event | EVENT | 0.96+ |
early 80s | DATE | 0.96+ |
4trans | TITLE | 0.96+ |
second world war | EVENT | 0.95+ |
IBM Machine Learning Launch Event | EVENT | 0.94+ |
second era | QUANTITY | 0.94+ |
IBM Machine Learning Launch | EVENT | 0.93+ |
Stu | PERSON | 0.92+ |
first jobs | QUANTITY | 0.92+ |
middle 50s | DATE | 0.91+ |
couple years ago | DATE | 0.89+ |
agile | TITLE | 0.87+ |
petabyte | QUANTITY | 0.85+ |
BAL | TITLE | 0.84+ |
this decade | DATE | 0.81+ |
three eras | QUANTITY | 0.78+ |
last 10 years | DATE | 0.78+ |
this millennium | DATE | 0.75+ |
third era | QUANTITY | 0.72+ |