Julie Lockner, IBM | IBM DataOps 2020
>>from the Cube Studios in Palo Alto and Boston connecting with thought leaders all around the world. This is a cube conversation. >>Hi, everybody. This is Dave Volante with Cuban. Welcome to the special digital presentation. We're really digging into how IBM is operational izing and automating the AI and data pipeline not only for its clients, but also for itself. And with me is Julie Lockner, who looks after offering management and IBM Data and AI portfolio really great to see you again. >>Great, great to be here. Thank you. Talk a >>little bit about the role you have here at IBM. >>Sure, so my responsibility in offering >>management and the data and AI organization is >>really twofold. One is I lead a team that implements all of the back end processes, really the operations behind any time we deliver a product from the Data and AI team to the market. So think about all of the release cycle management are seeing product management discipline, etcetera. The other role that I play is really making sure that I'm We are working with our customers and making sure they have the best customer experience and a big part of that is developing the data ops methodology. It's something that I needed internally >>from my own line of business execution. But it's now something that our customers are looking for to implement in their shops as well. >>Well, good. I really want to get into that. So let's let's start with data ops. I mean, I think you know, a lot of people are familiar with Dev Ops. Not maybe not everybody's familiar with data ops. What do we need to know about data? >>Well, I mean, you bring up the point that everyone knows Dev ops. And in fact, I think you know what data ops really >>does is bring a lot of the benefits that Dev Ops did for application >>development to the data management organizations. So when we look at what is data ops, it's a data management. Uh, it is a data management set of principles that helps organizations bring business ready data to their consumers. Quickly. It takes it borrows from Dev ops. Similarly, where you have a data pipeline that associates a business value requirement. I have this business initiative. It's >>going to drive this much revenue or this must cost >>savings. This is the data that I need to be able to deliver it. How do I develop that pipeline and map to the data sources Know what data it is? Know that I can trust it. So ensuring >>that it has the right quality that I'm actually using, the data that it was meant >>for and then put it to use. So in in history, most data management practices deployed a waterfall like methodology. Our implementation methodology and what that meant is all the data pipeline >>projects were implemented serially, and it was done based on potentially a first in first out program management office >>with a Dev Ops mental model and the idea of being able to slice through all of the different silos that's required to collect the data, to organize it, to integrate it, the validate its quality to create those data integration >>pipelines and then present it to the dashboard like if it's a Cognos dashboard >>or a operational process or even a data science team, that whole end to end process >>gets streamlined through what we're pulling data ops methodology. >>So I mean, as you well know, we've been following this market since the early days of Hadoop people struggle with their data pipelines. It's complicated for them, there's a a raft of tools and and and they spend most of their time wrangling data preparing data moving data quality, different roles within the organization. So it sounds like, you know, to borrow from from Dev Ops Data offices is all about streamlining that data pipeline, helping people really understand and communicate across. End the end, as you're saying, But but what's the ultimate business outcome that you're trying to drive? >>So when you think about projects that require data to again cut costs Teoh Artemia >>business process or drive new revenue initiatives, >>how long does it take to get from having access to the data to making it available? That duration for every time delay that is spent wasted trying to connect to data sources, trying to find subject matter experts that understand what the data means and can verify? It's quality, like all of those steps along those different teams and different disciplines introduces delay in delivering high quality data fat, though the business value of data ops is always associated with something that the business is trying to achieve but with a time element so if it's for every day, we don't have this data to make a decision where either making money or losing money, that's the value proposition of data ops. So it's about taking things that people are already doing today and figuring out the quickest way to do it through automation or work flows and just cutting through all the political barriers >>that often happens when these data's cross different organizational boundaries. >>Yes, sir, speed, Time to insights is critical. But in, you know, with Dev Ops, you really bringing together of the skill sets into, sort of, you know, one Super Dev or one Super ops. It sounds with data ops. It's really more about everybody understanding their role and having communication and line of sight across the entire organization. It's not trying to make everybody else, Ah, superhuman data person. It's the whole It's the group. It's the team effort, Really. It's really a team game here, isn't it? >>Well, that's a big part of it. So just like any type of practice, there's people, aspects, process, aspects and technology, right? So people process technology, and while you're you're describing it, like having that super team that knows everything about the data. The only way that's possible is if you have a common foundation of metadata. So we've seen a surgeons in the data catalog market in the last, you know, 67 years. And what what the what? That the innovation in the data catalog market has actually enabled us to be able >>to drive more data ops pipelines. >>Meaning as you identify data assets you captured the metadata capture its meaning. You capture information that can be shared, whether they're stakeholders, it really then becomes more of a essential repository for people don't really quickly know what data they have really quickly understand what it means in its quality and very quickly with the right proper authority, like privacy rules included. Put it to use >>for models, um, dashboards, operational processes. >>Okay. And we're gonna talk about some examples. And one of them, of course, is IBM's own internal example. But help us understand where you advise clients to start. I want to get into it. Where do I get started? >>Yeah, I mean, so traditionally, what we've seen with these large data management data governance programs is that sometimes our customers feel like this is a big pill to swallow. And what we've said is, Look, there's an operator. There's an opportunity here to quickly define a small project, align into high value business initiative, target something that you can quickly gain access to the data, map out these pipelines and create a squad of skills. So it includes a person with Dev ops type programming skills to automate an instrument. A lot of the technology. A subject matter expert who understands the data sources in it's meeting the line of business executive who translate bringing that information to the business project and associating with business value. So when we say How do you get started? We've developed A I would call it a pretty basic maturity model to help organizations figure out. Where are they in terms of the technology, where are they in terms of organizationally knowing who the right people should be involved in these projects? And then, from a process perspective, we've developed some pretty prescriptive project plans. They help you nail down. What are the data elements that are critical for this business business initiative? And then we have for each role what their jobs are to consolidate the data sets map them together and present them to the consumer. We find that six week projects, typically three sprints, are perfect times to be able to a timeline to create one of these very short, quick win projects. Take that as an opportunity to figure out where your bottlenecks are in your own organization, where your skill shortages are, and then use the outcome of that six week sprint to then focus on billing and gaps. Kick off the next project and iterating celebrate the success and promote the success because >>it's typically tied to a business value to help them create momentum for the next one. >>That's awesome. I want to get into some examples, I mean, or we're both Massachusetts based. Normally you'd be in our studio and we'd be sitting here for face to face of obviously with Kobe. 19. In this crisis world sheltering in place, you're up somewhere in New England. I happened to be in my studio, but I'm the only one here, so relate this to cove it. How would data ops, or maybe you have a, ah, a concrete example in terms of how it's helped, inform or actually anticipate and keep up to date with what's happening with both. >>Yeah, well, I mean, we're all experiencing it. I don't think there's a person >>on the planet who hasn't been impacted by what's been going on with this Cupid pandemic prices. >>So we started. We started down this data obscurity a year ago. I mean, this isn't something that we just decided to implement a few weeks ago. We've been working on developing the methodology, getting our own organization in place so that we could respond the next time we needed to be able todo act upon a data driven decision. So part of the step one of our journey has really been working with our global chief data officer, Interpol, who I believe you have had an opportunity to meet with an interview. So part of this year Journey has been working with with our corporate organization. I'm in a line of business organization where we've established the roles and responsibilities we've established the technology >>stack based on our cloud pack for data and Watson knowledge padlock. >>So I use that as the context. For now, we're faced with a pandemic prices, and I'm being asked in my business unit to respond very quickly. How can we prioritize the offerings that are going to help those in critical need so that we can get those products out to market? We can offer a 90 day free use for governments and hospital agencies. So in order for me to do that as a operations lead or our team, I needed to be able to have access to our financial data. I needed to have access to our product portfolio information. I needed to understand our cloud capacity. So in order for me to be able to respond with the offers that we recently announced and you'll you can take a look at some of the examples with our Watson Citizen Assistant program, where I was able to provide the financial information required for >>us to make those products available from governments, hospitals, state agencies, etcetera, >>that's a That's a perfect example. Now, to set the stage back to the corporate global, uh, the chief data office organization, they implemented some technology that allowed us to, in just data, automatically classify it, automatically assign metadata, automatically associate data quality so that when my team started using that data, we knew what the status of that information >>was when we started to build our own predictive models. >>And so that's a great example of how we've been partnered with a corporate central organization and took advantage of the automated, uh, set of capabilities without having to invest in any additional resources or head count and be able to release >>products within a matter of a couple of weeks. >>And in that automation is a function of machine intelligence. Is that right? And obviously, some experience. But you couldn't you and I when we were consultants doing this by hand, we couldn't have done this. We could have done it at scale anyway. It is it is it Machine intelligence and AI that allows us to do this. >>That's exactly right. And you know, our organization is data and AI, so we happen to have the research and innovation teams that are building a lot of this technology, so we have somewhat of an advantage there, but you're right. The alternative to what I've described is manual spreadsheets. It's querying databases. It's sending emails to subject matter experts asking them what this data means if they're out sick or on vacation. You have to wait for them to come back, and all of this was a manual process. And in the last five years, we've seen this data catalog market really become this augmented data catalog, and the augmentation means it's automation through AI. So with years of experience and natural language understanding, we can home through a lot of the metadata that's available electronically. We can calm for unstructured data, but we can categorize it. And if you have a set of business terms that have industry standard definitions through machine learning, we can automate what you and I did as a consultant manually in a matter of seconds. That's the impact that AI is have in our organization, and now we're bringing this to the market, and >>it's a It's a big >>part of where I'm investing. My time, both internally and externally, is bringing these types >>of concepts and ideas to the market. >>So I'm hearing. First of all, one of the things that strikes me is you've got multiple data, sources and data that lives everywhere. You might have your supply chain data in your er p. Maybe that sits on Prem. You might have some sales data that's sitting in a sas in a cloud somewhere. Um, you might have, you know, weather data that you want to bring in in theory. Anyway, the more data that you have, the better insights that you could gather assuming you've got the right data quality. But so let me start with, like, where the data is, right? So So it's it's anywhere you don't know where it's going to be, but you know you need it. So that's part of this right? Is being able >>to get >>to the data quickly. >>Yeah, it's funny. You bring it up that way. I actually look a little differently. It's when you start these projects. The data was in one place, and then by the time you get through the end of a project, you >>find out that it's moved to the cloud, >>so the data location actually changes. While we're in the middle of projects, we have many or even during this this pandemic crisis. We have many organizations that are using this is an opportunity to move to SAS. So what was on Prem is now cloud. But that shouldn't change the definition of the data. It shouldn't change. It's meaning it might change how you connect to it. It might also change your security policies or privacy laws. Now, all of a sudden, you have to worry about where is that data physically located? And am I allowed to share it across national boundaries right before we knew physically where it waas. So when you think about data ops, data ops is a process that sits on top of where the data physically resides. And because we're mapping metadata and we're looking at these data pipelines and automated work flows, part of the design principles are to set it up so that it's independent of where it resides. However, you have to have placeholders in your metadata and in your tool chain, where we're automating these work flows so that you can accommodate when the data decides to move. Because the corporate policy change >>from on prem to cloud. >>And that's a big part of what Data ops offers is the same thing. By the way, for Dev ops, they've had to accommodate building in, you know, platforms as a service versus on from the development environments. It's the same for data ops, >>and you know, the other part that strikes me and listening to you is scale, and it's not just about, you know, scale with the cloud operating model. It's also about what you were talking about is you know, the auto classification, the automated metadata. You can't do that manually. You've got to be able to do that. Um, in order to scale with automation, That's another key part of data office, is it not? >>It's a well, it's a big part of >>the value proposition and a lot of the part of the business case. >>Right then you and I started in this business, you know, and big data became the thing. People just move all sorts of data sets to these Hadoop clusters without capturing the metadata. And so as a result, you know, in the last 10 years, this information is out there. But nobody knows what it means anymore. So you can't go back with the army of people and have them were these data sets because a lot of the contact was lost. But you can use automated technology. You can use automated machine learning with natural, understand natural language, understanding to do a lot of the heavy lifting for you and a big part of data ops, work flows and building these pipelines is to do what we call management by exception. So if your algorithms say 80% confident that this is a phone number and your organization has a low risk tolerance, that probably will go to an exception. But if you have a you know, a match algorithm that comes back and says it's 99% sure this is an email address, right, and you have a threshold that's 98%. It will automate much of the work that we used to have to do manually. So that's an example of how you can automate, eliminate manual work and have some human interaction based on your risk threshold. >>That's awesome. I mean, you're right, the no schema on write said. I throw it into a data lake. Data Lake becomes a data swamp. We all know that joke. Okay, I want to understand a little bit, and maybe you have some other examples of some of the use cases here, but there's some of the maturity of where customers are. It seems like you've got to start by just understanding what data you have, cataloging it. You're getting your metadata act in order. But then you've got you've got a data quality component before you can actually implement and get yet to insight. So, you know, where are customers on the maturity model? Do you have any other examples that you can share? >>Yeah. So when we look at our data ops maturity model, we tried to simplify, and I mentioned this earlier that we try to simplify it so that really anybody can get started. They don't have to have a full governance framework implemented to to take advantage of the benefits data ops delivers. So what we did is we said if you can categorize your data ops programs into really three things one is how well do you know your data? Do you even know what data you have? The 2nd 1 is, and you trust it like, can you trust it's quality? Can you trust it's meeting? And the 3rd 1 is Can you put it to use? So if you really think about it when you begin with what data do you know, write? The first step is you know, how are you determining what data? You know? The first step is if you are using spreadsheets. Replace it with a data catalog. If you have a department line of business catalog and you need to start sharing information with the department's, then start expanding to an enterprise level data catalog. Now you mentioned data quality. So the first step is do you even have a data quality program, right. Have you even established what your criteria are for high quality data? Have you considered what your data quality score is comprised of? Have you mapped out what your critical data elements are to run your business? Most companies have done that for there. They're governed processes. But for these new initiatives And when you identify, I'm in my example with the covert prices, what products are we gonna help bring to market quickly? I need to be able to >>find out what the critical data elements are. And can I trust it? >>Have I even done a quality scan and have teams commented on it's trustworthiness to be used in this case, If you haven't done anything like that in your organization, that might be the first place to start. Pick the critical data elements for this initiative, assess its quality, and then start to implement the work flows to re mediate. And then when you get to putting it to use, there's several methods for making data available. One is simply making a gate, um, are available to a small set of users. That's what most people do Well, first, they make us spreadsheet of the data available, But then, if they need to have multiple people access it, that's when, like a Data Mart might make sense. Technology like data virtualization eliminates the need for you to move data as you're in this prototyping phase, and that's a great way to get started. It doesn't cost a lot of money to get a virtual query set up to see if this is the right join or the right combination of fields that are required for this use case. Eventually, you'll get to the need to use a high performance CTL tool for data integration. But Nirvana is when you really get to that self service data prep, where users can weary a catalog and say these are the data sets I need. It presents you a list of data assets that are available. I can point and click at these columns I want as part of my data pipeline and I hit go and automatically generates that output or data science use cases for it. Bad news, Dashboard. Right? That's the most mature model and being able to iterate on that so quickly that as soon as you get feedback that that data elements are wrong or you need to add something, you can do it. Push button. And that's where data obscurity should should bring organizations too. >>Well, Julie, I think there's no question that this covert crisis is accentuated the importance of digital. You know, we talk about digital transformation a lot, and it's it's certainly riel, although I would say a lot of people that we talk to we'll say, Well, you know, not on my watch. Er, I'll be retired before that all happens. Well, this crisis is accelerating. That transformation and data is at the heart of it. You know, digital means data. And if you don't have data, you know, story together and your act together, then you're gonna you're not gonna be able to compete. And data ops really is a key aspect of that. So give us a parting word. >>Yeah, I think This is a great opportunity for us to really assess how well we're leveraging data to make strategic decisions. And if there hasn't been a more pressing time to do it, it's when our entire engagement becomes virtual like. This interview is virtual right. Everything now creates a digital footprint that we can leverage to understand where our customers are having problems where they're having successes. You know, let's use the data that's available and use data ops to make sure that we can generate access. That data? No, it trust it, Put it to use so that we can respond to >>those in need when they need it. >>Julie Lockner, your incredible practitioner. Really? Hands on really appreciate you coming on the Cube and sharing your knowledge with us. Thank you. >>Thank you very much. It was a pleasure to be here. >>Alright? And thank you for watching everybody. This is Dave Volante for the Cube. And we will see you next time. >>Yeah, yeah, yeah, yeah, yeah
SUMMARY :
from the Cube Studios in Palo Alto and Boston connecting with thought leaders all around the world. portfolio really great to see you again. Great, great to be here. from the Data and AI team to the market. But it's now something that our customers are looking for to implement I mean, I think you know, I think you know what data ops really Similarly, where you have a data pipeline that associates a This is the data that I need to be able to deliver it. for and then put it to use. So it sounds like, you know, that the business is trying to achieve but with a time element so if it's for every you know, with Dev Ops, you really bringing together of the skill sets into, sort of, in the data catalog market in the last, you know, 67 years. Meaning as you identify data assets you captured the metadata capture its meaning. But help us understand where you advise clients to start. So when we say How do you get started? it's typically tied to a business value to help them create momentum for the next or maybe you have a, ah, a concrete example in terms of how it's helped, I don't think there's a person on the planet who hasn't been impacted by what's been going on with this Cupid pandemic Interpol, who I believe you have had an opportunity to meet with an interview. So in order for me to Now, to set the stage back to the corporate But you couldn't you and I when we were consultants doing this by hand, And if you have a set of business terms that have industry part of where I'm investing. Anyway, the more data that you have, the better insights that you could The data was in one place, and then by the time you get through the end of a flows, part of the design principles are to set it up so that it's independent of where it for Dev ops, they've had to accommodate building in, you know, and you know, the other part that strikes me and listening to you is scale, and it's not just about, So you can't go back with the army of people and have them were these data I want to understand a little bit, and maybe you have some other examples of some of the use cases So the first step is do you even have a data quality program, right. And can I trust it? able to iterate on that so quickly that as soon as you get feedback that that data elements are wrong And if you don't have data, you know, Put it to use so that we can respond to Hands on really appreciate you coming on the Cube and sharing Thank you very much. And we will see you next time.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Julie | PERSON | 0.99+ |
Julie Lockner | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Dave Volante | PERSON | 0.99+ |
New England | LOCATION | 0.99+ |
90 day | QUANTITY | 0.99+ |
99% | QUANTITY | 0.99+ |
80% | QUANTITY | 0.99+ |
Massachusetts | LOCATION | 0.99+ |
Data Mart | ORGANIZATION | 0.99+ |
first step | QUANTITY | 0.99+ |
98% | QUANTITY | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
Boston | LOCATION | 0.99+ |
67 years | QUANTITY | 0.99+ |
six week | QUANTITY | 0.99+ |
Cube Studios | ORGANIZATION | 0.99+ |
both | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
a year ago | DATE | 0.99+ |
first | QUANTITY | 0.98+ |
Dev Ops | ORGANIZATION | 0.98+ |
2nd 1 | QUANTITY | 0.97+ |
One | QUANTITY | 0.97+ |
First | QUANTITY | 0.97+ |
Interpol | ORGANIZATION | 0.97+ |
one place | QUANTITY | 0.97+ |
each role | QUANTITY | 0.97+ |
Hadoop | TITLE | 0.95+ |
Kobe | PERSON | 0.95+ |
SAS | ORGANIZATION | 0.95+ |
Cupid pandemic | EVENT | 0.94+ |
today | DATE | 0.93+ |
3rd 1 | QUANTITY | 0.93+ |
this year | DATE | 0.93+ |
few weeks ago | DATE | 0.88+ |
Prem | ORGANIZATION | 0.87+ |
last five years | DATE | 0.87+ |
2020 | DATE | 0.85+ |
three sprints | QUANTITY | 0.81+ |
one Super | QUANTITY | 0.8+ |
Nirvana | ORGANIZATION | 0.79+ |
Cuban | ORGANIZATION | 0.77+ |
three things | QUANTITY | 0.76+ |
pandemic | EVENT | 0.74+ |
step one | QUANTITY | 0.71+ |
one of them | QUANTITY | 0.7+ |
last 10 years | DATE | 0.69+ |
Dev Ops | TITLE | 0.69+ |
Teoh Artemia | ORGANIZATION | 0.68+ |
Cognos | ORGANIZATION | 0.61+ |
Watson Citizen Assistant | TITLE | 0.6+ |
Dev ops | TITLE | 0.6+ |
Cube | COMMERCIAL_ITEM | 0.57+ |
ops | ORGANIZATION | 0.54+ |
weeks | QUANTITY | 0.48+ |
Cube | ORGANIZATION | 0.47+ |
couple | QUANTITY | 0.47+ |
Watson | TITLE | 0.42+ |
UNLISTED FOR REVIEW Julie Lockner, IBM | DataOps In Action
from the cube studios in Palo Alto in Boston connecting with thought leaders all around the world this is a cube conversation hi everybody this is David on tape with the cube and welcome to the special digital presentation we're really digging into how IBM is operationalizing and automating the AI and data pipeline not only for its clients but also for itself and with me is Julie Lochner who looks after offering management and IBM's data and AI portfolio Julie great to see you again okay great to be here thank you talk a little bit about the role you have here at IBM sure so my responsibility in offering management in the data and AI organization is really twofold one is I lead a team that implements all of the back-end processes really the operations behind anytime we deliver a product from the data AI team to the market so think about all of the release cycle management pricing product management discipline etc the other roles that I play is really making sure that um we are working with our customers and making sure they have the best customer experience and a big part of that is developing the data ops methodology it's something that I needed internally from my own line of business execution but it's now something that our customers are looking for to implement in their shops as well well good I really want to get into that and so let's let's start with data ops I mean I think you know a lot of people are familiar with DevOps not maybe not everybody's familiar with the data Ops what do we need to know about data well I mean you bring up the point that everyone knows DevOps and and then in fact I think you know what data Ops really does is bring a lot of the benefits that DevOps did for application development to the data management organizations so when we look at what is data ops it's a data management it's a it's a data management set of principles that helps organizations bring business ready data to their consumers quickly it takes it borrows from DevOps similarly where you have a data pipeline that associates a business value requirement I have this business initiative it's gonna drive this much revenue or this much cost savings this is the data that I need to be able to deliver it how do I develop that pipeline and map to the data sources know what data it is know that I can trust it so ensuring that it has the right quality that I'm actually using the data that it was meant for and then put it to use so in in history most dated management practices deployed a waterfall like methodology or implementation methodology and what that meant is all the data pipeline projects were implemented serially and it was dawn based on potentially a first-in first-out program management office with a DevOps mental model and the idea of being able to slice through all of the different silos that's required to collect the data to organize it to integrate it to validate its quality to create those data integration pipelines and then present it to the dashboard like if it's a Cognos dashboard for a operational process or even a data science team that whole end-to-end process gets streamlined through what we're calling data ops methodology so I mean as you well know we've been following this market since the early days of a dupe and people struggle with their data pipelines it's complicated for them there's a raft of tools and and and they spend most of their time wrangling data preparing data improving data quality different roles within the organization so it sounds like you know to borrow from from DevOps data OPS's is all about REME lining that data pipeline helping people really understand and communicate across end to end as you're saying but but what's the ultimate business outcome that you're trying to drive so when you think about projects that require data to again cut cost to automate a business process or drive new revenue initiatives how long does it take to get from having access to the data to making it available that duration for every time delay that is spent wasted trying to connect to data sources trying to find subject matter experts that understand what the data means and can verify its quality like all of those steps along those different teams and different disciplines introduces delay in delivering high quality data fast so the business value of data Ops is always associated with something that the business is trying to achieve but with a time element so if it's for every day we don't have this data to make a decision we're either making money or losing money that's the value proposition of data ops so it's about taking things that people are already doing today and figuring out the quickest way to do it through automation through workflows and just cutting through all of the political barriers that often happens when these data's cross different organizational boundaries yeah so speed time to insights is critical but to in and then you know with DevOps you're really bringing together the skill sets into sort of you know one super dev or one super ops it sounds with data ops it's really more about everybody understanding their role and having communication and line-of-sight across the entire organization it's not trying to make everybody a superhuman data person it's the whole it's the group it's the team effort really it's really a team game here isn't it well that's a big part of it so just like any type of practice there's people aspects process aspects and technology right so people process technology and while you're you're describing it like having that super team that knows everything about the data the only way that's possible is if you have a common foundation of metadata so we've seen a surgeons in the data catalog market and last you know six seven years and what what the what that the innovation in the data catalog market has actually enabled us to be able to drive more data ops pipelines meaning as you identify data assets you've captured the metadata you capture its meaning you capture information that can be shared whether they're stakeholders it really then becomes more of a essential repository for people to really quickly know what data they have really quickly understand what it means in its quality and very quickly with the right proper authority like privacy rules included put it to use for models you know dashboards operational processes okay and and we're gonna talk about some examples and one of them of course is ibm's own internal example but but help us understand where you advise clients to start I want to get into it where do I get started yeah I mean so traditionally what we've seen with these large data management data governance programs is that sometimes our customers feel like this is a big pill to swallow and what we've said is look there's an opera there's an opportunity here to quickly define a small project align it to a high-value business initiative target something that you can quickly gain access to the data map out these pipelines and create a squad of skills so it includes a person with DevOps type programming skills to automate an instrument a lot of the technology a subject matter expert who understands the data sources and its meaning a line of business executive who can translate bringing that information to the business project and associating with business value so when we say how do you get started we've developed a I would call it a pretty basic maturity model to help organizations figure out where are they in terms of the technology where are they in terms of organizationally knowing who the right people should be involved in these projects and then from a process perspective we've developed some pretty prescriptive project plans that help you nail down what are the data elements that are critical for this business business initiative and then we have for each role what their jobs are to consolidate the datasets map them together and present them to the consumer we find that six-week projects typically three sprints are perfect times to be able to in a timeline to create one of these very short quick win projects take that as an opportunity to figure out where your bottlenecks are in your own organization where your skill shortages are and then use the outcome of that six-week sprint to then focus on filling in gaps kick off the next project and iterate celebrate the success and promote the success because it's typically tied to a business value to help them create momentum for the next one all right that's awesome I want to now get into some examples I mean or you're we're both massachusetts-based normally you'd be in our studio and we'd be sitting here face-to-face obviously with kovat 19 in this crisis we're all sheltering in place you're up in somewhere in New England I happen to be in my studio believe it but I'm the only one here so relate this to kovat how would data ops or maybe you have a concrete example in in terms of how it's helped inform or actually anticipate and keep up-to-date with what's happening with building yeah well I mean we're all experiencing it I don't think there's a person on the planet who hasn't been impacted by what's been going on with this coded pandemic crisis so we started we started down this data obscurity a year ago I mean this isn't something that we just decided to implement a few weeks ago we've been working on developing the methodology getting our own organization in place so that we could respond the next time we needed to be able to you know act upon a data-driven decision so part of step one of our journey has really been working with our global chief data officer Interpol who I believe you have had an opportunity to meet with an interview so part of this year journey has been working with with our corporate organization I'm in the line of business organization where we've established the roles and responsibilities we've established the technology stack based on our cloud pack for data and Watson knowledge catalog so I use that as the context for now we're faced with a pandemic crisis and I'm being asked in my business unit to respond very quickly how can we prioritize the offerings that are gonna help those in critical need so that we can get those products out to market we can offer a you know 90-day free use for governments and Hospital agencies so in order for me to do that as a operations lead for our team I needed to be able to have access to our financial data I needed to have access to our product portfolio information I needed to understand our cloud capacity so in order for me to be able to respond with the offers that we recently announced you know you can take a look at some of the examples with our Watson citizen assistant program where I was able to provide the financial information required for us to make those products available for governments hospitals state agencies etc that's a that's a perfect example now to to set the stage back to the corporate global chief data office organization they implemented some technology that allowed us to ingest data automatically classify it automatically assign metadata automatically associate data quality so that when my team started using that data we knew what the status of that information was when we started to build our own predictive models and so that's a great example of how we've partnered with a corporate central organization and took advantage of the automated set of capabilities without having to invest in any additional resources or headcount and be able to release products within a matter of a couple of weeks and in that automation is a function of machine intelligence is that right and obviously some experience but but you couldn't you and I when we were consultants doing this by hand we couldn't have done this we could have done it at scale anyways it is it machine intelligence an AI that allows us to do this that's exactly right and as you know our organization is data and AI so we happen to have the a research and innovation teams that are building a lot of this technology so we have somewhat of an advantage there but you're right the alternative to what I've described is manual spreadsheets it's querying databases it's sending emails to subject matter experts asking them what this data means if they're out sick or on vacation you have to wait for them to come back and all of this was a manual process and in the last five years we've seen this data catalog market really become this augmented data catalog and that augmentation means it's automation through AI so with years of experience and natural language understanding we can comb through a lot of the metadata that's available electronically we can comb through unstructured data we can categorize it and if you have a set of business terms that have industry standard definitions through machine learning we can automate what you and I did as a consultant manually in a matter of seconds that's the impact the AI is had in our organization and now we're bringing this to the market and it's a it's a big part of where I'm investing my time both internally and externally is bringing these types of concepts and ideas to the market so I'm hearing first of all one of the things that strikes me is you've got multiple data sources and data lives everywhere you might have your supply chain data and your ERP maybe that sits on Prem you might have some sales data that's sitting in the SAS store in a cloud somewhere you might have you know a weather data that you want to bring in in theory anyway the more data that you have the better insights that you can gather assuming you've got the right data quality but so let me start with like where the data is right so so it sits anywhere you don't know where it's gonna be but you know you need it so that that's part of this right is being able to read it quickly yeah it's funny you bring it up that way I actually look a little differently it's when you start these projects the data was in one place and then by the time you get through the end of a project you find out that it's a cloud so the data location actually changes while we're in the middle of projects we have many or coming even during this this pandemic crisis we have many organizations that are using this as an opportunity to move to SAS so what was on Prem is now cloud but that shouldn't change the definition of the data it shouldn't change its meaning it might change how you connect to it um it might also change your security policies or privacy laws now all of a sudden you have to worry about where is that data physically located and am I allowed to share it across national boundaries right before we knew physically where it was so when you think about data ops data ops is a process that sits on top of where the data physically resides and because we're mapping metadata and we're looking at these data pipelines and automated workflows part of the design principles are to set it up so that it's independent of where it resides however you have to have placeholders in your metadata and in your tool chain where we oughta mating these workflows so that you can accommodate when the data decides to move because of corporate policy change from on-prem to cloud then that's a big part of what data Ops offers it's the same thing by the way for DevOps they've had to accommodate you know building in you know platforms as a service versus on from the development environments it's the same for data ops and you know the other part that strikes me and listening to you is scale and it's not just about you know scale with the cloud operating model it's also about what you're talking about is you know the auto classification the automated metadata you can't do that manually you've got to be able to do that in order to scale with automation that's another key part of data Ops is it not it's well it's a big part of the value proposition and a lot of a part of the business base right then you and I started in this business you know and Big Data became the thing people just move all sorts of data sets to these Hadoop clusters without capturing the metadata and so as a result you know in the last 10 years this information is out there but nobody knows what it means anymore so you can't go back with the army of people and have them query these data sets because a lot of the contact was lost but you can use automated technology you can use automated machine learning with natural under Snatcher Alang guaa Jing to do a lot of the heavy lifting for you and a big part of data ops workflows and building these pipelines is to do what we call management-by-exception so if your algorithms say you know 80% confident that this is a phone number and your organization has a you know low risk tolerance that probably will go to an exception but if you have a you know a match algorithm that comes back and says it's 99 percent sure this is an email address right and you I have a threshold that's 98% it will automate much of the work that we used to have to do manually so that's an example of how you can automate eliminate manual work and have some human interaction based on your risk threshold now that's awesome I mean you're right the no schema on right said I throw it into a data leg the data link becomes the data swap we all know that joke okay I want to understand a little bit and maybe you have some other examples of some of the use cases here but there's some of the maturity of where customers are I mean it seems like you got to start by just understanding what data you have cataloging it you're getting your metadata act in order but then you've got a you've got a data quality component before you can actually implement and get yet to insight so you know where our customers on the on the maturity model do you have any other examples that you can share yeah so when we look at our data ops maturity model we tried to simplify it I mentioned this earlier that we try to simplify it so that really anybody can get started they don't have to have a full governance framework implemented to take advantage of the benefits data ops delivers so what we did we said if you can categorize your data ops programs into really three things one is how well do you know your data do you even know what data you have the second one is and you trust it like can you trust its quality can you trust its meeting and the third one is can you put it to use so if you really think about it when you begin with what data do you know right the first step is you know how are you determining what data you know the first step is if you are using spreadsheets replace it with a data catalog if you have a department line of business catalog and you need to start sharing information with the departments then start expanding to an enterprise level data catalog now you mentioned data quality so the first step is do you even have a data quality program right have you even established what your criteria are for high quality data have you considered what your data quality score is comprised of have you mapped out what your critical data elements are to run your business most companies have done that for they're they're governed processes but for these new initiatives and when you identify I'm in my example with the Kovach crisis what products are we gonna help bring to market quickly I need to be able to find out what the critical data elements are and can I trust it have I even done a quality scan and have teams commented on its trustworthiness to be used in this case if you haven't done anything like that in your organization that might be the first place to start pick the critical data elements for this initiative assess its quality and then start to implement the workflows to remediate and then when you get to putting it to use there's several methods for making data available you know one is simply making a data Mart available to a small set of users that's what most people do well first they make a spreadsheet of the data available but then if they need to have multiple people access it that's when like a data Mart might make sense technology like data virtualization eliminates the need for you to move data as you're in this prototyping phase and that's a great way to get started it doesn't cost a lot of money to get a virtual query set up to see if this is the right join or the right combination of fields that are required for this use case eventually you'll get to the need to use a high performance ETL tool for data integration but Nirvana is when you really get to that self-service data prep where users can query a catalog and say these are the data sets I need it presents you a list of data assets that are available I can point and click at these columns I want as part of my you know data pipeline and I hit go and it automatically generates that output for data science use cases for a Cognos dashboard right that's the most mature model and being able to iterate on that so quickly that as soon as you get feedback that that data elements are wrong or you need to add something you can do it push button and that's where data observation to bring organizations to well Julie I think there's no question that this kovat crisis is accentuated the importance of digital you know we talk about digital transformation a lot and it's it's certainly real although I would say a lot of people that we talk to will say well you know not on my watch or I'll be retired before that all happens will this crisis is accelerating that transformation and data is at the heart of it you know digital means data and if you don't have your data you know story together and your act together then you're gonna you're not going to be able to compete and data ops really is a key aspect of that so you know give us a parting word all right I think this is a great opportunity for us to really assess how well we're leveraging data to make strategic decisions and if there hasn't been a more pressing time to do it it's when our entire engagement becomes virtual like this interview is virtual write everything now creates a digital footprint that we can leverage to understand where our customers are having problems where they're having successes you know let's use the data that's available and use data ops to make sure that we can iterate access that data know it trust it put it to use so that we can respond to those in need when they need it Julie Locker your incredible practitioner really hands-on really appreciate you coming on the Kuban and sharing your knowledge with us thank you okay thank you very much it was a pleasure to be here all right and thank you for watching everybody this is Dave Volante for the cube and we will see you next time [Music]
**Summary and Sentiment Analysis are not been shown because of improper transcript**
ENTITIES
Entity | Category | Confidence |
---|---|---|
Julie Lochner | PERSON | 0.99+ |
Dave Volante | PERSON | 0.99+ |
Julie Lockner | PERSON | 0.99+ |
90-day | QUANTITY | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
99 percent | QUANTITY | 0.99+ |
Julie Locker | PERSON | 0.99+ |
80% | QUANTITY | 0.99+ |
six-week | QUANTITY | 0.99+ |
first step | QUANTITY | 0.99+ |
New England | LOCATION | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
first step | QUANTITY | 0.99+ |
98% | QUANTITY | 0.99+ |
Julie | PERSON | 0.99+ |
DevOps | TITLE | 0.99+ |
a year ago | DATE | 0.99+ |
Boston | LOCATION | 0.99+ |
David | PERSON | 0.98+ |
Watson | TITLE | 0.98+ |
second one | QUANTITY | 0.98+ |
six seven years | QUANTITY | 0.97+ |
Interpol | ORGANIZATION | 0.97+ |
third one | QUANTITY | 0.97+ |
one | QUANTITY | 0.97+ |
both | QUANTITY | 0.96+ |
Mart | ORGANIZATION | 0.94+ |
first place | QUANTITY | 0.93+ |
today | DATE | 0.92+ |
each role | QUANTITY | 0.91+ |
first | QUANTITY | 0.91+ |
a couple of weeks | QUANTITY | 0.88+ |
pandemic | EVENT | 0.88+ |
kovat | PERSON | 0.87+ |
three sprints | QUANTITY | 0.87+ |
three things | QUANTITY | 0.84+ |
step one | QUANTITY | 0.8+ |
guaa Jing | PERSON | 0.8+ |
few weeks ago | DATE | 0.78+ |
OPS | ORGANIZATION | 0.77+ |
one place | QUANTITY | 0.77+ |
ibm | ORGANIZATION | 0.75+ |
Nirvana | ORGANIZATION | 0.74+ |
last five years | DATE | 0.72+ |
DevOps | ORGANIZATION | 0.71+ |
this year | DATE | 0.7+ |
pandemic crisis | EVENT | 0.7+ |
last 10 years | DATE | 0.69+ |
a lot of people | QUANTITY | 0.68+ |
Cognos | TITLE | 0.66+ |
lot of money | QUANTITY | 0.66+ |
Kuban | LOCATION | 0.56+ |
DataOps | ORGANIZATION | 0.55+ |
Kovach | ORGANIZATION | 0.55+ |
Snatcher | PERSON | 0.51+ |
kovat | ORGANIZATION | 0.49+ |
lot | QUANTITY | 0.46+ |
19 | PERSON | 0.44+ |
massachusetts | PERSON | 0.42+ |
SAS | ORGANIZATION | 0.37+ |
Alang | PERSON | 0.31+ |
Jay Limburn, IBM & Julie Lockner, IBM | IBM Think 2019
>> Live from San Francisco, it's theCUBE! Covering IBM Think 2019. Brought to you by IBM. >> Welcome back, live here in San Francisco, it's theCUBE's coverage of IBM Think 2019. I'm John Furrier--Stu Miniman. Stu, four days, we're on our fourth day, the sun's shining, they've shut down Howard Street here at IBM. Big event for IBM, in San Francisco, not Las Vegas. Lot of great cloud action, lot of great AI data developers. Great story, good to see you again. Our next two guests, Julie Lockner, Director, Offering Management, Portfolio Operations at IBM, Data+AI, great to see you. >> Thank you, it's great to see you too, thank you. >> And Jay Limburn, Director of Offering Management, IBM Data+AI, thanks for coming on. >> Hey guys, great to be here. >> So, we've chatted many times at events, the role of data. So, we're religious about data, data flows through our blood, but IBM has put it all together now. All the reorgs are over, everyone's kind of, the table is set for IBM. The data path is clear, it's part of applications. It's feeding the apps. AI's the key workload inside the application. This is now a fully set-up group, give us the update, what's the focus? >> Yeah, it's really exciting because, if you think about it, before, we were called IBM Analytics, and that really is only a part of what we do. Now that we're Data+AI, that means that not only are we responsible for delivering data assets, and technology that supports those data assets to our customers, but infusing AI, not only in the technologies that we have, but also helping them build applications so they can fuse AI into their business processes. >> It's pretty broad, I mean, data's very much a broad swath of things. Analytics, you know, wrangling data, setting things up, cataloging them. Take me through how you guys set this up. How do you present it to the marketplace? How are clients engaged with it? Because it's pretty broad. But it could be, it needs to be specific. Take us through the methodology. >> So, you probably heard a lot of people today talk about the ladder to AI, right? This is IBM's view of how we explain our client's journey towards AI. It really starts at the bottom rung of the ladder, where we've got the collection of information. Collect your data. Once you've collected your data, you move up to the next rung, which is the Organize. And this is really where all the governance stuff comes in. This is how we can provide a view across that data, understand that data, provide trust to that data, and then serve that up to the consumers of that information, so they can actually use that in AI. That's where all the data science capabilities come in, allowing people to actually be able to consume that information. >> So, the bottom set is just really all the hard and heavy lifting that data scientists actually don't want to do. >> And writing algorithms, the collecting, the ingesting of data from any source, that's the bottom? And then, tell me about that next layer up, from the collection-- >> So, Collect is the physical assets or the collection of the data that you're going to be using for AI. If you don't get that foundation right, it doesn't really make sense. You have to have the data first. The piece in the middle that Jay was referring to, that's called Organize, our whole divisions are actually organized around these ladders to AI, so, Collect, Organize, Analyze, Infuse. On the Organize side, as Jay was mentioning, it's all about inventorying the data assets, knowing what data you have, then providing data quality rules, governance, compliance-type offerings, that allow organizations to not just know your data, trust your data, but then make it available so you can use your data, and the users are those data scientists, they're the analytics teams, they're the operation organizations that need to be able to build their solutions on top of trusted data. >> So, where does the Catalog fit in? Which level does that come into? >> Yeah, so, think of the Data Catalog as the DNS for data, all right? It's the way in which you can provide a full view of all of your information. Whether it's structured information, unstructured information, data you've got on PRAM and data you've got in a cloud somewhere. >> That's in the Organize layer, right? >> That's all in the Organize layer. So, if you can collect that information, you can then provide capabilities that allow you to understand the quality of that data, know where that data's come from, and then, finally, if you serve that up inside a compelling, business-friendly experience, so that a data scientist can go to one place, quickly make a decision on if that's the right data for them, and allow them to go and be productive by building a data science model, then we're really able to move the needle on making those data science organizations efficient, allowing us to build better models to transform their business. >> Yeah, and a big part of that is, if you think about what makes Amazon successful, it's because they know where all their products are, from the vendor, to when it shows up on the doorstep. What the Catalog provides is really the similar capability of, I would call it inventory management of your data assets, where we know where the data came from, its source--in that Collect layer-- who's transformed it, who's accessed it, if they're even allowed to see it, so, data privacy policies are part of that, and then being able to just serve up that data to those users. Being able to see that whole end-to-end lineage is a key point, critical point of the ladder to AI. Especially when you start to think about things like bias detection, which is a big part of the Analyze layer. >> But one of the things we've been digging into on theCUBE is, is data the next flywheel of innovation? You know, it used to be I just had my information, many years ago we started talking about, "Okay, I need to be able to access all that other information." We hear things like 80% of the data out there isn't really searchable today. So, how do you see data, data gravity, all those pieces, as the next flywheel of innovation? >> Yeah, I think it's key. I mean, we've talked a lot about how, you can't do AI without information architecture. And it's absolutely true. And getting that view of that data in a single location, so it is like the DNS of the internet. So you know exactly where to search, you can get hold of that data, and then you've got tools that give you self-service access to actually get hold of the data without any need of support from IT to get access to it. It's really a key-- >> Yeah, but to the point you were just asking about, data gravity? I mean, being able to do this where the data resides. So, for example, we have a lot of our customers that are mergers and acquisitions. Some teams have a lot of data assets that are on-premises, others have large data lakes in AWS or Azure. How do you inventory those assets and really have a view of what you have available across that landscape? Part of what we've been focusing on this year is making our technology work across all of those clouds. And having a single view of your assets but knowing where it resides. >> So, Julie, this environment is a bit more complicated than the old data warehousing, or even what we were looking at with big data and Hadoop and all those pieces. >> Isn't that the truth? >> Help explain why we're actually going to be able to get the information, leverage and drive new business value out of data today, when we've struggled so many times in the past. >> Well, I think the biggest thing that's changed is the adoption of DevOps, and when I say adoption of DevOps and things like containerization and Docker containers, Kubernetes, the ability to provision data assets very quickly, no matter where they are, build these very quick value-producing applications based on AI, Artificial Intelligence APIs, is what's allowing us to take advantage of this multi-cloud landscape. If you didn't have that DevOps foundation, you'd still be building ETL jobs in data warehouses, and that was 20 years ago. Today, it's much more about these microservices-based architecture, building up these AI-- >> Well, that's the key point, and the "Fuse" part of the stack, I think, or ladder. Stack? Ladder? >> Ladder. (laughs) >> Ladder to success! Is key, because you're seeing the applications that have data native into the app, where it has to have certain characteristics, whether it's a realtime healthcare app, or retail app, and we had the retail folks on earlier, it's like, oh my god, this now has to be addressable very fast, so, the old fenced-off data warehouse-- "Hey, give me that data!"--pull it over. You need a sub-second latency, or milliseconds. So, this is now a requirement. >> That's right. >> So, how are people getting there? What are some use cases? >> Sure. I'll start with the healthcare 'cause you brought that up. One of the big use cases for technology that we provide is really around taking information that might be realtime, or batch data, and providing the ability to analyze that data very quickly in realtime to the point where you can predict when someone might potentially have a cardiac arrest. And yesterday's keynote that Rob Thomas presented, a demonstration that showed the ability to take data from a wearable device, combine it with data that's sitting in an Amazon... MySQL database, be able to predict who is the most at-risk of having a potential cardiac arrest! >> That's me! >> And then present that to a call center of cardiologists. So, this company that we work with, iCure, really took that entire stack, Organize, Collect, Organize, Analyze, Infuse, and built an application in a matter of six weeks. Now, that's the most compelling part. We were able to build the solution, inventory their data assets, tie it to the industry model, healthcare industry model, and predict when someone might potentially-- >> Do you have that demo on you? The device? >> Of course I do. I know, I know. So, here is, this is called a BraveHeart Life Sensor. And essentially, it's a Bluetooth device. I know! If you put it on! (laughs) >> If I put it on, it'll track... Biometric? It'll start capturing information about your heart, ECG, and on Valentine's Day, right? My heart to yours, happy Valentine's Day to my husband, of course. The ability to be able to capture all this data here on the device, stream it to an AI engine that can then immediately classify whether or not someone has an anomaly in their ECG signal. You couldn't do that without having a complete ladder to AI capability. >> So, realtime telemetry from the heart. So, I see timing's important if you're about to have a heart attack. >> Yeah. >> Pretty important. >> And that's a great example of, you mentioned the speed. It's all about being able to capture that data in whatever form it's coming in, understand what that data is, know if you can trust that data, and then put it in the hands of the individuals that can do something valuable with the analysis from that data. >> Yeah, you have to able to trust it. Especially-- >> So, you brought up earlier bias in data. So, I want to bring that up in context of this. This is just one example of wearables, Fitbits, all kinds of things happening. >> New sources of tech, yeah. >> In healthcare, retail, all kinds of edge, realtime, is bias of data. And the other one's privacy because now you have a new kind of data source going into the cloud. And then, so, this fits into what part of the ladder? So, the ladder needs a secure piece. >> Tell me about that. >> Yeah, it does. So, that really falls into that Organize piece of that ladder, the governance aspects around it. If you're going to make data available for self-service, you've got to still make sure that that data's protected, and that you're not going to go and break any kind of regulatory law around that data. So, we actually can use technology now to understand what that data is, whether it contains sensitive information, credit card numbers, and expose that information out to those consumers, yet still masking the key elements that should be protected. And that's really important, because data science is a hugely inefficient business. Data scientists are spending too much time looking for information. And worse than that, they actually don't have all the information available that they need, because certain information needs to be protected. But what we can do now is expose information that wasn't previously available, but protect just the key parts of that information, so we're still ensuring it's safe. >> That's a really key point. It's the classic iceberg, right? What you see: "Oh, data science is going to "change the game of our business!" And then when they realize what's underneath the water, it's like, all this set-up, incompatible data, dirty data, data cleaning, and then all of a sudden it just doesn't work, right? This is the reality. Are you guys seeing this? Do you see that? >> Yeah, absolutely. I think we're only just really at the beginning of a crest of a wave, here. I think organizations know they want to get to AI, the ladder to AI really helps explain and it helps to understand how they can get there. And we're able then to solve that through our technology, and help them get there and drive those efficiencies that they need. >> And just to add to that, I mean, now that there's more data assets available, you can't manually classify, tag and inventory all that data, determine whether or not it contains sensitive data. And that's where infusing machine learning into our products has really allowed our customers to automate the process. I mentioned, the only way that we were able to deploy this application in six weeks, is because we used a lot of the embedded machine learning to identify the patient data that was considered sensitive, tag it as patient data, and then, when the data scientists were actually building the models in that same environment, it was masked. So, they knew that they had access to the data, but they weren't allowed to see it. It's perfectly--especially with HIMSS' conference this week as well! You were talking about this there. >> Great use case with healthcare. >> Love to hear you speak about the ecosystem being built around this. Everything, open APIs, I'm guessing? >> Oh, yeah. What kind of partners are-- >> Jay, talk a little bit-- >> Yeah, so, one of the key things we're doing is ensuring that we're able to keep this stuff open. We don't want to curate a proprietary system. We're already big supporters of open source, as you know, in IBM. One of the things that we're heavily-invested in is our open metadata strategy. Open metadata is part of the open source ODPi Foundation. Project Egeria defines a standard for common metadata interchange. And what that means is that, any of these metadata systems that adopt this standard can freely share and exchange metadata across that landscape, so that wherever your data is, whichever systems it's stored in, wherever that metadata is harvested, it can play part of that network and share that metadata across those systems. >> I'd like to get your thoughts on something, Julie. You've been on the analyst side, you're now at IBM. Jay, if you can weigh in on this too, that'd be great. We, here, we see all the trends and go to all the events and one of the things that's popping up that's clear within the IBM ecosystem because you guys have a lot of business customers, is that a new kind of business app developer's coming in. And we've seen data science highlight the citizen data scientist, so if data is code, part of the application, and all the ladder stuff kind of falls into place, that means we're going to see new kinds of applications. So, how are you guys looking at, this is kind of a, not like the cloud-native, hardcore DevOps developer. It's the person that says, "Hey, I can innovate "a business model." I see a business model innovation that's not so much about building technology, it's about using insight and a unique... Formula or algorithm, to tweak something. That's not a lot of programming involved. 'Cause with Cloud and Cloud Private, all these back end systems, that's an ecosystem partner opportunity for you guys, but it's not your classic ISV. So, there's a new breed of business apps that we see coming, your thoughts on this? >> Yeah, it's almost like taking business process optimization as a discipline, and turning it into micro-applications. You want to be able to leverage data that's available and accessible, be able to insert that particular Artificial Intelligence machine learning algorithm to optimize that business process, and then get out of the way. Because if you try to reinvent your entire business process, culture typically gets in the way of some of these things. >> I thought, as an application value, 'cause there's value creation here, right? >> Absolutely. >> You were talking about, so, is this a new kind of genre of developer, or-- >> It really is, I mean... If you take the citizen data scientist, an example that you mentioned earlier. It's really about lowering the entry point to that technology. How can you allow individuals with lower levels of skills to actually get in and be productive and create something valuable? It shouldn't be just a practice that's held away for the hardcore developer anymore. It's about lowering the entry point with the set of tools. One of the things we have in Watson Studio, for example, our data science platform, is just that. It's about providing wizards and walkthroughs to allow people to develop productive use models very easily, without needing hardcore coding skills. >> Yeah, I also think, though, that, in order for these value-added applications to be built, the data has to be business-ready. That's how you accelerate these application development life cycles. That's how you get the new class of application developers productive, is making sure that they start with a business-ready foundation. >> So, how are you guys going to go after this new market? What's the marketing strategy? Again, this is like, forward-pioneering kind of things happening. What's the strategy, how are you going to enable this, what's the plan? >> Well, there's two parts of it. One is, when Jay was mentioning the Open Metadata Repository Services, our key strategy is embedding Catalog everywhere and anywhere we can. We believe that having that open metadata exchange allows us to open up access to metadata across these applications. So, really, that's first and foremost, is making sure that we can catalog and inventory data assets that might not necessarily be in the IBM Cloud, or in IBM products. That's really the first step. >> Absolutely. The second step, I would say, is really taking all of our capabilities, making them, from the ground up, microservices-enabled, delivering them through Docker containers and making sure that they can port across whatever cloud deployment model our customers want to be able to execute on. And being able to optimize the runtime engines, whether it's data integration, data movement, data virtualization, based on data gravity, that you had mentioned-- >> So, something like a whole new developer program opportunity to bring to the market. >> Absolutely. I mean, there is, I think there is a huge opportunity for, from an education perspective, to help our customers build these applications. But it starts with understanding the data assets, understanding what they can do with it, and using self-service-type tools that Jay was referring to. >> And all of that underpinned with the trust. If you don't trust your data, the data scientist is not going to know whether or not they're using the right thing. >> So, the ladder's great. Great way for people to figure out where they are, it's like looking in the mirror, on the organization. How early is this? What inning are we in? How do you guys see the progression? How far along are we? Obviously, you have some data, examples, some people are doing it end-to-end. What's the maturity look like? What's the uptake? >> Go ahead, Jay. >> So, I think we're at the beginning of a crest of a wave. As I say, there's been a lot of discussion so far, even if you compare this year's conference to last year's. A lot of the discussion last year was, "What's possible with AI?" This year's conference is much more about, "What are we doing with AI?" And I think we're now getting to the point where people can actually start to be productive and really start to change their business through that. >> Yeah and, just to add to that, I mean, the ladder to AI was introduced last year, and it has gained so much adoption in the marketplace and our customers, they're actually organizing their business that way. So, the Collect divisions are the database teams, are now expanding to Hadoop and Cloudera, and Hortonworks and Mongo. They're organizing their data governance teams around the Organize pillar, where they're doing things like data integration, data replication. So, I feel like the maturity of this ladder to AI is really enabling our customers to achieve it much faster than-- >> I was talking to Dave Vellante about this, and we're seeing that, you know, we've been covering IBM since, it's the 10th year of theCUBE, all ten years. It's been, watching the progression. The past couple of years has been setting the table, everyone seems to be pumping, it makes sense, everything's hanging together, it's in one group. Data's not one, "This group, that group," it's all, Data, AI, all Analytics, all Watson. Smart, and the ladder just allows you to understand where a customer is, and then-- >> Well, and also, we mentioned the emphasis on open source. It allows our customers to take an inventory of, what do they have, internally, with IBM assets, externally, open source, so that they can actually start to architect their information architecture, using the same kind of analogy. >> And an opportunity for developers too, great. Julie, thanks for coming on. Jay, appreciate it. >> Thank you so much for the opportunity, happy Valentine's Day! Happy Valentine's Day, we're theCUBE. I'm John Furrier, Stu Miniman here, live in San Francisco at the Moscone Center, and the whole street's shut down, Howard Street. Huge event, 30,000 people, we'll be back with more Day Four coverage after this short break.
SUMMARY :
Brought to you by IBM. Great story, good to see you again. And Jay Limburn, Director of Offering Management, It's feeding the apps. not only in the technologies that we have, But it could be, it needs to be specific. talk about the ladder to AI, right? So, the bottom set is just really that need to be able to build their solutions It's the way in which you can provide so that a data scientist can go to one place, of the ladder to AI. is data the next flywheel of innovation? get hold of the data without any need Yeah, but to the point you were than the old data warehousing, going to be able to get the information, the ability to provision data assets of the stack, I think, or ladder. (laughs) that have data native into the app, the ability to analyze that data And then present that to a call center of cardiologists. If you put it on! The ability to be able to capture So, realtime telemetry from the heart. It's all about being able to capture that data Yeah, you have to able to trust it. So, you brought up earlier bias in data. And the other one's privacy because now you have of that ladder, the governance aspects around it. This is the reality. the ladder to AI really helps explain I mentioned, the only way that we were able Love to hear you speak about What kind of partners are-- One of the things that we're heavily-invested in and one of the things that's popping up be able to insert that particular One of the things we have in Watson Studio, for example, to be built, the data has to be business-ready. What's the strategy, how are you That's really the first step. that you had mentioned-- opportunity to bring to the market. from an education perspective, to help And all of that underpinned with the trust. So, the ladder's great. A lot of the discussion last year was, So, I feel like the maturity of this ladder to AI Smart, and the ladder just allows you It allows our customers to take an inventory of, And an opportunity for developers too, great. and the whole street's shut down, Howard Street.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Julie Lockner | PERSON | 0.99+ |
Jay Limburn | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
Jay | PERSON | 0.99+ |
Julie | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
San Francisco | LOCATION | 0.99+ |
80% | QUANTITY | 0.99+ |
Las Vegas | LOCATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Hortonworks | ORGANIZATION | 0.99+ |
last year | DATE | 0.99+ |
yesterday | DATE | 0.99+ |
first step | QUANTITY | 0.99+ |
second step | QUANTITY | 0.99+ |
two parts | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
Hadoop | ORGANIZATION | 0.99+ |
Howard Street | LOCATION | 0.99+ |
fourth day | QUANTITY | 0.99+ |
Moscone Center | LOCATION | 0.99+ |
10th year | QUANTITY | 0.99+ |
ODPi Foundation | ORGANIZATION | 0.99+ |
six weeks | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
Cloudera | ORGANIZATION | 0.99+ |
ten years | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
30,000 people | QUANTITY | 0.99+ |
Rob Thomas | PERSON | 0.99+ |
Mongo | ORGANIZATION | 0.99+ |
four days | QUANTITY | 0.98+ |
Today | DATE | 0.98+ |
Stu | PERSON | 0.98+ |
MySQL | TITLE | 0.98+ |
Valentine's Day | EVENT | 0.98+ |
20 years ago | DATE | 0.98+ |
iCure | ORGANIZATION | 0.97+ |
two guests | QUANTITY | 0.97+ |
Watson Studio | TITLE | 0.97+ |
2019 | DATE | 0.97+ |
this year | DATE | 0.96+ |
today | DATE | 0.96+ |
DevOps | TITLE | 0.95+ |
one group | QUANTITY | 0.95+ |
one | QUANTITY | 0.94+ |
Cloud | TITLE | 0.93+ |
single location | QUANTITY | 0.92+ |
IBM Data | ORGANIZATION | 0.92+ |
Project Egeria | ORGANIZATION | 0.9+ |
this week | DATE | 0.9+ |
one example | QUANTITY | 0.9+ |