Image Title

Search Results for Granger:

Vimal Endiran, Global Data Business Group Ecosystem Lead, Accenture @AccentureTech


 

>> Live from San Jose, in the heart of Silicon Valley, it's theCube. Covering Datawork Summit 2018. Brought to you by Hortonworks. >> Welcome back to theCube's live coverage of Dataworks here in San Jose, California. I'm your host, Rebecca Knight along with my cohost James Kobielus. We have with us Vimal Endiran. He is the Global Business Data Group Ecosystem Lead, at Accenture. He's coming to us straight from the Motor City. So, welcome Vimal. >> Thank you, thank you Rebecca. Thank you Jim. Looking forward to talk to you for the next ten minutes. >> So, before the cameras were rolling we were talking about how data veracity and how managers can actually know that the data that they're getting, that they're seeing, is trustworthy. What's your take on that right now? >> So, in the today's age the data is coming at you in a velocity that you never thought about, right. So today, the organizations are gathering data probably in the magnitude of petabytes. This is a new normal. We used to talk about gigs and now it's in petabytes. And the data coming in the form of images, video files, from the edge, you know edge devices, sensors, social media and everything. So, the amount of data, this is becoming the fuel for the new economy, right. So that companies, who can find a way to take advantage and figure out a way to use this data going to have a competitive advantage over their competitors. So, for that purpose, even though it's coming at that volume and velocity doesn't mean it's useful. So the thing is if they can find a way to make the data can be trustworthy, by the organization, and at the same time it's governed and secured. That's what's going to happen. It used to be it's called data quality, we call it when the structure it's okay, everything is maintained in SAP or some system. It's good it's coming to you. But now, you need to take advantage of the tools like machine learning, artificial intelligence, combining these algorithms and tool sets and abilities of people's mind, putting that in there and making it somewhat... Things can happen to itself at the same time it's trustworthy, we have offerings around that Accenture is developing place... It differs from industry to industry. Given the fact if the data coming in is something it's only worth for 15 seconds. After that it has no use other than understanding how to prevent something, from a sense of data. So, we have our offerings putting into place to make the data in a trustworthy, governed, secured, for an organization to use it and help the organization to get there. That's what we are doing. >> The standard user of your tools is it a data steward in the traditional sense or is it a data scientist or data engineer who's trying to, for example, compile a body of training data for use in building and training machine learning models? Do you see those kinds of customers for your data veracity offerings, that customer segment growing? >> Yes. We see both sides pretty much all walk of customers in our life. So, you hit the nail on the head, yes. We do see that type of aspects and also becoming, the data scientists you're also getting another set of people, the citizen data scientist. The people--- >> What is that? That's a controversial term. I've used that term on a number of occasions and a lot of my colleagues and peers in terms of other analysts bat me down and say, "No, that demeans the profession of data science by calling it..." But you tell me what how Accenture's defining that. >> The thing is, it's not demeaning. The fact is to become a citizen data scientist you need the help of data scientists. Basically, every time you need to build a model. And then you feed some data to learn. And then have an outcome to put that out. So you have a data scientist creating algorithms. What a citizen data scientist means, say if I'm not a data scientist, I should be able to take advantage of a model built for my business scenario, feed something data in, whatever I need to feed in, get an output and that program, that tool's going to tell me, go do this or don't do this, kind of things. So I become a data scientist by using a predefined model that's developed by an expert. Minds of many experts together. But rather than me going and hiring hundred experts, I go and buy a model and able to have one person maintain or tweak this model continuously. So, how can I enable that large volume of people by using more models. That's what-- >> If a predictive analytics tool that you would license from whatever vendor. If that includes prebuilt machine learning models for a particular tasks in it does that... Do you as a user of that tool, do you become automatically a citizen data scientist or do you need to do some actual active work with that model or data to live up to the notion of being a citizen data scientist? >> It's a good question. In my mind, I don't want to do it, my job is something else. To make something for the company. So, my job is not creating a model and doing that. My job is, I know my sets of data, I want to feed it in. I want to get the outcome that I can go and say increase my profit, increase my sales. That's what I want to do. So I may become a citizen data scientist without me knowing. I won't even be told that I'm using a model. I will take this set of data, feed it in here, it's going to tell you something. So, our data veracity point of view, we have these models built into some of platforms. That can be a tool from foreign works, taking advantage of the data storage tool or any other... In our own algorithms put in that helps you to create and maintain the data veracity to a scale of, if you say one to five, one is being low, five is being bad, to maintain at the five level. So that's the objective of that. >> So you're democratizing the tools of data science for the rest of us to solve real business problems. >> Right. >> So the data veracity aside, you're saying the user of these tools is doing something to manage, to correct or enhance or augment the data that's used to feed into these prebuilt models to achieve these outcomes? >> Yes. The augmented data, the feed data and the training data it comes out with an outcome to say, go do something. It tells you to perform something or do not perform. It's still an action. Comes out with an action to achieve a target. That's what it's going to be. >> You mention Hortonworks and since we are here at Dataworks and the Hortonworks show, tell us a little bit about your relationship with that company. >> Definitely. So Hortonworks is one of our premiere strategic partners. We've been the number one implementers, the partners for last two years in a row, implementing their technology across many of our clients. From partnership point of view, we have jointly developed offerings. What Accenture is best at, we're very good at industry knowledge. So with our industry knowledge and with their technology together what we're doing is we're creating some offerings that you can take to market. For example, we used to have data warehouses like using Teradata and older technology data warehouses. They're still good but at the same time, people also want to take the structured, unstructured data, images files and able to incorporate into the existing data warehouses. And how I can get the value out of the whole thing together. That's where Hortonworks' type of tools comes to play. So we have developed offerings called Modern Data Warehouse, taking advantage of your legacy systems you have plus this new data coming together and immediately you can create an analytics case, used case to do something. So, we have prebuilt programs and different scripts that take in different types of data. Moving into a data lake, Hortonworks data lake and then use your existing legacy data and all those together help you to create analytics use cases. So we have that called data modernization offering, we have one of that. Then we have-- >> So that's a prebuilt model for a specific vertical industry requirements or a specific business function, predictive analytics, anomaly detection and natural language processing, am I understanding correctly? >> Yes. We have industry based solutions as well but also to begin with, the data supply chain itself. To bring the data into the lake to use it. That's one of the offerings we play-- >> ...Pipeline and prepackaged models and rules and so forth. >> Right, prepackaged data ingestion, transformation, that prepackaged to take advantage with the new data sets along with your legacy data. That's one offering called data modernization offering. That to cloud. So, we can take to cloud. Hortonworks in a cloud it can be a joure, WS, HP, any cloud plus moving data. So that's one type of offering. Today actually we announced another offering jointly with Hortonworks, Atlas and Grainger Tool to help GDPR compliance. >> Will you explain what that tool does specifically to help customers with GDPR points. Does it work out of the box with Hortonworks data stewards studio? >> Well, to me I can get your answers from my colleagues who are much more technical on that but the fact is I can tell you functionally what the tool does is. >> Okay, please. >> So you, today the GDPR is basically, there's account regulations about you need to know about your personal data and you have your own destiny about your personal data. You can call the company and say, "Forget about me." If you are an EU resident. Or say, "Modify my data." They have to do it within certain time frame. If not they get fined. The fine can be up to 4% of the company's... So it's going to be a very large fine. >> Total revenue, yeah. >> So what we do is, basically take this tool. Put it in, working with Hortonworks this Atlas and Granger tool, we can go in and scan your data leak and we can scan at the metadata level and come into showcase. Then you know where is your personal data information about a consumer lies and now I know everything. Because what used to be in a legacy situation, the data originated someplace, somebody takes it and puts a system then somebody else downloads to an X file, somebody will put in an access data base and this kind of things. So now your data's pulling it across, you don't know where that lies. In this case, in the lake we can scan it, put this information, the meta data and the lineage information. Now, you immediately know where the data lies when somebody calls. Rebecca calls and says, "No longer use my information." I exactly know it's stored in this place in this table, in this column, let me go and take it out from here so that Rebecca doesn't exist anymore. Or whoever doesn't exist anymore. So that's the idea behind it. Also, we can catalog the entire data lake and we know not just personal information, other information, everything about other dimensions as well. And we can use it for our business advantage. So that's what we announced today. >> We're almost out of time but I want to finally ask you about talent because this is a pressing issue in Silicon Valley and beyond in really the tech industry, finding the right people, putting them in the right jobs and then keeping them happy there. So recruiting, retaining, what's Accenture's approach? >> This area, talent is the hardest one. >> Yes! >> Thanks to Hortonworks and Hortonworks point of view >> Send them to Detroit where the housing is far less expensive. >> Not a bad idea. >> Exactly! But the fact is-- >> We're both for Detroiters. >> What we did was, Hortonworks, Accenture has access to Hortonworks University, all their educational aspects. So we decided we're going to take that advantage and we going to enhance our talent by bringing the people from our... Retraining the people, taking the people to the new. People who know the legacy data aspects. So take them to see how we take the new world. So then we have a plan to use Hortonworks together the University, the materials and the people help, together we going to train about 500 people in different geos, 500 per piece and also our the development centers in India, Philippines, these places, so we have a larger plan to retrain the legacy into new. So, let's go and get people from out of the college and stuff, start building them from there, from an analyst to a consultant to a technical level and so that's the best way we are doing and actually the group I work with. Our group technology officer Sanjiv Vohra, he's basically in charge of training about 90,000 people on different technologies in and around that space. So the magnet is high but that's our approach to go and try and people and take it to that. >> Are you training them to be well rounded professionals in all things data or are you training them for specific specialties? >> Very, very good question. We do have this call master data architect program, so basically in the different levels after these trainings people go through specially you have to do so many projects, come back have an interview with a panel of people and you get certified, within the company, at certain level. At the master architect level you go and help a customer transform their data transformation, architecture vision where do you want to go to, that level. So we have the program with a university and that's the way we've taken it step by step to people to that level. >> Great. Vimal, thank you so much for coming on theCube. >> Thank you. >> It was really fun talking to you. >> Thank you so much, thank you for having me. Thank you. >> I'm Rebecca Knight for James Kobielus we will have more, well we actually will not be having any more coming up from Dataworks. This has been the Dataworks show. Thank you for tuning in. >> Signing off for now. >> And we'll see you next time.

Published Date : Jun 21 2018

SUMMARY :

Brought to you by Hortonworks. He is the Global Business Data Group Ecosystem Lead, Looking forward to talk to you for the next ten minutes. and how managers can actually know that the data and help the organization to get there. the data scientists "No, that demeans the profession of data science So you have a data scientist creating algorithms. or do you need to do some actual active work with that model and maintain the data veracity to a scale of, for the rest of us to solve real business problems. The augmented data, the feed data and the training data and the Hortonworks show, and immediately you can create an analytics case, To bring the data into the lake to use it. that prepackaged to take advantage with the new data sets to help customers with GDPR points. I can tell you functionally what the tool does is. and you have your own destiny about your personal data. So that's the idea behind it. and beyond in really the tech industry, Send them to Detroit and so that's the best way we are doing At the master architect level you go Vimal, thank you so much for coming on theCube. Thank you so much, thank you for having me. This has been the Dataworks show.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
RebeccaPERSON

0.99+

James KobielusPERSON

0.99+

VimalPERSON

0.99+

Rebecca KnightPERSON

0.99+

JimPERSON

0.99+

Sanjiv VohraPERSON

0.99+

HortonworksORGANIZATION

0.99+

IndiaLOCATION

0.99+

Vimal EndiranPERSON

0.99+

15 secondsQUANTITY

0.99+

Silicon ValleyLOCATION

0.99+

TodayDATE

0.99+

San JoseLOCATION

0.99+

Hortonworks UniversityORGANIZATION

0.99+

AccentureORGANIZATION

0.99+

fiveQUANTITY

0.99+

hundred expertsQUANTITY

0.99+

San Jose, CaliforniaLOCATION

0.99+

DetroitLOCATION

0.99+

HPORGANIZATION

0.99+

oneQUANTITY

0.99+

todayDATE

0.99+

both sidesQUANTITY

0.99+

Hortonworks,ORGANIZATION

0.99+

Hortonworks'ORGANIZATION

0.99+

bothQUANTITY

0.98+

WSORGANIZATION

0.98+

about 90,000 peopleQUANTITY

0.98+

500 per pieceQUANTITY

0.97+

TeradataORGANIZATION

0.97+

one personQUANTITY

0.97+

GDPRTITLE

0.97+

about 500 peopleQUANTITY

0.96+

Global Business Data Group EcosystemORGANIZATION

0.95+

five levelQUANTITY

0.93+

up to 4%QUANTITY

0.93+

EULOCATION

0.93+

Datawork Summit 2018EVENT

0.93+

DataworksORGANIZATION

0.93+

DetroitersPERSON

0.92+

@AccentureTechORGANIZATION

0.91+

Atlas and Grainger ToolORGANIZATION

0.88+

Global Data Business Group Ecosystem LeadORGANIZATION

0.86+

theCubeORGANIZATION

0.83+

PhilippinesLOCATION

0.8+

masterTITLE

0.77+

one typeQUANTITY

0.74+

petabytesQUANTITY

0.73+

SAPORGANIZATION

0.61+

last twoDATE

0.58+

ten minutesQUANTITY

0.58+

AtlasORGANIZATION

0.52+

yearsQUANTITY

0.5+

data architect programOTHER

0.48+

GrangerORGANIZATION

0.46+

Pandit Prasad, IBM | DataWorks Summit 2018


 

>> From San Jose, in the heart of Silicon Valley, it's theCube. Covering DataWorks Summit 2018. Brought to you by Hortonworks. (upbeat music) >> Welcome back to theCUBE's live coverage of Data Works here in sunny San Jose, California. I'm your host Rebecca Knight along with my co-host James Kobielus. We're joined by Pandit Prasad. He is the analytics, projects, strategy, and management at IBM Analytics. Thanks so much for coming on the show. >> Thanks Rebecca, glad to be here. >> So, why don't you just start out by telling our viewers a little bit about what you do in terms of in relationship with the Horton Works relationship and the other parts of your job. >> Sure, as you said I am in Offering Management, which is also known as Product Management for IBM, manage the big data portfolio from an IBM perspective. I was also working with Hortonworks on developing this relationship, nurturing that relationship, so it's been a year since the Northsys partnership. We announced this partnership exactly last year at the same conference. And now it's been a year, so this year has been a journey and aligning the two portfolios together. Right, so Hortonworks had HDP HDF. IBM also had similar products, so we have for example, Big Sequel, Hortonworks has Hive, so how Hive and Big Sequel align together. IBM has a Data Science Experience, where does that come into the picture on top of HDP, so it means before this partnership if you look into the market, it has been you sell Hadoop, you sell a sequel engine, you sell Data Science. So what this year has given us is more of a solution sell. Now with this partnership we go to the customers and say here is NTN experience for you. You start with Hadoop, you put more analytics on top of it, you then bring Big Sequel for complex queries and federation visualization stories and then finally you put Data Science on top of it, so it gives you a complete NTN solution, the NTN experience for getting the value out of the data. >> Now IBM a few years back released a Watson data platform for team data science with DSX, data science experience, as one of the tools for data scientists. Is Watson data platform still the core, I call it dev ops for data science and maybe that's the wrong term, that IBM provides to market or is there sort of a broader dev ops frame work within which IBM goes to market these tools? >> Sure, Watson data platform one year ago was more of a cloud platform and it had many components of it and now we are getting a lot of components on to the (mumbles) and data science experience is one part of it, so data science experience... >> So Watson analytics as well for subject matter experts and so forth. >> Yes. And again Watson has a whole suit of side business based offerings, data science experience is more of a a particular aspect of the focus, specifically on the data science and that's been now available on PRAM and now we are building this arm from stack, so we have HDP, HDF, Big Sequel, Data Science Experience and we are working towards adding more and more to that portfolio. >> Well you have a broader reference architecture and a stack of solutions AI and power and so for more of the deep learning development. In your relationship with Hortonworks, are they reselling more of those tools into their customer base to supplement, extend what they already resell DSX or is that outside of the scope of the relationship? >> No it is all part of the relationship, these three have been the core of what we announced last year and then there are other solutions. We have the whole governance solution right, so again it goes back to the partnership HDP brings with it Atlas. IBM has a whole suite of governance portfolio including the governance catalog. How do you expand the story from being a Hadoop-centric story to an enterprise data-like story, and then now we are taking that to the cloud that's what Truata is all about. Rob Thomas came out with a blog yesterday morning talking about Truata. If you look at it is nothing but a governed data-link hosted offering, if you want to simplify it. That's one way to look at it caters to the GDPR requirements as well. >> For GDPR for the IBM Hortonworks partnership is the lead solution for GDPR compliance, is it Hortonworks Data Steward Studio or is it any number of solutions that IBM already has for data governance and curation, or is it a combination of all of that in terms of what you, as partners, propose to customers for soup to nuts GDPR compliance? Give me a sense for... >> It is a combination of all of those so it has a HDP, its has HDF, it has Big Sequel, it has Data Science Experience, it had IBM governance catalog, it has IBM data quality and it has a bunch of security products, like Gaurdium and it has some new IBM proprietary components that are very specific towards data (cough drowns out speaker) and how do you deal with the personal data and sensitive personal data as classified by GDPR. I'm supposed to query some high level information but I'm not allowed to query deep into the personal information so how do you blog those queries, how do you understand those, these are not necessarily part of Data Steward Studio. These are some of the proprietary components that are thrown into the mix by IBM. >> One of the requirements that is not often talked about under GDPR, Ricky of Formworks got in to it a little bit in his presentation, was the notion that the requirement that if you are using an UE citizen's PII to drive algorithmic outcomes, that they have the right to full transparency. It's the algorithmic decision paths that were taken. I remember IBM had a tool under the Watson brand that wraps up a narrative of that sort. Is that something that IBM still, it was called Watson Curator a few years back, is that a solution that IBM still offers, because I'm getting a sense right now that Hortonworks has a specific solution, not to say that they may not be working on it, that addresses that side of GDPR, do you know what I'm referring to there? >> I'm not aware of something from the Hortonworks side beyond the Data Steward Studio, which offers basically identification of what some of the... >> Data lineage as opposed to model lineage. It's a subtle distinction. >> It can identify some of the personal information and maybe provide a way to tag it and hence, mask it, but the Truata offering is the one that is bringing some new research assets, after GDPR guidelines became clear and then they got into they are full of how do we cater to those requirements. These are relatively new proprietary components, they are not even being productized, that's why I am calling them proprietary components that are going in to this hosting service. >> IBM's got a big portfolio so I'll understand if you guys are still working out what position. Rebecca go ahead. >> I just wanted to ask you about this new era of GDPR. The last Hortonworks conference was sort of before it came into effect and now we're in this new era. How would you say companies are reacting? Are they in the right space for it, in the sense of they're really still understand the ripple effects and how it's all going to play out? How would you describe your interactions with companies in terms of how they're dealing with these new requirements? >> They are still trying to understand the requirements and interpret the requirements coming to terms with what that really means. For example I met with a customer and they are a multi-national company. They have data centers across different geos and they asked me, I have somebody from Asia trying to query the data so that the query should go to Europe, but the query processing should not happen in Asia, the query processing all should happen in Europe, and only the output of the query should be sent back to Asia. You won't be able to think in these terms before the GDPR guidance era. >> Right, exceedingly complicated. >> Decoupling storage from processing enables those kinds of fairly complex scenarios for compliance purposes. >> It's not just about the access to data, now you are getting into where the processing happens were the results are getting displayed, so we are getting... >> Severe penalties for not doing that so your customers need to keep up. There was announcement at this show at Dataworks 2018 of an IBM Hortonwokrs solution. IBM post-analytics with with Hortonworks. I wonder if you could speak a little bit about that, Pandit, in terms of what's provided, it's a subscription service? If you could tell us what subset of IBM's analytics portfolio is hosted for Hortonwork's customers? >> Sure, was you said, it is a a hosted offering. Initially we are starting of as base offering with three products, it will have HDP, Big Sequel, IBM DB2 Big Sequel and DSX, Data Science Experience. Those are the three solutions, again as I said, it is hosted on IBM Cloud, so customers have a choice of different configurations they can choose, whether it be VMs or bare metal. I should say this is probably the only offering, as of today, that offers bare metal configuration in the cloud. >> It's geared to data scientist developers and machine-learning models will build the models and train them in IBM Cloud, but in a hosted HDP in IBM Cloud. Is that correct? >> Yeah, I would rephrase that a little bit. There are several different offerings on the cloud today and we can think about them as you said for ad-hoc or ephemeral workloads, also geared towards low cost. You think about this offering as taking your on PRAM data center experience directly onto the cloud. It is geared towards very high performance. The hardware and the software they are all configured, optimized for providing high performance, not necessarily for ad-hoc workloads, or ephemeral workloads, they are capable of handling massive workloads, on sitcky workloads, not meant for I turned this massive performance computing power for a couple of hours and then switched them off, but rather, I'm going to run these massive workloads as if it is located in my data center, that's number one. It comes with the complete set of HDP. If you think about it there are currently in the cloud you have Hive and Hbase, the sequel engines and the stories separate, security is optional, governance is optional. This comes with the whole enchilada. It has security and governance all baked in. It provides the option to use Big Sequel, because once you get on Hadoop, the next experience is I want to run complex workloads. I want to run federated queries across Hadoop as well as other data storage. How do I handle those, and then it comes with Data Science Experience also configured for best performance and integrated together. As a part of this partnership, I mentioned earlier, that we have progress towards providing this story of an NTN solution. The next steps of that are, yeah I can say that it's an NTN solution but are the product's look and feel as if they are one solution. That's what we are getting into and I have featured some of those integrations. For example Big Sequel, IBM product, we have been working on baking it very closely with HDP. It can be deployed through Morey, it is integrated with Atlas and Granger for security. We are improving the integrations with Atlas for governance. >> Say you're building a Spark machine learning model inside a DSX on HDP within IH (mumbles) IBM hosting with Hortonworks on HDP 3.0, can you then containerize that machine learning Sparks and then deploy into an edge scenario? >> Sure, first was Big Sequel, the next one was DSX. DSX is integrated with HDP as well. We can run DSX workloads on HDP before, but what we have done now is, if you want to run the DSX workloads, I want to run a Python workload, I need to have Python libraries on all the nodes that I want to deploy. Suppose you are running a big cluster, 500 cluster. I need to have Python libraries on all 500 nodes and I need to maintain the versioning of it. If I upgrade the versions then I need to go and upgrade and make sure all of them are perfectly aligned. >> In this first version will you be able build a Spark model and a Tesorflow model and containerize them and deploy them. >> Yes. >> Across a multi-cloud and orchestrate them with Kubernetes to do all that meshing, is that a capability now or planned for the future within this portfolio? >> Yeah, we have that capability demonstrated in the pedestal today, so that is a new one integration. We can run virtual, we call it virtual Python environment. DSX can containerize it and run data that's foreclosed in the HDP cluster. Now we are making use of both the data in the cluster, as well as the infrastructure of the cluster itself for running the workloads. >> In terms of the layers stacked, is also incorporating the IBM distributed deep-learning technology that you've recently announced? Which I think is highly differentiated, because deep learning is increasingly become a set of capabilities that are across a distributed mesh playing together as is they're one unified application. Is that a capability now in this solution, or will it be in the near future? DPL distributed deep learning? >> No, we have not yet. >> I know that's on the AI power platform currently, gotcha. >> It's what we'll be talking about at next year's conference. >> That's definitely on the roadmap. We are starting with the base configuration of bare metals and VM configuration, next one is, depending on how the customers react to it, definitely we're thinking about bare metal with GPUs optimized for Tensorflow workloads. >> Exciting, we'll be tuned in the coming months and years I'm sure you guys will have that. >> Pandit, thank you so much for coming on theCUBE. We appreciate it. I'm Rebecca Knight for James Kobielus. We will have, more from theCUBE's live coverage of Dataworks, just after this.

Published Date : Jun 19 2018

SUMMARY :

Brought to you by Hortonworks. Thanks so much for coming on the show. and the other parts of your job. and aligning the two portfolios together. and maybe that's the wrong term, getting a lot of components on to the (mumbles) and so forth. a particular aspect of the focus, and so for more of the deep learning development. No it is all part of the relationship, For GDPR for the IBM Hortonworks partnership the personal information so how do you blog One of the requirements that is not often I'm not aware of something from the Hortonworks side Data lineage as opposed to model lineage. It can identify some of the personal information if you guys are still working out what position. in the sense of they're really still understand the and interpret the requirements coming to terms kinds of fairly complex scenarios for compliance purposes. It's not just about the access to data, I wonder if you could speak a little that offers bare metal configuration in the cloud. It's geared to data scientist developers in the cloud you have Hive and Hbase, can you then containerize that machine learning Sparks on all the nodes that I want to deploy. In this first version will you be able build of the cluster itself for running the workloads. is also incorporating the IBM distributed It's what we'll be talking next one is, depending on how the customers react to it, I'm sure you guys will have that. Pandit, thank you so much for coming on theCUBE.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
RebeccaPERSON

0.99+

James KobielusPERSON

0.99+

Rebecca KnightPERSON

0.99+

EuropeLOCATION

0.99+

IBMORGANIZATION

0.99+

AsiaLOCATION

0.99+

Rob ThomasPERSON

0.99+

San JoseLOCATION

0.99+

Silicon ValleyLOCATION

0.99+

PanditPERSON

0.99+

last yearDATE

0.99+

PythonTITLE

0.99+

yesterday morningDATE

0.99+

HortonworksORGANIZATION

0.99+

three solutionsQUANTITY

0.99+

RickyPERSON

0.99+

NorthsysORGANIZATION

0.99+

HadoopTITLE

0.99+

Pandit PrasadPERSON

0.99+

GDPRTITLE

0.99+

IBM AnalyticsORGANIZATION

0.99+

first versionQUANTITY

0.99+

bothQUANTITY

0.99+

one year agoDATE

0.98+

HortonworkORGANIZATION

0.98+

threeQUANTITY

0.98+

todayDATE

0.98+

DSXTITLE

0.98+

FormworksORGANIZATION

0.98+

this yearDATE

0.98+

AtlasORGANIZATION

0.98+

firstQUANTITY

0.98+

GrangerORGANIZATION

0.97+

GaurdiumORGANIZATION

0.97+

oneQUANTITY

0.97+

Data Steward StudioORGANIZATION

0.97+

two portfoliosQUANTITY

0.97+

TruataORGANIZATION

0.96+

DataWorks Summit 2018EVENT

0.96+

one solutionQUANTITY

0.96+

one wayQUANTITY

0.95+

next yearDATE

0.94+

500 nodesQUANTITY

0.94+

NTNORGANIZATION

0.93+

WatsonTITLE

0.93+

HortonworksPERSON

0.93+