Vasanth Kumar, MongoDB Principal Solutions Architect | Io-Tahoe Episode 7
>> Okay. We're here with Vasanth Kumar who's the Principal Solutions Architect for MongoDB. Vasanth, welcome to "theCube." >> Thanks Dave. >> Hey, listen, I feel like you were born to be an architect in technology. I mean, you've worked for big SIs, you've worked with many customers, you have experience in financial services and banking. Tell us, the audience, a little bit more about yourself, and what you're up to these days. >> Yeah. Hi, thanks for the for inviting me for this discussion. I'm based out of Bangalore, India, having around 18 years experience in IT industry, building enterprise products for different domains, verticals, finance built and enterprise banking applications, IOT platforms, digital experience solutions. Now being with MongoDB nearly two years, been working in a partner team as a principal solutions architect, especially working with ISBs to build the best practices of handling the data and embed the right database as part of their product. I also worked with technology partners to integrate the compatible technology compliance with MongoDB. And also worked with the private cloud providers to provide a database as a service. >> Got it. So, you know, I have to Vasanth, I think Mongo, you kind of nailed it. They were early on with the trends of managing unstructured data, making it really simple. There was always a developer appeal, which has lasted and then doing so with an architecture that scales out, and back in the early days when Mongo was founded, I remember those days, I mean, digital transformation, wasn't a thing, it wasn't a buzz word, but it just so happens that Mongo's approach, it dovetails very nicely with a digital business. So I wonder if you could talk about that, talk about the fit and how MongoDB thinks about accelerating digital transformation and why you're different from like a traditional RDBMS. >> Sure, exactly, yeah. You had a right understanding, let me elaborate it. So we all know that the customer expectation changes day by day, because of the business agility functionality changes, how they want to experience the applications, or in apps that changes okay. And obviously this yields to the agility of the information which transforms between the multiple systems or layers. And to achieve this, obviously the way of architecting or developing the product as completely a different shift, might be moving from the monolith to microservices or event-based architecture and so on. And obviously the database has to be opt for these environment to adopt these changes, to adopt the scale of load and the other thing. Okay. And also like we see that the common, the protocol for the information exchange is JSON, and something like you, you adopt it. The database adopts it natively to that is a perfect fit. Okay. So that's where the MongoDB fits perfectly for billing or transforming the modern applications, because it's a general purpose database which accepts the JSON as a payload and stores it in a BSON format. You don't need to be, suppose like to develop any particular application or to transfer an existing application, typically they see the what is the effort required and how much, what is the cost involved in it, and how quickly I can do that. That's main important thing without disturbing the functionality here where, since it is a multimodal database in a JSON format, you don't easily build an application. Okay? Don't need a lot of transformation in case of an RDBMS, you get the JSON payload, you transform into a tabular structure or a different format, and then probably you build an ORM layer and then map it and save it. There are lot of work involved in it. There are a lot of components need to be written in between. But in case of MongoDB, what they can do is you get the information from the multiple sources. And as is, you can put it in a DB based on where, or you can transform it based on the access patterns. And then you can store it quickly. >> Dave: Got it. And I tell Dave, because today you haven't context data, which has a selected set of information. Probably tomorrow the particular customer has more information to put it. So how do you capture that? In case of an RDBMS, you need to change the schema. Once you scheme change the schema, your application breaks down. But here it magically adopts it. Like you pass the extra information, it's open for extension. It adopts it easily. You don't need to redeploy or change the schema or do something like that. >> Right. That's the genius of Mongo. And then of course, you know, in the early days people say, oh, you know, Mongo, it won't scale. And then of course we, through the cloud. And I follow very closely Atlas. I look at the numbers every quarter. I mean, overall cloud adoption is increasing like crazy, you know, our Wiki Bon analyst team. We got the big four cloud vendors just in IAS growing beyond a 115 billion this year. That's 35% on top of, you know, 80-90 billion last year. So talk more about how MongoDB fits with the cloud and how it helps with the whole migration story. 'Cause you're killing it in that space. >> Yeah. Sure. Just to add one more point on the previous question. So for continuously, for past four to five years, we have been the number one in the wanted database. >> Dave: Right Okay. That that's how like the popularity is getting done. That's how the adoption has happened. >> Dave: Right. >> I'm coming back to your question- >> Yeah let's talk about the cloud and database as a service, you guys actually have packaged that very nicely I have to say. >> Yeah. So we have spent lot of effort and time in developing Atlas, our managed database as a service, which typically gives the customer the way of just concentrating on their application rather than maintaining and managing the whole set of database or how to scale infrastructure. All those things on work is taken care. You don't need to be an expert of DB, like when you are using an Atlas. So we provide the managed database in three major cloud providers, AWS, GCP, and Azure, and also it's a purely a multicloud, you know, like you can have a primary in AWS and you have the replicated nodes in GCP or Azure. It's a purely multicloud. So that like, you don't have a cloud blocking. You feel that, okay, your business is, I mean, if this is the right for your business you are choosing the model, you think that I need to move to GCP. You don't need to bother, you easily migrate this to GCP. Okay. No vendor lock in, no cloud lock in this particular- >> So Vasanth, maybe you could talk a little bit more about Atlas and some of the differentiable features and things that you can do with Atlas that maybe people don't know about. >> Yeah, sure Dave like, Atlas is not just a manage database as a service, you know, like it's a complete data platform and it provides many features. Like for example, you build an application and probably down the line of three years, the data which you captured three years back might be an old data. Like how do you do it? Like there's no need for you to manually purge or do thing. Like we do have an online archival where you configure the data. So that like the data, which is older than two years, just purge it. So automatically this is taken care. So that like you have hot data kept in Atlas cluster and the cold data moved up to an ARKit. And also like we have a data lake where you can run a federated queries . For example, you've done an archival, but what if people want to access the data? So with data lake, what it can do is, on a single connection, you can fire a- you can run a federated queries both on the active and the archival data. That's the beauty, like you archive the data, but still you can able to query it. And we do also have a charts where like, you can build in visualization on top of the data, what you have captured. You can build in graphs or you can build in graphs and also embed these graphs as part of your application, or you can collaborate to the customers, to the CXOs and other theme. >> Dave: Got it. >> It's a complete data platform. >> Okay. Well, speaking of data platform, let's talk about Io-Tahoe's data RPA platform, and coupling that with Mongo DB. So maybe you could help us understand how you're helping with process automation, which is a very hot topic and just this whole notion of a modern application development. >> Sure. See, the process automation is more with respect to the data and how you manage this data and what to derive and build a business process on top of it. I see there are two parts into it. Like one is the source of data. How do you identify, how do you discover the data? How do you enrich the context or transform it, give a business context to it. And then you build a business rules or act on it, and then you store the data or you derive the insights or enrich it and store it into DB. The first part is completely taken by Io-Tahoe, where you can tag the data for the multiple data sources. For example, if we take an customer 360 view, you can grab the data from multiple data sources using Io-Tahoe and you discover this data, you can tag it, you can label it and you build a view of the complete customer context, and use a realm web book and then the data is ingested back to Mongo. So that's all like more sort of like server-less fashion. You can build this particular customer 360 view for example. And just to talk about the realm I spoke, right? The realm web book, realm is a backend APA that you can create on top of the data on Mongo cluster, which is available in addclass. Okay. Then once you run, the APS are ready. Data as a service, you build it as a data as a service, and you fully secure APIs, which are available. These APS can be integrated within a mobile app or an web application to build in a built in modern application. But what left out is like, just build a UI artifacts and integrate these APIs. >> Yeah, I mean we live in this API economy companies. People throw that out as sort of a buzz phrase, but Mongo lives that. I mean, that's why developers really like the Mongo. So what's your take on DevOps? Maybe you could talk a little bit about, you know, your perspective there, how you help Devs and data engineers build faster pipelines. >> Yeah, sure. Like, okay, this is the most favorite topic. Like, no, and it's a buzzword along, like all the DevOps moving out from the traditional deployment, what I learned online. So like we do support like the deployment automation in multiple ways okay, and also provide the diagnostic under the hood. We have two options in Mongo DB. One is an enterprise option, which is more on the on-prem's version. And Atlas is more with respect to the cloud one manage database service. Okay. In case of an enterprise advanced, like we do have an Ops manager and the Kubernetes operator, like a Ops manager will manage all sort of deployment automation. Upgrades, provides your diagnostics, both with respect to the hardwares, and also with respect to the MongoDB gives you a profiling, slow running queries and what you can get a context of what's working on the data using that. I'm using an enterprise operator. You can integrate with existing Kubernetes cluster, either in a different namespace on an existing namespace. And orchestrate the deployment. And in case of Atlas, we do have an Atlas-Kubernetes operator, which helps you to integrate your Kubernetes operator. And you don't need to leave your Kubernetes. And also we have worked with the cloud providers. For example, we have we haven't cloud formation templates where you can just in one click, you can just roll out an Atlas cluster with a complete platform. So that's one, like we are continuously working, evolving on the DevOps site to roll out the might be a helm chart, or we do have an operator, which has a standard (indistinct) for different types of deployments. >> You know, some really important themes here. Obviously, anytime you talk about Mongo, simplicity comes in, automation, you know, that big, big push that Io-Tahoe was making. What you said about data context was interesting because a lot of data systems, organizations, they lack context and context is very important. So auto classification and things like that. And the other thing you said about federated queries I think fits very well into the trend toward decentralized data architecture. So very important there. And of course, hybridisity. I call it hybridisity. On-prem, cloud, abstracting that complexity away and allowing people to really focus on their digital transformations. I tell ya, Vasanth, it's great stuff. It's always a pleasure chatting with Io-Tahoe partners, and really getting into the tech with folks like yourself. So thanks so much for coming on theCube. >> Thanks. Thanks, Dave. Thanks for having a nice discussion with you. >> Okay. Stay right there. We've got one more quick session that you don't want to miss.
SUMMARY :
Okay. We're here with Vasanth Kumar you have experience in of handling the data and and back in the early days And then you can store it quickly. So how do you capture that? And then of course, you know, on the previous question. That's how the adoption has happened. you guys actually have So that like, you don't So Vasanth, maybe you could talk the data which you So maybe you could help us and then you store the data little bit about, you know, and what you can get a context And the other thing you discussion with you. that you don't want to miss.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Vasanth Kumar | PERSON | 0.99+ |
Mongo | ORGANIZATION | 0.99+ |
Dave | PERSON | 0.99+ |
two parts | QUANTITY | 0.99+ |
Vasanth | PERSON | 0.99+ |
35% | QUANTITY | 0.99+ |
tomorrow | DATE | 0.99+ |
last year | DATE | 0.99+ |
115 billion | QUANTITY | 0.99+ |
first part | QUANTITY | 0.99+ |
Bangalore, India | LOCATION | 0.99+ |
three years | QUANTITY | 0.99+ |
JSON | TITLE | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Io-Tahoe | ORGANIZATION | 0.99+ |
80-90 billion | QUANTITY | 0.99+ |
MongoDB | ORGANIZATION | 0.99+ |
ARKit | TITLE | 0.98+ |
two options | QUANTITY | 0.98+ |
one click | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
Atlas | TITLE | 0.98+ |
this year | DATE | 0.98+ |
both | QUANTITY | 0.97+ |
older than two years | QUANTITY | 0.97+ |
around 18 years | QUANTITY | 0.96+ |
One | QUANTITY | 0.96+ |
nearly two years | QUANTITY | 0.96+ |
Azure | TITLE | 0.96+ |
Wiki Bon | ORGANIZATION | 0.96+ |
MongoDB | TITLE | 0.95+ |
three years back | DATE | 0.94+ |
Vasanth | ORGANIZATION | 0.93+ |
Io-Tahoe | TITLE | 0.92+ |
DevOps | TITLE | 0.91+ |
Atlas | ORGANIZATION | 0.91+ |
Kubernetes | TITLE | 0.91+ |
360 view | QUANTITY | 0.89+ |
one | QUANTITY | 0.89+ |
single connection | QUANTITY | 0.88+ |
five years | QUANTITY | 0.85+ |
one more quick session | QUANTITY | 0.83+ |
GCP | ORGANIZATION | 0.83+ |
four cloud vendors | QUANTITY | 0.82+ |
GCP | TITLE | 0.79+ |
three major cloud providers | QUANTITY | 0.76+ |
one more point | QUANTITY | 0.73+ |
Io | TITLE | 0.72+ |
Azure | ORGANIZATION | 0.72+ |
-Tahoe | ORGANIZATION | 0.68+ |
four | QUANTITY | 0.67+ |
Mongo DB | TITLE | 0.65+ |
APA | TITLE | 0.58+ |
ISBs | ORGANIZATION | 0.54+ |
Tiji Mathew, Patrick Zimet and Senthil Karuppaiah | Io-Tahoe Data Quality Active DQ
(upbeat music), (logo pop up) >> Narrator: From around the globe it's theCUBE. Presenting active DQ intelligent automation for data quality brought to you by IO-Tahoe. >> Are you ready to see active DQ on Snowflake in action? Let's get into the show and tell him, do the demo. With me or Tiji Matthew, the Data Solutions Engineer at IO-Tahoe. Also joining us is Patrick Zeimet Data Solutions Engineer at IO-Tahoe and Senthilnathan Karuppaiah, who's the Head of Production Engineering at IO-Tahoe. Patrick, over to you let's see it. >> Hey Dave, thank you so much. Yeah, we've seen a huge increase in the number of organizations interested in Snowflake implementation. Were looking for an innovative, precise and timely method to ingest their data into Snowflake. And where we are seeing a lot of success is a ground up method utilizing both IO-Tahoe and Snowflake. To start you define your as is model. By leveraging IO-Tahoe to profile your various data sources and push the metadata to Snowflake. Meaning we create a data catalog within Snowflake for a centralized location to document items such as source system owners allowing you to have those key conversations and understand the data's lineage, potential blockers and what data is readily available for ingestion. Once the data catalog is built you have a much more dynamic strategies surrounding your Snowflake ingestion. And what's great is that while you're working through those key conversations IO-Tahoe will maintain that metadata push and partnered with Snowflake ability to version the data. You can easily incorporate potential scheme changes along the way. Making sure that the information that you're working on stays as current as the systems that you're hoping to integrate with Snowflake. >> Nice, Patrick I wonder if you could address how you IO-Tahoe Platform Scales and maybe in what way it provides a competitive advantage for customers. >> Great question where IO-Tahoe shines is through its active DQ or the ability to monitor your data's quality in real time. Marking which roads need remediation. According to the customized business rules that you can set. Ensuring that the data quality standards meet the requirements of your organizations. What's great is through our use of RPA. We can scale with an organization. So as you ingest more data sources we can allocate more robotic workers meaning the results will continue to be delivered in the same timely fashion you've grown used to. What's Morrisons IO-Tahoe is doing the heavy lifting on monitoring data quality. That's frees up your data experts to focus on the more strategic tasks such as remediation that augmentations and analytics developments. >> Okay, maybe Tiji, you could address this. I mean, how does all this automation change the operating model that we were talking to to Aj and Dunkin before about that? I mean, if it involves less people and more automation what else can I do in parallel? >> I'm sure the participants today will also be asking the same question. Let me start with the strategic tasks Patrick mentioned, Io-Tahoe does the heavy lifting. Freeing up data experts to act upon the data events generated by IO-Tahoe. Companies that have teams focused on manually building their inventory of the data landscape. Leads to longer turnaround times in producing actionable insights from their own data assets. Thus, diminishing the value realized by traditional methods. However, our operating model involves profiling and remediating at the same time creating a catalog data estate that can be used by business or IT accordingly. With increased automation and fewer people. Our machine learning algorithms augment the data pipeline to tag and capture the data elements into a comprehensive data catalog. As IO-Tahoe automatically catalogs the data estate in a centralized view, the data experts can partly focus on remediating the data events generated from validating against business rules. We envision that data events coupled with this drillable and searchable view will be a comprehensive one to assess the impact of bad quality data. Let's briefly look at the image on screen. For example, the view indicates that bad quality zip code data impacts the contact data which in turn impacts other related entities in systems. Now contrast that with a manually maintained spreadsheet that drowns out the main focus of your analysis. >> Tiji, how do you tag and capture bad quality data and stop that from you've mentioned these printed dependencies. How do you stop that from flowing downstream into the processes within the applications or reports? >> As IO-Tahoe builds the data catalog across source systems. We tag the elements that meet the business rule criteria while segregating the failed data examples associated with the elements that fall below a certain threshold. The elements that meet the business rule criteria are tagged to be searchable. Thus, providing an easy way to identify data elements that may flow through the system. The segregated data examples on the other hand are used by data experts to triage for the root cause. Based on the root cause potential outcomes could be one, changes in the source system to prevent that data from entering the system in the first place. Two, add data pipeline logic, to sanitize bad data from being consumed by downstream applications and reports or just accept the risk of storing bad data and address it when it meets a certain threshold. However, Dave as for your question about preventing bad quality data from flowing into the system? IO-Tahoe will not prevent it because the controls of data flowing between systems is managed outside of IO-Tahoe. Although, IO-Tahoe will alert and notify the data experts to events that indicate bad data has entered the monitored assets. Also we have redesigned our product to be modular and extensible. This allows data events generated by IO-Tahoe to be consumed by any system that wants to control the targets from bad data. Does IO-Tahoe empowers the data experts to control the bad data from flowing into their system. >> Thank you for that. So, one of the things that we've noticed, we've written about is that you've got these hyper specialized roles within the data, the centralized data organization. And wonder how do the data folks get involved here if at all, and how frequently do they get involved? Maybe Senthilnathan you could take that. >> Thank you, Dave for having me here. Well, based on whether the data element in question is in data cataloging or monitoring phase. Different data folks gets involved. When it isn't in the data cataloging stage. The data governance team, along with enterprise architecture or IT involved in setting up the data catalog. Which includes identifying the critical data elements business term identification, definition, documentation data quality rules, and data even set up data domain and business line mapping, lineage PA tracking source of truth. So on and so forth. It's typically in one time set up review certify then govern and monitor. But while when it is in the monitoring phase during any data incident or data issues IO-Tahoe broadcast data signals to the relevant data folks to act and remedy it as quick as possible. And alerts the consumption team it could be the data science, analytics, business opts are both a potential issue so that they are aware and take necessary preventative measure. Let me show you an example, critical data element from data quality dashboard view to lineage view to data 360 degree view for a zip code for conformity check. So in this case the zip code did not meet the past threshold during the technical data quality check and was identified as non-compliant item and notification was sent to the ID folks. So clicking on the zip code. Will take to the lineage view to visualize the dependent system, says that who are producers and who are the consumers. And further drilling down will take us to the detailed view, that a lot of other information's are presented to facilitate for a root cause analysis and not to take it to a final closure. >> Thank you for that. So Tiji? Patrick was talking about the as is to be. So I'm interested in how it's done now versus before. Do you need a data governance operating model for example? >> Typically a company that decides to make an inventory of the data assets would start out by manually building a spreadsheet managed by data experts of the company. What started as a draft now get break into the model of a company. This leads to loss of collaboration as each department makes a copy of their catalog for their specific needs. This decentralized approach leads to loss of uniformity which each department having different definitions which ironically needs a governance model for the data catalog itself. And as the spreadsheet grows in complexity the skill level needed to maintain. It also increases thus leading to fewer and fewer people knowing how to maintain it. About all the content that took so much time and effort to build is not searchable outside of that spreadsheet document. >> Yeah, I think you really hit the nail on my head Tiji. Now companies want to move away from the spreadsheet approach. IO-Tahoe addresses the shortcoming of the traditional approach enabling companies to achieve more with less. >> Yeah, what the customer reaction has been. We had Webster Bank, on one of the early episodes for example, I mean could they have achieved. What they did without something like active data quality and automation maybe Senthilnathan you could address that? >> Sure, It is impossible to achieve full data quality monitoring and remediation without automation or digital workers in place reality that introverts they don't have the time to do the remediation manually because they have to do an analysis conform fix on any data quality issues, as fast as possible before it gets bigger and no exception to Webster. That's why Webster implemented IO-Tahoe's active DQ to set up the business, metadata management and data quality monitoring and remediation in the Snowflake cloud data Lake. We help and building the center of excellence in the data governance, which is managing the data catalog schedule on demand and in-flight data quality checks, but Snowflake, no pipe on stream are super beneficial to achieve in flight quality checks. Then the data assumption monitoring and reporting last but not the least the time saver is persisting the non-compliant records for every data quality run within the Snowflake cloud, along with remediation script. So that during any exceptions the respect to team members is not only alerted. But also supplied with necessary scripts and tools to perform remediation right from the IO-Tahoe's Active DQ. >> Very nice. Okay guys, thanks for the demo. Great stuff. Now, if you want to learn more about the IO-Tahoe platform and how you can accelerate your adoption of Snowflake book some time with a data RPA expert all you got to do is click on the demo icon on the right of your screen and set a meeting. We appreciate you attending this latest episode of the IO-Tahoe data automation series. Look, if you missed any of the content that's all available on demand. This is Dave Vellante theCUBE. Thanks for watching. (upbeat music)
SUMMARY :
the globe it's theCUBE. and tell him, do the demo. and push the metadata to Snowflake. if you could address or the ability to monitor the operating model on remediating the data events generated into the processes within the data experts to events that indicate So, one of the things that So clicking on the zip code. Thank you for that. the skill level needed to maintain. of the traditional approach one of the early episodes So that during any exceptions the respect of the IO-Tahoe data automation series.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Patrick | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Tiji Matthew | PERSON | 0.99+ |
Tiji Mathew | PERSON | 0.99+ |
Senthil Karuppaiah | PERSON | 0.99+ |
Patrick Zimet | PERSON | 0.99+ |
IO-Tahoe | ORGANIZATION | 0.99+ |
Io-Tahoe | ORGANIZATION | 0.99+ |
Tiji | PERSON | 0.99+ |
360 degree | QUANTITY | 0.99+ |
Senthilnathan Karuppaiah | PERSON | 0.99+ |
each department | QUANTITY | 0.99+ |
Snowflake | TITLE | 0.99+ |
today | DATE | 0.99+ |
Webster | ORGANIZATION | 0.99+ |
Aj | PERSON | 0.99+ |
Dunkin | PERSON | 0.98+ |
Two | QUANTITY | 0.98+ |
IO | ORGANIZATION | 0.97+ |
Patrick Zeimet | PERSON | 0.97+ |
Webster Bank | ORGANIZATION | 0.97+ |
one | QUANTITY | 0.97+ |
one time | QUANTITY | 0.97+ |
both | QUANTITY | 0.96+ |
Senthilnathan | PERSON | 0.96+ |
IO-Tahoe | TITLE | 0.93+ |
first place | QUANTITY | 0.89+ |
IO | TITLE | 0.72+ |
Snowflake | EVENT | 0.71+ |
Tahoe | ORGANIZATION | 0.69+ |
Data Solutions | ORGANIZATION | 0.69+ |
-Tahoe | TITLE | 0.64+ |
Tahoe | TITLE | 0.63+ |
Snowflake | ORGANIZATION | 0.6+ |
Morrisons | ORGANIZATION | 0.6+ |
Ajay Vohora and Duncan Turnbull | Io-Tahoe ActiveDQ Intelligent Automation for Data Quality
>>From around the globe, but it's the cube presenting active DQ, intelligent automation for data quality brought to you by IO Tahoe. >>Now we're going to look at the role automation plays in mobilizing your data on snowflake. Let's welcome. And Duncan Turnbull who's partner sales engineer at snowflake and AIG Vihara is back CEO of IO. Tahoe is going to share his insight. Gentlemen. Welcome. >>Thank you, David. Good to have you back. Yeah, it's great to have you back >>A J uh, and it's really good to CIO Tao expanding the ecosystem so important. Um, now of course bringing snowflake and it looks like you're really starting to build momentum. I mean, there's progress that we've seen every month, month by month, over the past 12, 14 months, your seed investors, they gotta be happy. >>They are all that happy. And then I can see that we run into a nice phase of expansion here and new customers signing up. And now you're ready to go out and raise that next round of funding. I think, um, maybe think of a slight snowflake five years ago. So we're definitely on track with that. A lot of interest from investors and, um, we're right now trying to focus in on those investors that can partner with us, understand AI data and, and automation. >>So personally, I mean, you've managed a number of early stage VC funds. I think four of them, uh, you've taken several comp, uh, software companies through many funding rounds and growth and all the way to exit. So, you know how it works, you have to get product market fit, you know, you gotta make sure you get your KPIs, right. And you gotta hire the right salespeople, but, but what's different this time around, >>Uh, well, you know, the fundamentals that you mentioned though, those are never change. And, um, what we can say, what I can say that's different, that's shifted, uh, this time around is three things. One in that they used to be this kind of choice of, do we go open source or do we go proprietary? Um, now that has turned into, um, a nice hybrid model where we've really keyed into, um, you know, red hat doing something similar with Santos. And the idea here is that there is a core capability of technology that independence a platform, but it's the ability to then build an ecosystem around that made a pervade community. And that community may include customers, uh, technology partners, other tech vendors, and enabling the platform adoption so that all of those folks in that community can build and contribute, um, while still maintaining the core architecture and platform integrity, uh, at the core of it. >>And that's one thing that's changed was fitting a lot of that type of software company, um, emerge into that model, which is different from five years ago. Um, and then leveraging the cloud, um, every cloud snowflake cloud being one of them here in order to make use of what customers, uh, and customers and enterprise software are moving towards. Uh, every CIO is now in some configuration of a hybrid. Um, it is state whether those cloud multi-cloud on prem. That's just the reality. The other piece is in dealing with the CIO is legacy. So the past 15, 20 years they've purchased many different platforms, technologies, and some of those are still established and still, how do you, um, enable that CIO to make purchase while still preserving and in some cases building on and extending the, the legacy, um, material technology. So they've invested their people's time and training and financial investment into solving a problem, customer pain point, uh, with technology, but, uh, never goes out of fashion >>That never changes. You have to focus like a laser on that. And of course, uh, speaking of companies who are focused on solving problems, don't can turn bill from snowflake. You guys have really done a great job and really brilliantly addressing pain points, particularly around data warehousing, simplified that you're providing this new capability around data sharing, uh, really quite amazing. Um, Dunkin AAJ talks about data quality and customer pain points, uh, in, in enterprise. It, why is data quality been such a problem historically? >>Oh, sorry. One of the biggest challenges that's really affected by it in the past is that because to address everyone's need for using data, they've evolved all these kinds of different places to store all these different silos or data marts or all this kind of clarification of places where data lives and all of those end up with slightly different schedules to bringing data in and out. They end up with slightly different rules for transforming that data and formatting it and getting it ready and slightly different quality checks for making use of it. And this then becomes like a big problem in that these different teams are then going to have slightly different or even radically different ounces to the same kinds of questions, which makes it very hard for teams to work together, uh, on their different data problems that exist inside the business, depending on which of these silos they end up looking at and what you can do. If you have a single kind of scalable system for putting all of your data into it, you can kind of sidestep along to this complexity and you can address the data quality issues in a, in a single and a single way. >>Now, of course, we're seeing this huge trend in the market towards robotic process automation, RPA, that adoption is accelerating. Uh, you see, in UI paths, I IPO, you know, 35 plus billion dollars, uh, valuation, you know, snowflake like numbers, nice cops there for sure. Uh, agent you've coined the phrase data RPA, what is that in simple terms? >>Yeah, I mean, it was born out of, uh, seeing how in our ecosystem concern community developers and customers, uh, general business users for wanting to adopt and deploy a tar hose technology. And we could see that, um, I mean, there's not monkeying out PA we're not trying to automate that piece, but wherever there is a process that was tied into some form of a manual overhead with handovers and so on. Um, that process is something that we were able to automate with, with our ties technology and, and the deployment of AI and machine learning technologies specifically to those data processes almost as a precursor to getting into financial automation that, um, that's really where we're seeing the momentum pick up, especially in the last six months. And we've kept it really simple with snowflake. We've kind of stepped back and said, well, you know, the resource that a snowflake can leverage here is, is the metadata. So how could we turn snowflake into that repository of being the data catalog? And by the way, if you're a CIO looking to purchase a data catalog tool stop, there's no need to, um, working with snowflake, we've enable that intelligence to be gathered automatically and to be put, to use within snowflake. So reducing that manual effort, and I'm putting that data to work. And, um, and that's where we've packaged this with, uh, AI machine learning specific to those data tasks. Um, and it made sense that's, what's resonated with, with our customers. >>You know, what's interesting here, just a quick aside, as you know, I've been watching snowflake now for awhile and, and you know, of course the, the competitors come out and maybe criticize why they don't have this feature. They don't have that feature. And it's snowflake seems to have an answer. And the answer oftentimes is, well, its ecosystem ecosystem is going to bring that because we have a platform that's so easy to work with though. So I'm interested Duncan in what kind of collaborations you are enabling with high quality data. And of course, you know, your data sharing capability. >>Yeah. So I think, uh, you know, the ability to work on, on datasets, isn't just limited to inside the business itself or even between different business units. And we were kind of discussing maybe with their silos. Therefore, when looking at this idea of collaboration, we have these where we want to be >>Able to exploit data to the greatest degree possible, but we need to maintain the security, the safety, the privacy, and governance of that data. It could be quite valuable. It could be quite personal depending on the application involved. One of these novel applications that we see between organizations of data sharing is this idea of data clean rooms. And these data clean rooms are safe, collaborative spaces, which allow multiple companies or even divisions inside a company where they have particular, uh, privacy requirements to bring two or more data sets together for analysis. But without having to actually share the whole unprotected data set with each other, and this lets you to, you know, when you do this inside of snowflake, you can collaborate using standard tool sets. You can use all of our SQL ecosystem. You can use all of the data science ecosystem that works with snowflake. >>You can use all of the BI ecosystem that works with snowflake, but you can do that in a way that keeps the confidentiality that needs to be presented inside the data intact. And you can only really do these kinds of, uh, collaborations, especially across organization, but even inside large enterprises, when you have good reliable data to work with, otherwise your analysis just isn't going to really work properly. A good example of this is one of our large gaming customers. Who's an advertiser. They were able to build targeting ads to acquire customers and measure the campaign impact in revenue, but they were able to keep their data safe and secure while doing that while working with advertising partners, uh, the business impact of that was they're able to get a lifted 20 to 25% in campaign effectiveness through better targeting and actually, uh, pull through into that of a reduction in customer acquisition costs because they just didn't have to spend as much on the forms of media that weren't working for them. >>So, ha I wonder, I mean, you know, with, with the way public policy shaping out, you know, obviously GDPR started it in the States, you know, California, consumer privacy act, and people are sort of taking the best of those. And, and, and there's a lot of differentiation, but what are you seeing just in terms of, you know, the government's really driving this, this move to privacy, >>Um, government public sector, we're seeing a huge wake up an activity and, uh, across the whole piece that, um, part of it has been data privacy. Um, the other part of it is being more joined up and more digital rather than paper or form based. Um, we've all got stories of waiting in line, holding a form, taking that form to the front of the line and handing it over a desk. Now government and public sector is really looking to transform their services into being online, to show self service. Um, and that whole shift is then driving the need to, um, emulate a lot of what the commercial sector is doing, um, to automate their processes and to unlock the data from silos to put through into those, uh, those processes. Um, and another thing I can say about this is they, the need for data quality is as a Dunkin mentions underpins all of these processes, government pharmaceuticals, utilities, banking, insurance, the ability for a chief marketing officer to drive a, a loyalty campaign. >>They, the ability for a CFO to reconcile accounts at the end of the month. So do a, a, uh, a quick, accurate financial close. Um, also the, the ability of a customer operations to make sure that the customer has the right details about themselves in the right, uh, application that they can sell. So from all of that is underpinned by data and is effective or not based on the quality of that data. So whilst we're mobilizing data to snowflake cloud, the ability to then drive analytics, prediction, business processes off that cloud, um, succeeds or fails on the quality of that data. >>I mean it, and, you know, I would say, I mean, it really is table stakes. If you don't trust the data, you're not gonna use the data. The problem is it always takes so long to get to the data quality. There's all these endless debates about it. So we've been doing a fair amount of work and thinking around this idea of decentralized data, data by its very nature is decentralized, but the fault domains of traditional big data is that everything is just monolithic and the organizations monolithic technology's monolithic, the roles are very, you know, hyper specialized. And so you're hearing a lot more these days about this notion of a data fabric or what calls a data mesh. Uh, and we've kind of been leaning in to that and the ability to, to connect various data capabilities, whether it's a data warehouse or a data hub or a data Lake that those assets are discoverable, they're shareable through API APIs and they're governed on a federated basis. And you're using now bringing in a machine intelligence to improve data quality. You know, I wonder Duncan, if you could talk a little bit about Snowflake's approach to this topic. >>Sure. So I'd say that, you know, making use of all of your data, is there a key kind of driver behind these ideas that they can mesh into the data fabrics? And the idea is that you want to bring together not just your kind of strategic data, but also your legacy data and everything that you have inside the enterprise. I think I'd also like to kind of expand upon what a lot of people view as all of the data. And I think that a lot of people kind of miss that there's this whole other world of data they could be having access to, which is things like data from their business partners, their customers, their suppliers, and even stuff that's more in the public domain, whether that's, you know, demographic data or geographic or all these kinds of other types of data sources. And what I'd say to some extent is that the data cloud really facilitates the ability to share and gain access to this both kind of between organizations inside organizations. >>And you don't have to, you know, make lots of copies of the data and kind of worry about the storage and this federated, um, you know, idea of governance and all these things that it's quite complex to kind of manage this. Uh, you know, the snowflake approach really enables you to share data with your ecosystem all the world, without any latency with full control over what's shared without having to introduce new complexities or having complex attractions with APIs or software integration. The simple approach that we provide allows a relentless focus on creating the right data product to meet the challenges facing your business today. >>So, Andrea, the key here is to don't get to talking about it in my mind. Anyway, my cake takeaway is to simplicity. If you can take the complexity out of the equation, we're going to get more adoption. It really is that simple. >>Yeah, absolutely. Do you think that that whole journey, maybe five, six years ago, the adoption of data lakes was, was a stepping stone. Uh, however, the Achilles heel there was, you know, the complexity that it shifted towards consuming that data from a data Lake where there were many, many sets of data, um, to, to be able to cure rate and to, um, to consume, uh, whereas actually, you know, the simplicity of being able to go to the data that you need to do your role, whether you're in tax compliance or in customer services is, is key. And, you know, listen for snowflake by auto. One thing we know for sure is that our customers are super small and they're very capable. They're they're data savvy and know, want to use whichever tool and embrace whichever, um, cloud platform that is gonna reduce the barriers to solving. What's complex about that data, simplifying that and using, um, good old fashioned SQL, um, to access data and to build products from it to exploit that data. So, um, simplicity is, is key to it to allow people to, to, to make use of that data. And CIO is recognize that >>So Duncan, the cloud obviously brought in this notion of dev ops, um, and new methodologies and things like agile that brought that's brought in the notion of data ops, which is a very hot topic right now. Um, basically dev ops applies to data about how D how does snowflake think about this? How do you facilitate that methodology? >>Yeah, sorry. I agree with you absolutely. That they drops takes these ideas of agile development of >>Agile delivery and of the kind of dev ops world that we've seen just rise and rise, and it applies them to the data pipeline, which is somewhere where it kind of traditionally hasn't happened. And it's the same kinds of messages as we see in the development world, it's about delivering faster development, having better repeatability and really getting towards that dream of the data-driven enterprise, you know, where you can answer people's data questions, they can make better business decisions. And we have some really great architectural advantages that allow us to do things like allow cloning of data sets without having to copy them, allows us to do things like time travel so we can see what data looked like at some point in the past. And this lets you kind of set up both your own kind of little data playpen as a clone without really having to copy all of that data. >>So it's quick and easy, and you can also, again, with our separation of storage and compute, you can provision your own virtual warehouse for dev usage. So you're not interfering with anything to do with people's production usage of this data. So the, these ideas, the scalability, it just makes it easy to make changes, test them, see what the effect of those changes are. And we've actually seen this. You were talking a lot about partner ecosystems earlier. Uh, the partner ecosystem has taken these ideas that are inside snowflake and they've extended them. They've integrated them with, uh, dev ops and data ops tooling. So things like version control and get an infrastructure automation and things like Terraform. And they've kind of built that out into more of a data ops products that, that you can, you can make yourself so we can see there's a huge impact of, of these ideas coming into the data world. >>We think we're really well-placed to take advantage to them. The partner ecosystem is doing a great job with doing that. And it really allows us to kind of change that operating model for data so that we don't have as much emphasis on like hierarchy and change windows and all these kinds of things that are maybe use as a lot of fashioned. And we kind of taking the shift from this batch data integration into, you know, streaming continuous data pipelines in the cloud. And this kind of gets you away from like a once a week or once a month change window, if you're really unlucky to, you know, pushing changes, uh, in a much more rapid fashion as the needs of the business change. >>I mean, those hierarchical organizational structures, uh, w when we apply those to begin to that, what it actually creates the silos. So if you're going to be a silo Buster, which aji look at you guys in silo busters, you've got to put data in the hands of the domain experts, the business people, they know what data they want, if they have to go through and beg and borrow for a new data sets, et cetera. And so that's where automation becomes so key. And frankly, the technology should be an implementation detail, not the dictating factor. I wonder if you could comment on this. >>Yeah, absolutely. I think, um, making the, the technologies more accessible to the general business users >>Or those specialists business teams that, um, that's the key to unlocking is it is interesting to see is as people move from organization to organization where they've had those experiences operating in a hierarchical sense, I want to break free from that and, um, or have been exposed to, um, automation, continuous workflows, um, change is continuous in it. It's continuous in business, the market's continuously changing. So having that flow across the organization of work, using key components, such as get hub, similar to what you drive process Terraform to build in, um, code into the process, um, and automation and with a high Tahoe leveraging all the metadata from across those fragmented sources is, is, is good to say how those things are coming together. And watching people move from organization to organization say, Hey, okay, I've got a new start. I've got my first hundred days to impress my, my new manager. >>Uh, what kind of an impact can I, um, bring to this? And quite often we're seeing that as, you know, let me take away the good learnings from how to do it, or how not to do it from my previous role. And this is an opportunity for me to, to bring in automation. And I'll give you an example, David, you know, recently started working with a, a client in financial services. Who's an asset manager, uh, managing financial assets. They've grown over the course of the last 10 years through M and a, and each of those acquisitions have bought with it tactical data. It's saying instead of data of multiple CRM systems now multiple databases, multiple bespoke in-house created applications. And when the new CIO came in and had a look at those well, you know, yes, I want to mobilize my data. Yes, I need to modernize my data state because my CEO is now looking at these crypto assets that are on the horizon and the new funds that are emerging that around digital assets and crypto assets. >>But in order to get to that where absolutely data underpins and is the core asset, um, cleaning up that, that legacy situation mobilizing the relevant data into the Safelite cloud platform, um, is where we're giving time back, you know, that is now taking a few weeks, whereas that transitioned to mobilize that data, start with that, that new clean slate to build upon a new business as a, a digital crypto asset manager, as well as the legacy, traditional financial assets, bonds stocks, and fixed income assets, you name it, uh, is where we're starting to see a lot of innovation. >>Yeah. Tons of innovation. I love the crypto examples and FTS are exploding and, you know, let's face it, traditional banks are getting disrupted. Uh, and so I also love this notion of data RPA. I, especially because I've done a lot of work in the RPA space. And, and I want to, what I would observe is that the, the early days of RPA, I call it paving the cow path, taking existing processes and applying scripts, get letting software robots, you know, do its thing. And that was good because it reduced, you know, mundane tasks, but really where it's evolved is a much broader automation agenda. People are discovering new, new ways to completely transform their processes. And I see a similar, uh, analogy for data, the data operating model. So I'm wonder whenever you think about that, how a customer really gets started bringing this to their ecosystem, their data life cycles. >>Sure. Yeah. So step one is always the same is figuring out for the CIO, the chief data officer, what data do I have, um, and that's increasingly something that they want towards a mate, so we can help them there and, and do that automated data discovery, whether that is documents in the file, share backup archive in a relational data store, in a mainframe really quickly hydrating that and bringing that intelligence, the forefront of, of what do I have, and then it's the next step of, well, okay. Now I want to continually monitor and curate that intelligence with the platform that I've chosen. Let's say snowflake, um, in order such that I can then build applications on top of that platform to serve my, my internal, external customer needs and the automation around classifying data reconciliation across different fragmented data silos, building that in those insights into snowflake. >>Um, as you say, a little later on where we're talking about data quality, active DQ, allowing us to reconcile data from different sources, as well as look at the integrity of that data. Um, so they can go on to remediation, you know, I, I wanna, um, harness and leverage, um, techniques around traditional RPA. Um, but to get to that stage, I need to fix the data. So remediating publishing the data in snowflake, uh, allowing analysis to be formed performance snowflake. Th those are the key steps that we see and just shrinking that timeline into weeks, giving the organization that time back means they're spending more time on their customer and solving their customer's problem, which is where we want them to be. >>This is the brilliance of snowflake actually, you know, Duncan is, I've talked to him, then what does your view about this and your other co-founders and it's really that focus on simplicity. So, I mean, that's, you, you picked a good company to join my opinion. So, um, I wonder if you could, you know, talk about some of the industry sectors that are, again, going to gain the most from, from data RPA, I mean, traditional RPA, if I can use that term, you know, a lot of it was back office, a lot of, you know, financial w what are the practical applications where data RPA is going to impact, you know, businesses and, and the outcomes that we can expect. >>Yes, sir. So our drive is, is really to, to make that, um, business general user's experience of RPA simpler and, and using no code to do that, uh, where they've also chosen snowflake to build that their cloud platform. They've got the combination then of using a relatively simple script scripting techniques, such as SQL, uh, without no code approach. And the, the answer to your question is whichever sector is looking to mobilize their data. Uh, it seems like a cop-out, but to give you some specific examples, David, um, in banking where, uh, customers are looking to modernize their banking systems and enable better customer experience through, through applications and digital apps. That's where we're, we're seeing a lot of traction, uh, and this approach to, to pay RPA to data, um, health care, where there's a huge amount of work to do to standardize data sets across providers, payers, patients, uh, and it's an ongoing, um, process there for, for retail, um, helping to, to build that immersive customer experience. >>So recommending next best actions, um, providing an experience that is going to drive loyalty and retention, that's, that's dependent on understanding what that customer's needs intent, uh, being out to provide them with the content or the outfit at that point in time, or all data dependent utilities is another one great overlap there with, with snowflake where, you know, helping utilities, telecoms energy, water providers to build services on that data. And this is where the ecosystem just continues to, to expand. If we, if we're helping our customers turn their data into services for, for their ecosystem, that's, that's exciting. And they were more so exciting than insurance, which we always used to, um, think back to, uh, when insurance used to be very dull and mundane, actually, that's where we're seeing a huge amounts of innovation to create new flexible products that are priced to the day to the situation and, and risk models being adaptive when the data changes, uh, on, on events or circumstances. So across all those sectors that they're all mobilizing that data, they're all moving in some way, shape or form to a, a multi-cloud, um, set up with their it. And I think with, with snowflake and without Tahoe, being able to accelerate that and make that journey simple and as complex is, uh, is why we found such a good partner here. >>All right. Thanks for that. And then thank you guys. Both. We gotta leave it there. Uh, really appreciate Duncan you coming on and Aja best of luck with the fundraising. >>We'll keep you posted. Thanks, David. All right. Great. >>Okay. Now let's take a look at a short video. That's going to help you understand how to reduce the steps around your data ops. Let's watch.
SUMMARY :
intelligent automation for data quality brought to you by IO Tahoe. Tahoe is going to share his insight. Yeah, it's great to have you back Um, now of course bringing snowflake and it looks like you're really starting to build momentum. And then I can see that we run into a And you gotta hire the right salespeople, but, but what's different this time around, Uh, well, you know, the fundamentals that you mentioned though, those are never change. enable that CIO to make purchase while still preserving and in some And of course, uh, speaking of the business, depending on which of these silos they end up looking at and what you can do. uh, valuation, you know, snowflake like numbers, nice cops there for sure. We've kind of stepped back and said, well, you know, the resource that a snowflake can and you know, of course the, the competitors come out and maybe criticize why they don't have this feature. And we were kind of discussing maybe with their silos. the whole unprotected data set with each other, and this lets you to, you know, And you can only really do these kinds you know, obviously GDPR started it in the States, you know, California, consumer privacy act, insurance, the ability for a chief marketing officer to drive They, the ability for a CFO to reconcile accounts at the end of the month. I mean it, and, you know, I would say, I mean, it really is table stakes. extent is that the data cloud really facilitates the ability to share and gain access to this both kind Uh, you know, the snowflake approach really enables you to share data with your ecosystem all the world, So, Andrea, the key here is to don't get to talking about it in my mind. Uh, however, the Achilles heel there was, you know, the complexity So Duncan, the cloud obviously brought in this notion of dev ops, um, I agree with you absolutely. And this lets you kind of set up both your own kind So it's quick and easy, and you can also, again, with our separation of storage and compute, you can provision your own And this kind of gets you away from like a once a week or once a month change window, And frankly, the technology should be an implementation detail, not the dictating factor. the technologies more accessible to the general business users similar to what you drive process Terraform to build in, that as, you know, let me take away the good learnings from how to do um, is where we're giving time back, you know, that is now taking a And that was good because it reduced, you know, mundane tasks, that intelligence, the forefront of, of what do I have, and then it's the next step of, you know, I, I wanna, um, harness and leverage, um, This is the brilliance of snowflake actually, you know, Duncan is, I've talked to him, then what does your view about this and your but to give you some specific examples, David, um, the day to the situation and, and risk models being adaptive And then thank you guys. We'll keep you posted. That's going to help you understand how to reduce
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
David | PERSON | 0.99+ |
Andrea | PERSON | 0.99+ |
Duncan Turnbull | PERSON | 0.99+ |
Ajay Vohora | PERSON | 0.99+ |
Duncan | PERSON | 0.99+ |
20 | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
IO | ORGANIZATION | 0.99+ |
Both | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
first hundred days | QUANTITY | 0.99+ |
SQL | TITLE | 0.99+ |
both | QUANTITY | 0.99+ |
three things | QUANTITY | 0.98+ |
California | LOCATION | 0.98+ |
five years ago | DATE | 0.98+ |
one thing | QUANTITY | 0.98+ |
25% | QUANTITY | 0.97+ |
Terraform | ORGANIZATION | 0.97+ |
each | QUANTITY | 0.97+ |
one | QUANTITY | 0.96+ |
35 plus billion dollars | QUANTITY | 0.96+ |
five | DATE | 0.96+ |
Santos | ORGANIZATION | 0.96+ |
once a week | QUANTITY | 0.95+ |
GDPR | TITLE | 0.95+ |
Tahoe | PERSON | 0.95+ |
once a month | QUANTITY | 0.95+ |
consumer privacy act | TITLE | 0.94+ |
four | QUANTITY | 0.94+ |
step one | QUANTITY | 0.93+ |
IO Tahoe | ORGANIZATION | 0.93+ |
M | ORGANIZATION | 0.91+ |
agile | TITLE | 0.91+ |
last six months | DATE | 0.91+ |
14 months | QUANTITY | 0.9+ |
single | QUANTITY | 0.88+ |
six years ago | DATE | 0.88+ |
today | DATE | 0.88+ |
Io-Tahoe | ORGANIZATION | 0.87+ |
12 | QUANTITY | 0.84+ |
one of them | QUANTITY | 0.83+ |
AIG Vihara | ORGANIZATION | 0.82+ |
One thing | QUANTITY | 0.8+ |
single way | QUANTITY | 0.77+ |
last 10 years | DATE | 0.76+ |
Tons | QUANTITY | 0.75+ |
Agile | TITLE | 0.73+ |
years | QUANTITY | 0.73+ |
Tahoe | ORGANIZATION | 0.7+ |
Terraform | TITLE | 0.66+ |
every cloud | QUANTITY | 0.65+ |
Dunkin | ORGANIZATION | 0.61+ |
past 15, 20 | DATE | 0.59+ |
Tao | ORGANIZATION | 0.56+ |
Snowflake | ORGANIZATION | 0.56+ |
Safelite | ORGANIZATION | 0.54+ |
snowflake | TITLE | 0.53+ |
Dunkin AAJ | PERSON | 0.52+ |
people | QUANTITY | 0.51+ |
hat | ORGANIZATION | 0.5+ |
Glenn Grossman and Yusef Khan | Io-Tahoe ActiveDQ Intelligent Automation
>>from around the globe. It's the >>cube presenting >>active de que intelligent automation for data quality brought to you by Iota Ho >>Welcome to the sixth episode of the I. O. Tahoe data automation series. On the cube. We're gonna start off with a segment on how to accelerate the adoption of snowflake with Glenn Grossman, who is the enterprise account executive from Snowflake and yusef khan, the head of data services from Iota. Gentlemen welcome. >>Good afternoon. Good morning, Good evening. Dave. >>Good to see you. Dave. Good to see you. >>Okay glenn uh let's start with you. I mean the Cube hosted the snowflake data cloud summit in November and we heard from customers and going from love the tagline zero to snowflake, you know, 90 minutes very quickly. And of course you want to make it simple and attractive for enterprises to move data and analytics into the snowflake platform but help us understand once the data is there, how is snowflake helping to achieve savings compared to the data lake? >>Absolutely. dave. It's a great question, you know, it starts off first with the notion and uh kind of, we coined it in the industry or t shirt size pricing. You know, you don't necessarily always need the performance of a high end sports car when you're just trying to go get some groceries and drive down the street 20 mph. The t shirt pricing really aligns to, depending on what your operational workload is to support the business and the value that you need from that business? Not every day. Do you need data? Every second of the moment? Might be once a day, once a week through that t shirt size price and we can align for the performance according to the environmental needs of the business. What those drivers are the key performance indicators to drive that insight to make better decisions, It allows us to control that cost. So to my point, not always do you need the performance of a Ferrari? Maybe you need the performance and gas mileage of the Honda Civic if you would just get and deliver the value of the business but knowing that you have that entire performance landscape at a moments notice and that's really what what allows us to hold and get away from. How much is it going to cost me in a data lake type of environment? >>Got it. Thank you for that yussef. Where does Io Tahoe fit into this equation? I mean what's, what's, what's unique about the approach that you're taking towards this notion of mobilizing data on snowflake? >>Well, Dave in the first instance we profile the data itself at the data level, so not just at the level of metadata and we do that wherever that data lives. So it could be structured data could be semi structured data could be unstructured data and that data could be on premise. It could be in the cloud or it could be on some kind of SAAS platform. And so we profile this data at the source system that is feeding snowflake within snowflake itself within the end applications and the reports that the snowflake environment is serving. So what we've done here is take our machine learning discovery technology and make snowflake itself the repository for knowledge and insights on data. And this is pretty unique. Uh automation in the form of our P. A. Is being applied to the data both before after and within snowflake. And so the ultimate outcome is that business users can have a much greater degree of confidence that the data they're using can be trusted. Um The other thing we do uh which is unique is employee data R. P. A. To proactively detect and recommend fixes the data quality so that removes the manual time and effort and cost it takes to fix those data quality issues. Uh If they're left unchecked and untouched >>so that's key to things their trust, nobody's gonna use the data. It's not trusted. But also context. If you think about it, we've contextualized are operational systems but not our analytic system. So there's a big step forward glen. I wonder if you can tell us how customers are managing data quality when they migrate to snowflake because there's a lot of baggage in in traditional data warehouses and data lakes and and data hubs. Maybe you can talk about why this is a challenge for customers. And like for instance can you proactively address some of those challenges that customers face >>that we certainly can. They have. You know, data quality. Legacy data sources are always inherent with D. Q. Issues whether it's been master data management and data stewardship programs over the last really almost two decades right now, you do have systemic data issues. You have siloed data, you have information operational, data stores data marks. It became a hodgepodge when organizations are starting their journey to migrate to the cloud. One of the things that were first doing is that inspection of data um you know first and foremost even looking to retire legacy data sources that aren't even used across the enterprise but because they were part of the systemic long running operational on premise technology, it stayed there when we start to look at data pipelines as we onboard a customer. You know we want to do that era. We want to do QA and quality assurance so that we can, And our ultimate goal eliminate the garbage in garbage out scenarios that we've been plagued with really over the last 40, 50 years of just data in general. So we have to take an inspection where traditionally it was E. T. L. Now in the world of snowflake, it's really lt we're extracting were loading or inspecting them. We're transforming out to the business so that these routines could be done once and again give great business value back to making decisions around the data instead of spending all this long time. Always re architect ng the data pipeline to serve the business. >>Got it. Thank you. Glenda yourself of course. Snowflakes renowned for customers. Tell me all the time. It's so easy. It's so easy to spin up a data warehouse. It helps with my security. Again it simplifies everything but so you know, getting started is one thing but then adoption is also a key. So I'm interested in the role that that I owe. Tahoe plays in accelerating adoption for new customers. >>Absolutely. David. I mean as Ben said, you know every every migration to Snowflake is going to have a business case. Um uh and that is going to be uh partly about reducing spending legacy I. T. Servers, storage licenses, support all those good things um that see I want to be able to turn off entirely ultimately. And what Ayatollah does is help discover all the legacy undocumented silos that have been built up, as Glenn says on the data estate across a period of time, build intelligence around those silos and help reduce those legacy costs sooner by accelerating that that whole process. Because obviously the quicker that I. T. Um and Cdos can turn off legacy data sources the more funding and resources going to be available to them to manage the new uh Snowflake based data estate on the cloud. And so turning off the old building, the new go hand in hand to make sure those those numbers stack up the program is delivered uh and the benefits are delivered. And so what we're doing here with a Tahoe is improving the customers are y by accelerating their ability to adopt Snowflake. >>Great. And I mean we're talking a lot about data quality here but in a lot of ways that's table stakes like I said, if you don't trust the data, nobody's going to use it. And glenn, I mean I look at Snowflake and I see obviously the ease of use the simplicity you guys are nailing that the data sharing capabilities I think are really exciting because you know everybody talks about sharing data but then we talked about data as an asset, Everyone so high I to hold it. And so sharing is is something that I see as a paradigm shift and you guys are enabling that. So one of the things beyond data quality that are notable that customers are excited about that, maybe you're excited about >>David, I think you just cleared it out. It's it's this massive data sharing play part of the data cloud platform. Uh you know, just as of last year we had a little over about 100 people, 100 vendors in our data marketplace. That number today is well over 450 it is all about democratizing and sharing data in a world that is no longer held back by FTp s and C. S. V. S and then the organization having to take that data and ingested into their systems. You're a snowflake customer. want to subscribe to an S and P data sources an example, go subscribe it to it. It's in your account there was no data engineering, there was no physical lift of data and that becomes the most important thing when we talk about getting broader insights, data quality. Well, the data has already been inspected from your vendor is just available in your account. It's obviously a very simplistic thing to describe behind the scenes is what our founders have created to make it very, very easy for us to democratize not only internal with private sharing of data, but this notion of marketplace ensuring across your customers um marketplace is certainly on the type of all of my customers minds and probably some other areas that might have heard out of a recent cloud summit is the introduction of snow park and being able to do where all this data is going towards us. Am I in an ale, you know, along with our partners at Io Tahoe and R. P. A. Automation is what do we do with all this data? How do we put the algorithms and targets now? We'll be able to run in the future R and python scripts and java libraries directly inside Snowflake, which allows you to even accelerate even faster, Which people found traditionally when we started off eight years ago just as a data warehousing platform. >>Yeah, I think we're on the cusp of just a new way of thinking about data. I mean obviously simplicity is a starting point but but data by its very nature is decentralized. You talk about democratizing data. I like this idea of the global mesh. I mean it's very powerful concept and again it's early days but you know, keep part of this is is automation and trust, yussef you've worked with Snowflake and you're bringing active D. Q. To the market what our customers telling you so far? >>Well David the feedback so far has been great. Which is brilliant. So I mean firstly there's a point about speed and acceleration. Um So that's the speed to incite really. So where you have inherent data quality issues uh whether that's with data that was on premise and being brought into snowflake or on snowflake itself, we're able to show the customer results and help them understand their data quality better Within Day one which is which is a fantastic acceleration. I'm related to that. There's the cost and effort to get that insight is it's a massive productivity gain versus where you're seeing customers who've been struggling sometimes too remediate legacy data and legacy decisions that they've made over the past couple of decades, so that that cost and effort is much lower than it would otherwise have been. Um 3rdly, there's confidence and trust, so you can see Cdos and see IOS got demonstrable results that they've been able to improve data quality across a whole bunch of use cases for business users in marketing and customer services, for commercial teams, for financial teams. So there's that very quick kind of growth in confidence and credibility as the projects get moving. And then finally, I mean really all the use cases for the snowflake depend on data quality, really whether it's data science, uh and and the kind of snow park applications that Glenn has talked about, all those use cases work better when we're able to accelerate the ri for our joint customers by very quickly pushing out these data quality um insights. Um And I think one of the one of the things that the snowflake have recognized is that in order for C. I. O. Is to really adopt enterprise wide, um It's also as well as the great technology with Snowflake offers, it's about cleaning up that legacy data state, freeing up the budget for CIA to spend it on the new modern day to a state that lets them mobilise their data with snowflake. >>So you're seeing the Senate progression. We're simplifying the the the analytics from a tech perspective. You bring in Federated governance which which brings more trust. Then then you bring in the automation of the data quality piece which is fundamental. And now you can really start to, as you guys are saying, democratized and scale uh and share data. Very powerful guys. Thanks so much for coming on the program. Really appreciate your time. >>Thank you. I appreciate as well. Yeah.
SUMMARY :
It's the the head of data services from Iota. Good afternoon. Good to see you. I mean the Cube hosted the snowflake data cloud summit and the value that you need from that business? Thank you for that yussef. so not just at the level of metadata and we do that wherever that data lives. so that's key to things their trust, nobody's gonna use the data. Always re architect ng the data pipeline to serve the business. Again it simplifies everything but so you know, getting started is one thing but then I mean as Ben said, you know every every migration to Snowflake is going I see obviously the ease of use the simplicity you guys are nailing that the data sharing that might have heard out of a recent cloud summit is the introduction of snow park and I mean it's very powerful concept and again it's early days but you know, Um So that's the speed to incite And now you can really start to, as you guys are saying, democratized and scale uh and I appreciate as well.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
David | PERSON | 0.99+ |
Glenn Grossman | PERSON | 0.99+ |
Ben | PERSON | 0.99+ |
Io Tahoe | ORGANIZATION | 0.99+ |
Yusef Khan | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
20 mph | QUANTITY | 0.99+ |
Glenn | PERSON | 0.99+ |
CIA | ORGANIZATION | 0.99+ |
IOS | TITLE | 0.99+ |
Glenda | PERSON | 0.99+ |
90 minutes | QUANTITY | 0.99+ |
100 vendors | QUANTITY | 0.99+ |
Ferrari | ORGANIZATION | 0.99+ |
last year | DATE | 0.99+ |
One | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
first instance | QUANTITY | 0.99+ |
November | DATE | 0.99+ |
sixth episode | QUANTITY | 0.99+ |
once a day | QUANTITY | 0.99+ |
once a week | QUANTITY | 0.98+ |
Senate | ORGANIZATION | 0.98+ |
today | DATE | 0.98+ |
both | QUANTITY | 0.98+ |
eight years ago | DATE | 0.97+ |
yusef khan | PERSON | 0.97+ |
over | QUANTITY | 0.96+ |
one | QUANTITY | 0.95+ |
R. P. A. Automation | ORGANIZATION | 0.95+ |
python | TITLE | 0.95+ |
Tahoe | ORGANIZATION | 0.94+ |
I. O. Tahoe | TITLE | 0.93+ |
Honda | ORGANIZATION | 0.93+ |
Io-Tahoe | ORGANIZATION | 0.93+ |
one thing | QUANTITY | 0.91+ |
Io Tahoe | PERSON | 0.87+ |
firstly | QUANTITY | 0.87+ |
Civic | COMMERCIAL_ITEM | 0.87+ |
Snowflake | TITLE | 0.86+ |
Tahoe | PERSON | 0.85+ |
Ayatollah | PERSON | 0.84+ |
Snowflake | EVENT | 0.83+ |
past couple of decades | DATE | 0.82+ |
about 100 people | QUANTITY | 0.81+ |
two decades | QUANTITY | 0.8+ |
over 450 | QUANTITY | 0.79+ |
40, 50 years | QUANTITY | 0.76+ |
Day one | QUANTITY | 0.75+ |
glenn | PERSON | 0.74+ |
java | TITLE | 0.72+ |
snowflake | EVENT | 0.7+ |
Iota Ho | ORGANIZATION | 0.68+ |
P. | ORGANIZATION | 0.62+ |
ActiveDQ Intelligent Automation | ORGANIZATION | 0.61+ |
snowflake data cloud summit | EVENT | 0.6+ |
Iota | LOCATION | 0.58+ |
FTp | TITLE | 0.56+ |
Snowflake | ORGANIZATION | 0.54+ |
zero | QUANTITY | 0.53+ |
R | TITLE | 0.52+ |
O. | EVENT | 0.41+ |
C. | EVENT | 0.34+ |
Tiji Mathew, Patrick Zimet and Senthil Karuppaiah | Io-Tahoe Data Quality: Active DQ
(upbeat music), (logo pop up) >> Narrator: From around the globe it's theCUBE. Presenting active DQ intelligent automation for data quality brought to you by IO-Tahoe. >> Are you ready to see active DQ on Snowflake in action? Let's get into the show and tell him, do the demo. With me or Tiji Matthew, the Data Solutions Engineer at IO-Tahoe. Also joining us is Patrick Zeimet Data Solutions Engineer at IO-Tahoe and Senthilnathan Karuppaiah, who's the Head of Production Engineering at IO-Tahoe. Patrick, over to you let's see it. >> Hey Dave, thank you so much. Yeah, we've seen a huge increase in the number of organizations interested in Snowflake implementation. Were looking for an innovative, precise and timely method to ingest their data into Snowflake. And where we are seeing a lot of success is a ground up method utilizing both IO-Tahoe and Snowflake. To start you define your as is model. By leveraging IO-Tahoe to profile your various data sources and push the metadata to Snowflake. Meaning we create a data catalog within Snowflake for a centralized location to document items such as source system owners allowing you to have those key conversations and understand the data's lineage, potential blockers and what data is readily available for ingestion. Once the data catalog is built you have a much more dynamic strategies surrounding your Snowflake ingestion. And what's great is that while you're working through those key conversations IO-Tahoe will maintain that metadata push and partnered with Snowflake ability to version the data. You can easily incorporate potential scheme changes along the way. Making sure that the information that you're working on stays as current as the systems that you're hoping to integrate with Snowflake. >> Nice, Patrick I wonder if you could address how you IO-Tahoe Platform Scales and maybe in what way it provides a competitive advantage for customers. >> Great question where IO-Tahoe shines is through its active DQ or the ability to monitor your data's quality in real time. Marking which roads need remediation. According to the customized business rules that you can set. Ensuring that the data quality standards meet the requirements of your organizations. What's great is through our use of RPA. We can scale with an organization. So as you ingest more data sources we can allocate more robotic workers meaning the results will continue to be delivered in the same timely fashion you've grown used to. What's Morrisons IO-Tahoe is doing the heavy lifting on monitoring data quality. That's frees up your data experts to focus on the more strategic tasks such as remediation that augmentations and analytics developments. >> Okay, maybe Tiji, you could address this. I mean, how does all this automation change the operating model that we were talking to to Aj and Dunkin before about that? I mean, if it involves less people and more automation what else can I do in parallel? >> I'm sure the participants today will also be asking the same question. Let me start with the strategic task. Patrick mentioned IO-Tahoe does the heavy lifting. Freeing up data experts to act upon the data events generated by IO-Tahoe. Companies that have teams focused on manually building their inventory of the data landscape. Leads to longer turnaround times in producing actionable insights from their own data assets. Thus, diminishing the value realized by traditional methods. However, our operating model involves profiling and remediating at the same time creating a catalog data estate that can be used by business or IT accordingly. With increased automation and fewer people. Our machine learning algorithms augment the data pipeline to tag and capture the data elements into a comprehensive data catalog. As IO-Tahoe automatically catalogs the data estate in a centralized view, the data experts can partly focus on remediating the data events generated from validating against business rules. We envision that data events coupled with this drillable and searchable view will be a comprehensive one to assess the impact of bad quality data. Let's briefly look at the image on screen. For example, the view indicates that bad quality zip code data impacts the contact data which in turn impacts other related entities in systems. Now contrast that with a manually maintained spreadsheet that drowns out the main focus of your analysis. >> Tiji, how do you tag and capture bad quality data and stop that from you've mentioned these printed dependencies. How do you stop that from flowing downstream into the processes within the applications or reports? >> As IO-Tahoe builds the data catalog across source systems. We tag the elements that meet the business rule criteria while segregating the failed data examples associated with the elements that fall below a certain threshold. The elements that meet the business rule criteria are tagged to be searchable. Thus, providing an easy way to identify data elements that may flow through the system. The segregated data examples on the other hand are used by data experts to triage for the root cause. Based on the root cause potential outcomes could be one, changes in the source system to prevent that data from entering the system in the first place. Two, add data pipeline logic, to sanitize bad data from being consumed by downstream applications and reports or just accept the risk of storing bad data and address it when it meets a certain threshold. However, Dave as for your question about preventing bad quality data from flowing into the system? IO-Tahoe will not prevent it because the controls of data flowing between systems is managed outside of IO-Tahoe. Although, IO-Tahoe will alert and notify the data experts to events that indicate bad data has entered the monitored assets. Also we have redesigned our product to be modular and extensible. This allows data events generated by IO-Tahoe to be consumed by any system that wants to control the targets from bad data. Does IO-Tahoe empowers the data experts to control the bad data from flowing into their system. >> Thank you for that. So, one of the things that we've noticed, we've written about is that you've got these hyper specialized roles within the data, the centralized data organization. And wonder how do the data folks get involved here if at all, and how frequently do they get involved? Maybe Senthilnathan you could take that. >> Thank you, Dave for having me here. Well, based on whether the data element in question is in data cataloging or monitoring phase. Different data folks gets involved. When it doesn't the data cataloging stage. The data governance team, along with enterprise architecture or IT involved in setting up the data catalog. Which includes identifying the critical data elements business term identification, definition, documentation data quality rules, and data even set up data domain and business line mapping, lineage PA tracking source of truth. So on and so forth. It's typically in one time set up review certify then govern and monitor. But while when it is in the monitoring phase during any data incident or data issues IO-Tahoe broadcast data signals to the relevant data folks to act and remedy it as quick as possible. And alerts the consumption team it could be the data science, analytics, business opts are both a potential issue so that they are aware and take necessary preventative measure. Let me show you an example, critical data element from data quality dashboard view to lineage view to data 360 degree view for a zip code for conformity check. So in this case the zip code did not meet the past threshold during the technical data quality check and was identified as non-compliant item and notification was sent to the ID folks. So clicking on the zip code. Will take to the lineage view to visualize the dependent system, says that who are producers and who are the consumers. And further drilling down will take us to the detailed view, that a lot of other information's are presented to facilitate for a root cause analysis and not to take it to a final closure. >> Thank you for that. So Tiji? Patrick was talking about the as is to be. So I'm interested in how it's done now versus before. Do you need a data governance operating model for example? >> Typically a company that decides to make an inventory of the data assets would start out by manually building a spreadsheet managed by data experts of the company. What started as a draft now get break into the model of a company. This leads to loss of collaboration as each department makes a copy of their catalog for their specific needs. This decentralized approach leads to loss of uniformity which each department having different definitions which ironically needs a governance model for the data catalog itself. And as the spreadsheet grows in complexity the skill level needed to maintain. It also increases thus leading to fewer and fewer people knowing how to maintain it. About all the content that took so much time and effort to build is not searchable outside of that spreadsheet document. >> Yeah, I think you really hit the nail on my head Tiji. Now companies want to move away from the spreadsheet approach. IO-Tahoe addresses the shortcoming of the traditional approach enabling companies to achieve more with less. >> Yeah, what the customer reaction has been. We had Webster Bank, on one of the early episodes for example, I mean could they have achieved. What they did without something like active data quality and automation maybe Senthilnathan you could address that? >> Sure, It is impossible to achieve full data quality monitoring and remediation without automation or digital workers in place reality that introverts they don't have the time to do the remediation manually because they have to do an analysis conform fix on any data quality issues, as fast as possible before it gets bigger and no exception to Webster. That's why Webster implemented IO-Tahoe's active DQ to set up the business, metadata management and data quality monitoring and remediation in the Snowflake cloud data Lake. We help and building the center of excellence in the data governance, which is managing the data catalog schedule on demand and in-flight data quality checks, but Snowflake, no pipe on stream are super beneficial to achieve in flight quality checks. Then the data assumption monitoring and reporting last but not the least the time saver is persisting the non-compliant records for every data quality run within the Snowflake cloud, along with remediation script. So that during any exceptions the respect to team members is not only alerted. But also supplied with necessary scripts and tools to perform remediation right from the IO-Tahoe's Active DQ. >> Very nice. Okay guys, thanks for the demo. Great stuff. Now, if you want to learn more about the IO-Tahoe platform and how you can accelerate your adoption of Snowflake book some time with a data RPA expert all you got to do is click on the demo icon on the right of your screen and set a meeting. We appreciate you attending this latest episode of the IO-Tahoe data automation series. Look, if you missed any of the content that's all available on demand. This is Dave Vellante theCUBE. Thanks for watching. (upbeat music)
SUMMARY :
the globe it's theCUBE. and tell him, do the demo. and push the metadata to Snowflake. if you could address or the ability to monitor the operating model on remediating the data events generated into the processes within the data experts to events that indicate So, one of the things that So clicking on the zip code. Thank you for that. the skill level needed to maintain. of the traditional approach one of the early episodes So that during any exceptions the respect of the IO-Tahoe data automation series.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Patrick | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Tiji Matthew | PERSON | 0.99+ |
Tiji Mathew | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Patrick Zimet | PERSON | 0.99+ |
IO-Tahoe | ORGANIZATION | 0.99+ |
Senthil Karuppaiah | PERSON | 0.99+ |
360 degree | QUANTITY | 0.99+ |
Tiji | PERSON | 0.99+ |
Senthilnathan Karuppaiah | PERSON | 0.99+ |
each department | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
Snowflake | TITLE | 0.99+ |
Webster | ORGANIZATION | 0.99+ |
Aj | PERSON | 0.99+ |
Dunkin | PERSON | 0.98+ |
Two | QUANTITY | 0.98+ |
IO | ORGANIZATION | 0.97+ |
Patrick Zeimet | PERSON | 0.97+ |
one time | QUANTITY | 0.97+ |
Webster Bank | ORGANIZATION | 0.97+ |
one | QUANTITY | 0.97+ |
Io-Tahoe | ORGANIZATION | 0.96+ |
both | QUANTITY | 0.96+ |
Senthilnathan | PERSON | 0.96+ |
IO-Tahoe | TITLE | 0.95+ |
first place | QUANTITY | 0.89+ |
Snowflake | EVENT | 0.71+ |
Tahoe | ORGANIZATION | 0.69+ |
Data Solutions | ORGANIZATION | 0.69+ |
IO | TITLE | 0.68+ |
-Tahoe | TITLE | 0.64+ |
Snowflake | ORGANIZATION | 0.6+ |
Morrisons | ORGANIZATION | 0.6+ |
Tahoe | TITLE | 0.59+ |
Ajay Vohora and Duncan Turnbull | Io-Tahoe Data Quality: Active DQ
>> Announcer: From around the globe. It's the cube presenting active DQ, intelligent automation for data quality brought to you by Io Tahoe. (indistinct) >> Got it? all right if everybody is ready we'll opening on Dave in five, four, three. Now we're going to look at the role automation plays in mobilizing your data on snowflake. Let's welcome. And Duncan Turnbull who's partner sales engineer at snowflake, Ajay Vohora is back CEO of IO. Tahoe he's going to share his insight. Gentlemen. Welcome. >> Thank you, David good to be back. >> Yes it's great to have you back Ajay and it's really good to see Io Tahoe expanding the ecosystem so important now of course bringing snowflake in, it looks like you're really starting to build momentum. I mean, there's progress that we've seen every month month by month, over the past 12, 14 months. Your seed investors, they got to be happy. >> They are they're happy and they can see that we're running into a nice phase of expansion here new customers signing up, and now we're ready to go out and raise that next round of funding. Maybe think of us like Snowflake five years ago. So we're definitely on track with that. A lot of interest from investors and right now trying to focus in on those investors that can partner with us and understand AI data and an automation. >> Well, so personally, I mean you've managed a number of early stage VC funds. I think four of them. You've taken several comm software companies through many funding rounds and growth and all the way to exit. So you know how it works. You have to get product market fit, you got to make sure you get your KPIs, right. And you got to hire the right salespeople, but what's different this time around? >> Well, you know, the fundamentals that you mentioned those that never change. What I can see that's different that's shifted this time around is three things. One in that they used to be this kind of choice of do we go open source or do we go proprietary? Now that has turned into a nice hybrid model where we've really keyed into RedHat doing something similar with Centos. And the idea here is that there is a core capability of technology that underpins a platform, but it's the ability to then build an ecosystem around that made up of a community. And that community may include customers, technology partners, other tech vendors and enabling the platform adoption so that all of those folks in that community can build and contribute whilst still maintaining the core architecture and platform integrity at the core of it. And that's one thing that's changed. We're seeing a lot of that type of software company emerge into that model, which is different from five years ago. And then leveraging the Cloud, every Cloud, Snowflake Cloud being one of them here. In order to make use of what customers end customers in enterprise software are moving towards. Every CIO is now in some configuration of a hybrid. IT is state whether that is Cloud, multi-Cloud, on-prem. That's just the reality. The other piece is in dealing with the CIO, his legacy. So the past 15, 20 years I've purchased many different platforms, technologies, and some of those are still established and still (indistinct) How do you enable that CIO to make purchase whilst still preserving and in some cases building on and extending the legacy material technology. So they've invested their people's time and training and financial investment into. Yeah, of course solving a problem, customer pain point with technology that never goes out in a fashion >> That never changes. You have to focus like a laser on that. And of course, speaking of companies who are focused on solving problems, Duncan Turnbull from Snowflake. You guys have really done a great job and really brilliantly addressing pain points particularly around data warehousing, simplified that you're providing this new capability around data sharing really quite amazing. Duncan, Ajay talks about data quality and customer pain points in enterprise IT. Why is data quality been such a problem historically? >> So one of the biggest challenges that's really affected that in the past is that because to address everyone's needs for using data, they've evolved all these kinds of different places to store it, all these different silos or data marts or all this kind of pluralfiation of places where data lives and all of those end up with slightly different schedules for bringing data in and out, they end up with slightly different rules for transforming that data and formatting it and getting it ready and slightly different quality checks for making use of it. And this then becomes like a big problem in that these different teams are then going to have slightly different or even radically different ounces to the same kinds of questions, which makes it very hard for teams to work together on their different data problems that exist inside the business, depending on which of these silos they end up looking at. And what you can do. If you have a single kind of scalable system for putting all of your data, into it, you can kind of side step along this complexity and you can address the data quality issues in a single way. >> Now, of course, we're seeing this huge trend in the market towards robotic process automation, RPA that adoption is accelerating. You see in UI paths, IPO, 35 plus billion dollars, valuation, Snowflake like numbers, nice comms there for sure. Ajay you've coined the phrase data RPA what is that in simple terms? >> Yeah I mean, it was born out of seeing how in our ecosystem (indistinct) community developers and customers general business users for wanting to adopt and deploy Io Tahoe's technology. And we could see that. I mean, there's not marketing out here we're not trying to automate that piece but wherever there is a process that was tied into some form of a manual overhead with handovers. And so on, that process is something that we were able to automate with Io Tahoe's technology and the employment of AI and machine learning technologies specifically to those data processes, almost as a precursor to getting into marketing automation or financial information automation. That's really where we're seeing the momentum pick up especially in the last six months. And we've kept it really simple with snowflake. We've kind of stepped back and said, well, the resource that a Snowflake can leverage here is the metadata. So how could we turn Snowflake into that repository of being the data catalog? And by the way, if you're a CIO looking to purchase the data catalog tool, stop there's no need to. Working with Snowflake we've enabled that intelligence to be gathered automatically and to be put to use within snowflake. So reducing that manual effort and I'm putting that data to work. And that's where we've packaged this with our AI machine learning specific to those data tasks. And it made sense that's what's resonated with our customers. >> You know, what's interesting here just a quick aside, as you know I've been watching snowflake now for awhile and of course the competitors come out and maybe criticize, "Why they don't have this feature. They don't have that feature." And snowflake seems to have an answer. And the answer oftentimes is, well ecosystem, ecosystem is going to bring that because we have a platform that's so easy to work with. So I'm interested Duncan in what kind of collaborations you are enabling with high quality data. And of course, your data sharing capability. >> Yeah so I think the ability to work on datasets isn't just limited to inside the business itself or even between different business units you're kind of discussing maybe with those silos before. When looking at this idea of collaboration. We have these challenges where we want to be able to exploit data to the greatest degree possible, but we need to maintain the security, the safety, the privacy, and governance of that data. It could be quite valuable. It could be quite personal depending on the application involved. One of these novel applications that we see between organizations of data sharing is this idea of data clean rooms. And these data clean rooms are safe, collaborative spaces which allow multiple companies or even divisions inside a company where they have particular privacy requirements to bring two or more data sets together, for analysis. But without having to actually share the whole unprotected data set with each other. And this lets you to you know, when you do this inside of Snowflake you can collaborate using standard tool sets. You can use all of our SQL ecosystem. You can use all of the data science ecosystem that works with Snowflake. You can use all of the BI ecosystem that works with snowflake. But you can do that in a way that keeps the confidentiality that needs to be presented inside the data intact. And you can only really do these kinds of collaborations especially across organization but even inside large enterprises, when you have good reliable data to work with, otherwise your analysis just isn't going to really work properly. A good example of this is one of our large gaming customers. Who's an appetizer. They were able to build targeted ads to acquire customers and measure the campaign impact in revenue but they were able to keep their data safe and secure while doing that while working with advertising partners. The business impact of that was they're able to get a lift of 20 to 25% in campaign effectiveness through better targeting and actually pull through into that of a reduction in customer acquisition costs because they just didn't have to spend as much on the forms of media that weren't working for them. >> So, Ajay I wonder, I mean with the way public policy is shaping out, you know, obviously GDPR started it in the States, California consumer privacy Act, and people are sort of taking the best of those. And there's a lot of differentiation but what are you seeing just in terms of governments really driving this move to privacy. >> Government, public sector, we're seeing a huge wake up an activity and across (indistinct), part of it has been data privacy. The other part of it is being more joined up and more digital rather than paper or form based. We've all got, so there's a waiting in the line, holding a form, taking that form to the front of the line and handing it over a desk. Now government and public sector is really looking to transform their services into being online (indistinct) self service. And that whole shift is then driving the need to emulate a lot of what the commercial sector is doing to automate their processes and to unlock the data from silos to put through into those processes. And another thing that I can say about this is the need for data quality is as Duncan mentions underpins all of these processes government, pharmaceuticals, utilities, banking, insurance. The ability for a chief marketing officer to drive a a loyalty campaign, the ability for a CFO to reconcile accounts at the end of the month to do a quick accurate financial close. Also the ability of a customer operations to make sure that the customer has the right details about themselves in the right application that they can sell. So from all of that is underpinned by data and is effective or not based on the quality of that data. So whilst we're mobilizing data to the Snowflake Cloud the ability to then drive analytics, prediction, business processes of that Cloud succeeds or fails on the quality of that data. >> I mean it really is table stakes. If you don't trust the data you're not going to use the data. The problem is it always takes so long to get to the data quality. There's all these endless debates about it. So we've been doing a fair amount of work and thinking around this idea of decentralized data. Data by its very nature is decentralized but the fault domains of traditional big data is that everything is just monolithic. And the organizations monolithic that technology's monolithic, the roles are very, you know, hyper specialized. And so you're hearing a lot more these days about this notion of a data fabric or what Jimit Devani calls a data mesh and we've kind of been leaning into that and the ability to connect various data capabilities whether it's a data, warehouse or a data hub or a data lake, that those assets are discoverable, they're shareable through API APIs and they're governed on a federated basis. And you're using now bringing in a machine intelligence to improve data quality. You know, I wonder Duncan, if you could talk a little bit about Snowflake's approach to this topic >> Sure so I'd say that making use of all of your data is the key kind of driver behind these ideas of beta meshes or beta fabrics? And the idea is that you want to bring together not just your kind of strategic data but also your legacy data and everything that you have inside the enterprise. I think I'd also like to kind of expand upon what a lot of people view as all of the data. And I think that a lot of people kind of miss that there's this whole other world of data they could be having access to, which is things like data from their business partners, their customers, their suppliers, and even stuff that's, more in the public domain, whether that's, you know demographic data or geographic or all these kinds of other types of data sources. And what I'd say to some extent is that the data Cloud really facilitates the ability to share and gain access to this both kind of, between organizations, inside organizations. And you don't have to, make lots of copies of the data and kind of worry about the storage and this federated, idea of governance and all these things that it's quite complex to kind of manage. The snowflake approach really enables you to share data with your ecosystem or the world without any latency with full control over what's shared without having to introduce new complexities or having complex interactions with APIs or software integration. The simple approach that we provide allows a relentless focus on creating the right data product to meet the challenges facing your business today. >> So Ajay, the key here is Duncan's talking about it my mind and in my cake takeaway is to simplicity. If you can take the complexity out of the equation you're going to get more adoption. It really is that simple. >> Yeah, absolutely. I think that, that whole journey, maybe five, six years ago the adoption of data lakes was a stepping stone. However, the Achilles heel there was the complexity that it shifted towards consuming that data from a data lake where there were many, many sets of data to be able to cure rate and to consume. Whereas actually, the simplicity of being able to go to the data that you need to do your role, whether you're in tax compliance or in customer services is key. And listen for snowflake by Io Tahoe. One thing we know for sure is that our customers are super smart and they're very capable. They're data savvy and they'll want to use whichever tool and embrace whichever Cloud platform that is going to reduce the barriers to solving what's complex about that data, simplifying that and using good old fashioned SQL to access data and to build products from it to exploit that data. So simplicity is key to it to allow people to make use of that data and CIO is recognize that. >> So Duncan, the Cloud obviously brought in this notion of DevOps and new methodologies and things like agile that's brought in the notion of DataOps which is a very hot topic right now basically DevOps applies to data about how does Snowflake think about this? How do you facilitate that methodology? >> So I agree with you absolutely that DataOps takes these ideas of agile development or agile delivery and have the kind of DevOps world that we've seen just rise and rise. And it applies them to the data pipeline, which is somewhere where it kind of traditionally hasn't happened. And it's the same kinds of messages. As we see in the development world it's about delivering faster development having better repeatability and really getting towards that dream of the data-driven enterprise, where you can answer people's data questions they can make better business decisions. And we have some really great architectural advantages that allow us to do things like allow cloning of data sets without having to copy them, allows us to do things like time travel so we can see what the data looked like at some point in the past. And this lets you kind of set up both your own kind of little data playpen as a clone without really having to copy all of that data so it's quick and easy. And you can also, again with our separation of storage and compute, you can provision your own virtual warehouse for dev usage. So you're not interfering with anything to do with people's production usage of this data. So these ideas, the scalability, it just makes it easy to make changes, test them, see what the effect of those changes are. And we've actually seen this, that you were talking a lot about partner ecosystems earlier. The partner ecosystem has taken these ideas that are inside Snowflake and they've extended them. They've integrated them with DevOps and DataOps tooling. So things like version control and get an infrastructure automation and things like Terraform. And they've kind of built that out into more of a DataOps products that you can make use of. So we can see there's a huge impact of these ideas coming into the data world. We think we're really well-placed to take advantage to them. The partner ecosystem is doing a great job with doing that. And it really allows us to kind of change that operating model for data so that we don't have as much emphasis on like hierarchy and change windows and all these kinds of things that are maybe viewed as a lot as fashioned. And we kind of taken the shift from this batch stage of integration into streaming continuous data pipelines in the Cloud. And this kind of gets you away from like a once a week or once a month change window if you're really unlucky to pushing changes in a much more rapid fashion as the needs of the business change. >> I mean those hierarchical organizational structures when we apply those to begin to that it actually creates the silos. So if you're going to be a silo buster, which Ajay I look at you guys in silo busters, you've got to put data in the hands of the domain experts, the business people, they know what data they want, if they have to go through and beg and borrow for a new data sets cetera. And so that's where automation becomes so key. And frankly the technology should be an implementation detail not the dictating factor. I wonder if you could comment on this. >> Yeah, absolutely. I think making the technologies more accessible to the general business users or those specialists business teams that's the key to unlocking. So it is interesting to see is as people move from organization to organization where they've had those experiences operating in a hierarchical sense, I want to break free from that. And we've been exposed to automation. Continuous workflows change is continuous in IT. It's continuous in business. The market's continuously changing. So having that flow across the organization of work, using key components, such as GitHub and similar towards your drive process, Terraform to build in, code into the process and automation and with Io Tahoe, leveraging all the metadata from across those fragmented sources is good to see how those things are coming together. And watching people move from organization to organization say, "Hey okay, I've got a new start. I've got my first hundred days to impress my new manager. What kind of an impact can I bring to this?" And quite often we're seeing that as, let me take away the good learnings from how to do it or how not to do it from my previous role. And this is an opportunity for me to bring in automation. And I'll give you an example, David, recently started working with a client in financial services. Who's an asset manager, managing financial assets. They've grown over the course of the last 10 years through M&A and each of those acquisitions have bought with its technical debt, it's own set of data, that multiple CRM systems now multiple databases, multiple bespoke in-house created applications. And when the new CIO came in and had a look at those he thought well, yes I want to mobilize my data. Yes, I need to modernize my data state because my CEO is now looking at these crypto assets that are on the horizon and the new funds that are emerging that's around digital assets and crypto assets. But in order to get to that where absolutely data underpins that and is the core asset cleaning up that that legacy situation mobilizing the relevant data into the Snowflake Cloud platform is where we're giving time back. You know, that is now taking a few weeks whereas that transitioned to mobilize that data start with that new clean slate to build upon a new business as a digital crypto asset manager as well as the legacy, traditional financial assets, bonds, stocks, and fixed income assets, you name it is where we're starting to see a lot of innovation. >> Tons of innovation. I love the crypto examples, NFTs are exploding and let's face it. Traditional banks are getting disrupted. And so I also love this notion of data RPA. Especially because Ajay I've done a lot of work in the RPA space. And what I would observe is that the early days of RPA, I call it paving the cow path, taking existing processes and applying scripts, letting software robots do its thing. And that was good because it reduced mundane tasks, but really where it's evolved is a much broader automation agenda. People are discovering new ways to completely transform their processes. And I see a similar analogy for the data operating model. So I'm wonder what do you think about that and how a customer really gets started bringing this to their ecosystem, their data life cycles. >> Sure. Yeah. Step one is always the same. It's figuring out for the CIO, the chief data officer, what data do I have? And that's increasingly something that they want to automate, so we can help them there and do that automated data discovery whether that is documents in the file share backup archive in a relational data store in a mainframe really quickly hydrating that and bringing that intelligence the forefront of what do I have, and then it's the next step of, well, okay now I want to continually monitor and curate that intelligence with the platform that I've chosen let's say Snowflake. In order such that I can then build applications on top of that platform to serve my internal external customer needs. and the automation around classifying data, reconciliation across different fragmented data silos building that in those insights into Snowflake. As you say, a little later on where we're talking about data quality, active DQ, allowing us to reconcile data from different sources as well as look at the integrity of that data. So then go on to remediation. I want to harness and leverage techniques around traditional RPA but to get to that stage, I need to fix the data. So remediating publishing the data in Snowflake, allowing analysis to be formed, performed in Snowflake but those are the key steps that we see and just shrinking that timeline into weeks, giving the organization that time back means they're spending more time on their customer and solving their customer's problem which is where we want them to be. >> Well, I think this is the brilliance of Snowflake actually, you know, Duncan I've talked to Benoit Dageville about this and your other co-founders and it's really that focus on simplicity. So I mean, that's you picked a good company to join in my opinion. So I wonder Ajay, if you could talk about some of the industry sectors that again are going to gain the most from data RPA, I mean traditional RPA, if I can use that term, a lot of it was back office, a lot of financial, what are the practical applications where data RPA is going to impact businesses and the outcomes that we can expect. >> Yes, so our drive is really to make that business general user's experience of RPA simpler and using no code to do that where they've also chosen Snowflake to build their Cloud platform. They've got the combination then of using a relatively simple scripting techniques such as SQL without no code approach. And the answer to your question is whichever sector is looking to mobilize their data. It seems like a cop-out but to give you some specific examples, David now in banking, where our customers are looking to modernize their banking systems and enable better customer experience through applications and digital apps, that's where we're seeing a lot of traction in this approach to pay RPA to data. And health care where there's a huge amount of work to do to standardize data sets across providers, payers, patients and it's an ongoing process there. For retail helping to to build that immersive customer experience. So recommending next best actions. Providing an experience that is going to drive loyalty and retention, that's dependent on understanding what that customer's needs, intent are, being able to provide them with the content or the offer at that point in time or all data dependent utilities. There's another one great overlap there with Snowflake where helping utilities telecoms, energy, water providers to build services on that data. And this is where the ecosystem just continues to expand. If we're helping our customers turn their data into services for their ecosystem, that's exciting. Again, they were more so exciting than insurance which it always used to think back to, when insurance used to be very dull and mundane, actually that's where we're seeing a huge amounts of innovation to create new flexible products that are priced to the day to the situation and risk models being adaptive when the data changes on events or circumstances. So across all those sectors that they're all mobilizing their data, they're all moving in some way but for sure form to a multi-Cloud setup with their IT. And I think with Snowflake and with Io Tahoe being able to accelerate that and make that journey simple and less complex is why we've found such a good partner here. >> All right. Thanks for that. And thank you guys both. We got to leave it there really appreciate Duncan you coming on and Ajay best of luck with the fundraising. >> We'll keep you posted. Thanks, David. >> All right. Great. >> Okay. Now let's take a look at a short video. That's going to help you understand how to reduce the steps around your DataOps let's watch. (upbeat music)
SUMMARY :
brought to you by Io Tahoe. he's going to share his insight. and it's really good to see Io Tahoe and they can see that we're running and all the way to exit. but it's the ability to You have to focus like a laser on that. is that because to address in the market towards robotic and I'm putting that data to work. and of course the competitors come out that needs to be presented this move to privacy. the ability to then drive and the ability to connect facilitates the ability to share and in my cake takeaway is to simplicity. that is going to reduce the And it applies them to the data pipeline, And frankly the technology should be that's the key to unlocking. that the early days of RPA, and the automation and the outcomes that we can expect. And the answer to your question is We got to leave it there We'll keep you posted. All right. That's going to help you
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
David | PERSON | 0.99+ |
Ajay Vohora | PERSON | 0.99+ |
Duncan Turnbull | PERSON | 0.99+ |
Duncan Turnbull | PERSON | 0.99+ |
five | QUANTITY | 0.99+ |
Duncan | PERSON | 0.99+ |
two | QUANTITY | 0.99+ |
Dave | PERSON | 0.99+ |
IO | ORGANIZATION | 0.99+ |
Jimit Devani | PERSON | 0.99+ |
Ajay | PERSON | 0.99+ |
Io Tahoe | ORGANIZATION | 0.99+ |
20 | QUANTITY | 0.99+ |
Io-Tahoe | ORGANIZATION | 0.99+ |
One | QUANTITY | 0.99+ |
California consumer privacy Act | TITLE | 0.99+ |
Tahoe | PERSON | 0.99+ |
Benoit Dageville | PERSON | 0.99+ |
Snowflake | TITLE | 0.99+ |
five years ago | DATE | 0.99+ |
SQL | TITLE | 0.99+ |
first hundred days | QUANTITY | 0.98+ |
four | QUANTITY | 0.98+ |
GDPR | TITLE | 0.98+ |
each | QUANTITY | 0.98+ |
three | QUANTITY | 0.98+ |
both | QUANTITY | 0.98+ |
25% | QUANTITY | 0.97+ |
three things | QUANTITY | 0.97+ |
one | QUANTITY | 0.97+ |
M&A | ORGANIZATION | 0.97+ |
once a week | QUANTITY | 0.97+ |
one thing | QUANTITY | 0.96+ |
Snowflake | ORGANIZATION | 0.95+ |
once a month | QUANTITY | 0.95+ |
DevOps | TITLE | 0.95+ |
snowflake | TITLE | 0.94+ |
single | QUANTITY | 0.93+ |
last six months | DATE | 0.92+ |
States | TITLE | 0.92+ |
six years ago | DATE | 0.91+ |
single way | QUANTITY | 0.91+ |
Snowflake Cloud | TITLE | 0.9+ |
DataOps | TITLE | 0.9+ |
today | DATE | 0.86+ |
12 | QUANTITY | 0.85+ |
35 plus billion dollars | QUANTITY | 0.84+ |
five | DATE | 0.84+ |
Step one | QUANTITY | 0.83+ |
Tons | QUANTITY | 0.82+ |
RedHat | ORGANIZATION | 0.81+ |
Centos | ORGANIZATION | 0.8+ |
One thing | QUANTITY | 0.79+ |
14 months | QUANTITY | 0.79+ |
Io-Tahoe Episode 6: ActiveDQ™ Intelligent Automation for Data Quality Management promo 1
>>The data Lake concept was intriguing when first introduced in 2010, but people quickly realized that shoving data into a data Lake may data Lake stagnant, repositories that were essentially storage bins that were less expensive than traditional data warehouses. This is Dave Vellante joined me for IO. Tahoe's latest installment of the data automation series, active DQ, intelligent automation for data quality management. We'll talk to experts from snowflake about their data assessment utility from within the snowflake platform and how it scales to the demands of business. While also controlling costs. I have Tahoe CEO, AIG Hora will explain how IO Tahoe and snowflake together are bringing active DQ to market. And what the customers are saying about it. Save the date Thursday, April 29th for IO Tahoes data automation series active DQ, intelligent automation for data quality show streams promptly at 11:00 AM Eastern on the cube, the >>In high tech coverage.
SUMMARY :
the snowflake platform and how it scales to the demands of business.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
2010 | DATE | 0.99+ |
Thursday, April 29th | DATE | 0.99+ |
IO | ORGANIZATION | 0.98+ |
11:00 AM Eastern | DATE | 0.97+ |
first | QUANTITY | 0.96+ |
IO Tahoe | ORGANIZATION | 0.96+ |
AIG Hora | ORGANIZATION | 0.91+ |
Io-Tahoe | TITLE | 0.89+ |
IO Tahoes | ORGANIZATION | 0.89+ |
ActiveDQ™ | TITLE | 0.86+ |
Episode 6 | QUANTITY | 0.85+ |
Tahoe | ORGANIZATION | 0.83+ |
Tahoe | PERSON | 0.56+ |
CEO | PERSON | 0.54+ |
Io-Tahoe Episode 5: Enterprise Digital Resilience on Hybrid and Multicloud
>>from around the globe. It's the Cube presenting enterprise. Digital resilience on hybrid and multi cloud Brought to You by Iota Ho. Hello, everyone, and welcome to our continuing Siri's covering data automation brought to you by Io Tahoe. Today we're gonna look at how to ensure enterprise resilience for hybrid and multi cloud. Let's welcome in age. Eva Hora, who is the CEO of Iota A J. Always good to see you again. Thanks for coming on. >>Great to be back. David Pleasure. >>And he's joined by Fozzy Coons, who is a global principal architect for financial services. The vertical of financial services. That red hat. He's got deep experiences in that sector. Welcome, Fozzie. Good to see you. >>Thank you very much. Happy to be here. >>Fancy. Let's start with you. Look, there are a lot of views on cloud and what it is. I wonder if you could explain to us how you think about what is a hybrid cloud and and how it works. >>Sure, yes. So the hybrid cloud is a 90 architecture that incorporates some degree off workload, possibility, orchestration and management across multiple clouds. Those clouds could be private cloud or public cloud or even your own data centers. And how does it all work? It's all about secure interconnectivity and on demand. Allocation of resources across clouds and separate clouds can become hydrate when they're similarly >>interconnected. And >>it is that interconnectivity that allows the workloads workers to be moved and how management can be unified in off the street. You can work and how well you have. These interconnections has a direct impact on how well your hybrid cloud will work. >>Okay, so we'll fancy staying with you for a minute. So in the early days of Cloud that turned private Cloud was thrown a lot around a lot, but often just meant virtualization of an on PREM system and a network connection to the public cloud. Let's bring it forward. What, in your view, does a modern hybrid cloud architecture look like? >>Sure. So for modern public clouds, we see that, um, teams organizations need to focus on the portability off applications across clouds. That's very important, right? And when organizations build applications, they need to build and deploy these applications as small collections off independently, loosely coupled services, and then have those things run on the same operating system which means, in other words, running it on Lenox everywhere and building cloud native applications and being able to manage and orchestrate thes applications with platforms like KUBERNETES or read it open shit, for example. >>Okay, so that Z, that's definitely different from building a monolithic application that's fossilized and and doesn't move. So what are the challenges for customers, you know, to get to that modern cloud? Aziz, you've just described it. Is it skill sets? Is that the ability to leverage things like containers? What's your view there? >>So, I mean, from what we've seen around around the industry, especially around financial services, where I spent most of my time, we see that the first thing that we see is management right now because you have all these clouds and all these applications, you have a massive array off connections off interconnections. You also have massive array off integrations, possibility and resource allocations as well, and then orchestrating all those different moving pieces. Things like storage networks and things like those are really difficult to manage, right? That's one. What s O Management is the first challenge. The second one is workload, placement, placement. Where do you place this? How do you place this cloud? Native applications. Do you or do you keep on site on Prem? And what do you put in the cloud? That is the the the other challenge. The major one. The third one is security. Security now becomes the key challenge and concern for most customers. And we could talk about how hundreds? Yeah, >>we're definitely gonna dig into that. Let's bring a J into the conversation. A J. You know, you and I have talked about this in the past. One of the big problems that virtually every companies face is data fragmentation. Um, talk a little bit about how I owe Tahoe unifies data across both traditional systems legacy systems. And it connects to these modern I t environments. >>Yeah, sure, Dave. I mean, fancy just nailed it. There used to be about data of the volume of data on the different types of data. But as applications become or connected and interconnected at the location of that data really matters how we serve that data up to those those app. So working with red hat in our partnership with Red Hat being able Thio, inject our data Discovery machine learning into these multiple different locations. Would it be in AWS on IBM Cloud or A D. C p R. On Prem being able thio Automate that discovery? I'm pulling that. That single view of where is all my data then allows the CEO to manage cast that can do things like one. I keep the data where it is on premise or in my Oracle Cloud or in my IBM cloud on Connect. The application that needs to feed off that data on the way in which you do that is machine learning. That learns over time is it recognizes different types of data, applies policies to declassify that data. Andi and brings it all together with automation. >>Right? And that's one of the big themes and we've talked about this on earlier episodes. Is really simplification really abstracting a lot of that heavy lifting away so we can focus on things A. J A. Z. You just mentioned e nifaz e. One of the big challenges that, of course, we all talk about his governance across thes disparity data sets. I'm curious as your thoughts. How does Red Hat really think about helping customers adhere to corporate edicts and compliance regulations, which, of course, are are particularly acute within financial services. >>Oh, yeah, Yes. So for banks and the payment providers, like you've just mentioned their insurers and many other financial services firms, Um, you know, they have to adhere Thio standards such as a PC. I. D. S s in Europe. You've got the G g d p g d p r, which requires strange and tracking, reporting documentation. And you know, for them to to remain in compliance and the way we recommend our customers to address these challenges is by having an automation strategy. Right. And that type of strategy can help you to improve the security on compliance off the organization and reduce the risk after the business. Right. And we help organizations build security and compliance from the start without consulting services residencies. We also offer courses that help customers to understand how to address some of these challenges. And that's also we help organizations build security into their applications without open sources. Mueller, where, um, middle offerings and even using a platform like open shift because it allows you to run legacy applications and also continue rights applications in a unified platform right And also that provides you with, you know, with the automation and the truly that you need to continuously monitor, manage and automate the systems for security and compliance >>purposes. Hey, >>Jay, anything. Any color you could add to this conversation? >>Yeah, I'm pleased. Badly brought up Open shift. I mean, we're using open shift to be able. Thio, take that security application of controls to to the data level. It's all about context. So, understanding what data is there being able to assess it to say who should have access to it. Which application permission should be applied to it. Um, that za great combination of Red Hat tonight. Tahoe. >>But what about multi Cloud? Doesn't that complicate the situation even even further? Maybe you could talk about some of the best practices to apply automation across not only hybrid cloud, but multi >>cloud a swell. Yeah, sure. >>Yeah. So the right automation solution, you know, can be the difference between, you know, cultivating an automated enterprise or automation caress. And some of the recommendations we give our clients is to look for an automation platform that can offer the first thing is complete support. So that means have an automation solution that provides that provides, um, you know, promotes I t availability and reliability with your platform so that you can provide, you know, enterprise great support, including security and testing, integration and clear roadmaps. The second thing is vendor interoperability interoperability in that you are going to be integrating multiple clouds. So you're going to need a solution that can connect to multiple clouds. Simples lee, right? And with that comes the challenge off maintain ability. So you you you're going to need to look into a automation Ah, solution that that is easy to learn or has an easy learning curve. And then the fourth idea that we tell our customers is scalability in the in the hybrid cloud space scale is >>is >>a big, big deal here, and you need a to deploy an automation solution that can span across the whole enterprise in a constituent, consistent manner, right? And then also, that allows you finally to, uh, integrate the multiple data centers that you have, >>So A J I mean, this is a complicated situation, for if a customer has toe, make sure things work on AWS or azure or Google. Uh, they're gonna spend all their time doing that, huh? What can you add really? To simplify that that multi cloud and hybrid cloud equation? >>Yeah. I could give a few customer examples here Warming a manufacturer that we've worked with to drive that simplification Onda riel bonuses for them is has been a reduction cost. We worked with them late last year to bring the cost bend down by $10 million in 2021 so they could hit that reduced budget. Andre, What we brought to that was the ability thio deploy using open shift templates into their different environments. Where there is on premise on bond or in as you mentioned, a W s. They had G cps well, for their marketing team on a cross, those different platforms being out Thio use a template, use pre built scripts to get up and running in catalog and discover that data within minutes. It takes away the legacy of having teams of people having Thio to jump on workshop cause and I know we're all on a lot of teens. The zoom cause, um, in these current times, they just sent me is in in of hours in the day Thio manually perform all of this. So yeah, working with red hat applying machine learning into those templates those little recipes that we can put that automation toe work, regardless of which location the data is in allows us thio pull that unified view together. Right? >>Thank you, Fozzie. I wanna come back to you. So the early days of cloud, you're in the big apple, you know, financial services. Really well. Cloud was like an evil word within financial services, and obviously that's changed. It's evolved. We talked about the pandemic, has even accelerated that, Um And when you really, you know, dug into it when you talk to customers about their experiences with security in the cloud it was it was not that it wasn't good. It was great, whatever. But it was different. And there's always this issue of skill, lack of skills and multiple tools suck up teams, they're really overburdened. But in the cloud requires new thinking. You've got the shared responsibility model you've got obviously have specific corporate requirements and compliance. So this is even more complicated when you introduce multiple clouds. So what are the differences that you can share from your experience is running on a sort of either on Prem or on a mono cloud, um, or, you know, and versus across clouds. What? What? What do you suggest there? >>Yeah, you know, because of these complexities that you have explained here, Miss Configurations and the inadequate change control the top security threats. So human error is what we want to avoid because is, you know, as your clouds grow with complexity and you put humans in the mix, then the rate off eras is going to increase, and that is going to exposure to security threat. So this is where automation comes in because automation will streamline and increase the consistency off your infrastructure management. Also application development and even security operations to improve in your protection, compliance and change control. So you want to consistently configure resources according to a pre approved um, you know, pre approved policies and you want to proactively maintain a to them in a repeatable fashion over the whole life cycle. And then you also want to rapid the identified system that require patches and and reconfiguration and automate that process off patching and reconfiguring so that you don't have humans doing this type of thing, right? And you want to be able to easily apply patches and change assistant settings. According Thio, Pre defined, based on like explained before, you know, with the pre approved policies and also you want is off auditing and troubleshooting, right? And from a rate of perspective, we provide tools that enable you to do this. We have, for example, a tool called danceable that enables you to automate data center operations and security and also deployment of applications and also obvious shit yourself, you know, automates most of these things and obstruct the human beings from putting their fingers on, causing, uh, potentially introducing errors right now in looking into the new world off multiple clouds and so forth. The difference is that we're seeing here between running a single cloud or on prem is three main areas which is control security and compliance. Right control here it means if your on premise or you have one cloud, um, you know, in most cases you have control over your data and your applications, especially if you're on Prem. However, if you're in the public cloud, there is a difference there. The ownership, it is still yours. But your resources are running on somebody else's or the public clouds. You know, e w s and so forth infrastructure. So people that are going to do this need to really especially banks and governments need to be aware off the regulatory constraints off running, uh, those applications in the public cloud. And we also help customers regionalize some of these choices and also on security. You will see that if you're running on premises or in a single cloud, you have more control, especially if you're on Prem. You can control this sensitive information that you have, however, in the cloud. That's a different situation, especially from personal information of employees and things like that. You need to be really careful off that. And also again, we help you rationalize some of those choices. And then the last one is compliant. Aziz. Well, you see that if you're running on Prem or a single cloud, um, regulations come into play again, right? And if you're running a problem, you have control over that. You can document everything you have access to everything that you need. But if you're gonna go to the public cloud again, you need to think about that. We have automation, and we have standards that can help you, uh, you know, address some of these challenges for security and compliance. >>So that's really strong insights, Potsie. I mean, first of all, answerable has a lot of market momentum. Red hats in a really good job with that acquisition, your point about repeatability is critical because you can't scale otherwise. And then that idea you're you're putting forth about control, security compliance It's so true is I called it the shared responsibility model. And there was a lot of misunderstanding in the early days of cloud. I mean, yeah, maybe a W s is gonna physically secure the, you know, s three, but in the bucket. But we saw so many Miss configurations early on. And so it's key to have partners that really understand this stuff and can share the experiences of other clients. So this all sounds great. A j. You're sharp, you know, financial background. What about the economics? >>You >>know, our survey data shows that security it's at the top of the spending priority list, but budgets are stretched thin. E especially when you think about the work from home pivot and and all the areas that they had toe the holes that they had to fill their, whether it was laptops, you know, new security models, etcetera. So how do organizations pay for this? What's the business case look like in terms of maybe reducing infrastructure costs so I could, you know, pay it forward or there's a There's a risk reduction angle. What can you share >>their? Yeah. I mean, the perspective I'd like to give here is, um, not being multi cloud is multi copies of an application or data. When I think about 20 years, a lot of the work in financial services I was looking at with managing copies of data that we're feeding different pipelines, different applications. Now what we're saying I talk a lot of the work that we're doing is reducing the number of copies of that data so that if I've got a product lifecycle management set of data, if I'm a manufacturer, I'm just gonna keep that in one location. But across my different clouds, I'm gonna have best of breed applications developed in house third parties in collaboration with my supply chain connecting securely to that. That single version of the truth. What I'm not going to do is to copy that data. So ah, lot of what we're seeing now is that interconnectivity using applications built on kubernetes. Um, that decoupled from the data source that allows us to reduce those copies of data within that you're gaining from the security capability and resilience because you're not leaving yourself open to those multiple copies of data on with that. Couldn't come. Cost, cost of storage on duh cost of compute. So what we're seeing is using multi cloud to leverage the best of what each cloud platform has to offer That goes all the way to Snowflake and Hiroko on Cloud manage databases, too. >>Well, and the people cost to a swell when you think about yes, the copy creep. But then you know when something goes wrong, a human has to come in and figured out um, you brought up snowflake, get this vision of the data cloud, which is, you know, data data. I think this we're gonna be rethinking a j, uh, data architectures in the coming decade where data stays where it belongs. It's distributed, and you're providing access. Like you said, you're separating the data from the applications applications as we talked about with Fozzie. Much more portable. So it Z really the last 10 years will be different than the next 10 years. A. >>J Definitely. I think the people cast election is used. Gone are the days where you needed thio have a dozen people governing managing black policies to data. Ah, lot of that repetitive work. Those tests can be in power automated. We've seen examples in insurance were reduced teams of 15 people working in the the back office China apply security controls compliance down to just a couple of people who are looking at the exceptions that don't fit. And that's really important because maybe two years ago the emphasis was on regulatory compliance of data with policies such as GDP are in CCP a last year, very much the economic effect of reduce headcounts on on enterprises of running lean looking to reduce that cost. This year, we can see that already some of the more proactive cos they're looking at initiatives such as net zero emissions how they use data toe under understand how cape how they can become more have a better social impact. Um, and using data to drive that, and that's across all of their operations and supply chain. So those regulatory compliance issues that may have been external we see similar patterns emerging for internal initiatives that benefiting the environment, social impact and and, of course, course, >>great perspectives. Yeah, Jeff Hammer, Bucker once famously said, The best minds of my generation are trying to get people to click on ads and a J. Those examples that you just gave of, you know, social good and moving. Uh, things forward are really critical. And I think that's where Data is gonna have the biggest societal impact. Okay, guys, great conversation. Thanks so much for coming on the program. Really appreciate your time. Keep it right there from, or insight and conversation around, creating a resilient digital business model. You're watching the >>Cube digital resilience, automated compliance, privacy and security for your multi cloud. Congratulations. You're on the journey. You have successfully transformed your organization by moving to a cloud based platform to ensure business continuity in these challenging times. But as you scale your digital activities, there is an inevitable influx of users that outpaces traditional methods of cybersecurity, exposing your data toe underlying threats on making your company susceptible toe ever greater risk to become digitally resilient. Have you applied controls your data continuously throughout the data Lifecycle? What are you doing to keep your customer on supply data private and secure? I owe Tahoe's automated, sensitive data. Discovery is pre programmed with over 300 existing policies that meet government mandated risk and compliance standards. Thes automate the process of applying policies and controls to your data. Our algorithm driven recommendation engine alerts you to risk exposure at the data level and suggests the appropriate next steps to remain compliant on ensure sensitive data is secure. Unsure about where your organization stands In terms of digital resilience, Sign up for a minimal cost commitment. Free data Health check. Let us run our sensitive data discovery on key unmapped data silos and sources to give you a clear understanding of what's in your environment. Book time within Iot. Tahoe Engineer Now >>Okay, let's now get into the next segment where we'll explore data automation. But from the angle of digital resilience within and as a service consumption model, we're now joined by Yusuf Khan, who heads data services for Iot, Tahoe and Shirish County up in. Who's the vice president and head of U. S. Sales at happiest Minds? Gents, welcome to the program. Great to have you in the Cube. >>Thank you, David. >>Trust you guys talk about happiest minds. This notion of born digital, foreign agile. I like that. But talk about your mission at the company. >>Sure. >>A former in 2011 Happiest Mind is a born digital born a child company. The reason is that we are focused on customers. Our customer centric approach on delivering digitals and seamless solutions have helped us be in the race. Along with the Tier one providers, Our mission, happiest people, happiest customers is focused to enable customer happiness through people happiness. We have Bean ranked among the top 25 i t services company in the great places to work serving hour glass to ratings off 41 against the rating off. Five is among the job in the Indian nineties services company that >>shows the >>mission on the culture. What we have built on the values right sharing, mindful, integrity, learning and social on social responsibilities are the core values off our company on. That's where the entire culture of the company has been built. >>That's great. That sounds like a happy place to be. Now you said you had up data services for Iot Tahoe. We've talked in the past. Of course you're out of London. What >>do you what? Your >>day to day focus with customers and partners. What you focused >>on? Well, David, my team work daily with customers and partners to help them better understand their data, improve their data quality, their data governance on help them make that data more accessible in a self service kind of way. To the stakeholders within those businesses on dis is all a key part of digital resilience that will will come on to talk about but later. You're >>right, e mean, that self service theme is something that we're gonna we're gonna really accelerate this decade, Yussef and so. But I wonder before we get into that, maybe you could talk about the nature of the partnership with happiest minds, you know? Why do you guys choose toe work closely together? >>Very good question. Um, we see Hyo Tahoe on happiest minds as a great mutual fit. A Suresh has said, uh, happiest minds are very agile organization um, I think that's one of the key things that attracts their customers on Io. Tahoe is all about automation. Uh, we're using machine learning algorithms to make data discovery data cataloging, understanding, data done. See, uh, much easier on. We're enabling customers and partners to do it much more quickly. So when you combine our emphasis on automation with the emphasis on agility that happiest minds have that that's a really nice combination work works very well together, very powerful. I think the other things that a key are both businesses, a serious have said, are really innovative digital native type type companies. Um, very focused on newer technologies, the cloud etcetera on. Then finally, I think they're both Challenger brands on happiest minds have a really positive, fresh ethical approach to people and customers that really resonates with us at Ideo Tahoe to >>great thank you for that. So Russia, let's get into the whole notion of digital resilience. I wanna I wanna sort of set it up with what I see, and maybe you can comment be prior to the pandemic. A lot of customers that kind of equated disaster recovery with their business continuance or business resilient strategy, and that's changed almost overnight. How have you seen your clients respond to that? What? I sometimes called the forced march to become a digital business. And maybe you could talk about some of the challenges that they faced along the way. >>Absolutely. So, uh, especially during this pandemic, times when you say Dave, customers have been having tough times managing their business. So happiest minds. Being a digital Brazilian company, we were able to react much faster in the industry, apart from the other services company. So one of the key things is the organisation's trying to adopt onto the digital technologies. Right there has bean lot off data which has been to manage by these customers on There have been lot off threats and risk, which has been to manage by the CEO Seo's so happiest minds digital resilient technology, right where we bring in the data. Complaints as a service were ableto manage the resilience much ahead off other competitors in the market. We were ableto bring in our business continuity processes from day one, where we were ableto deliver our services without any interruption to the services. What we were delivered to our customers So that is where the digital resilience with business community process enabled was very helpful for us. Toe enable our customers continue their business without any interruptions during pandemics. >>So I mean, some of the challenges that customers tell me they obviously they had to figure out how to get laptops to remote workers and that that whole remote work from home pivot figure out how to secure the end points. And, you know, those were kind of looking back there kind of table stakes, But it sounds like you've got a digital business. Means a data business putting data at the core, I like to say, but so I wonder if you could talk a little bit more about maybe the philosophy you have toward digital resilience in the specific approach you take with clients? >>Absolutely. They seen any organization data becomes. The key on that, for the first step is to identify the critical data. Right. So we this is a six step process. What we following happiest minds. First of all, we take stock off the current state, though the customers think that they have a clear visibility off their data. How are we do more often assessment from an external point off view on see how critical their data is, then we help the customers to strategies that right. The most important thing is to identify the most important critical herself. Data being the most critical assert for any organization. Identification off the data's key for the customers. Then we help in building a viable operating model to ensure these identified critical assets are secure on monitor dearly so that they are consumed well as well as protected from external threats. Then, as 1/4 step, we try to bring in awareness, toe the people we train them >>at >>all levels in the organization. That is a P for people to understand the importance off the digital ourselves and then as 1/5 step, we work as a back up plan in terms of bringing in a very comprehensive and a holistic testing approach on people process as well as in technology. We'll see how the organization can withstand during a crisis time, and finally we do a continuous governance off this data, which is a key right. It is not just a one step process. We set up the environment, we do the initial analysis and set up the strategy on continuously govern this data to ensure that they are not only know managed will secure as well as they also have to meet the compliance requirements off the organization's right. That is where we help organizations toe secure on Meet the regulations off the organizations. As for the privacy laws, so this is a constant process. It's not on one time effort. We do a constant process because every organization goes towards their digital journey on. They have to face all these as part off the evolving environment on digital journey. And that's where they should be kept ready in terms off. No recovering, rebounding on moving forward if things goes wrong. >>So let's stick on that for a minute, and then I wanna bring yourself into the conversation. So you mentioned compliance and governance when when your digital business, you're, as you say, you're a data business, so that brings up issues. Data sovereignty. Uh, there's governance, this compliance. There's things like right to be forgotten. There's data privacy, so many things. These were often kind of afterthoughts for businesses that bolted on, if you will. I know a lot of executives are very much concerned that these air built in on, and it's not a one shot deal. So do you have solutions around compliance and governance? Can you deliver that as a service? Maybe you could talk about some of the specifics there, >>so some of way have offered multiple services. Tow our customers on digital against. On one of the key service is the data complaints. As a service here we help organizations toe map the key data against the data compliance requirements. Some of the features includes in terms off the continuous discovery off data right, because organizations keep adding on data when they move more digital on helping the helping and understanding the actual data in terms off the residents of data, it could be a heterogeneous data soldiers. It could be on data basis, or it could be even on the data legs. Or it could be a no even on compromise all the cloud environment. So identifying the data across the various no heterogeneous environment is very key. Feature off our solution. Once we identify classify this sensitive data, the data privacy regulations on the traveling laws have to be map based on the business rules So we define those rules on help map those data so that organizations know how critical their digital assets are. Then we work on a continuous marching off data for anomalies because that's one of the key teachers off the solution, which needs to be implemented on the day to day operational basis. So we're helping monitoring those anomalies off data for data quality management on an ongoing basis. On finally, we also bringing the automated data governance where we can manage the sensory data policies on their later relationships in terms off mapping on manage their business roots on we drive reputations toe Also suggest appropriate actions to the customers. Take on those specific data sets. >>Great. Thank you, Yousef. Thanks for being patient. I want to bring in Iota ho thio discussion and understand where your customers and happiest minds can leverage your data automation capability that you and I have talked about in the past. I'm gonna be great if you had an example is well, but maybe you could pick it up from there, >>John. I mean, at a high level, assertions are clearly articulated. Really? Um, Hyoty, who delivers business agility. So that's by, um accelerating the time to operationalize data, automating, putting in place controls and actually putting helping put in place digital resilience. I mean way if we step back a little bit in time, um, traditional resilience in relation to data often met manually, making multiple copies of the same data. So you have a d b A. They would copy the data to various different places, and then business users would access it in those functional style owes. And of course, what happened was you ended up with lots of different copies off the same data around the enterprise. Very inefficient. ONDA course ultimately, uh, increases your risk profile. Your risk of a data breach. Um, it's very hard to know where everything is. And I realized that expression. They used David the idea of the forced march to digital. So with enterprises that are going on this forced march, what they're finding is they don't have a single version of the truth, and almost nobody has an accurate view of where their critical data is. Then you have containers bond with containers that enables a big leap forward so you could break applications down into micro services. Updates are available via a p I s on. So you don't have the same need thio to build and to manage multiple copies of the data. So you have an opportunity to just have a single version of the truth. Then your challenge is, how do you deal with these large legacy data states that the service has been referring Thio, where you you have toe consolidate and that's really where I attack comes in. Um, we massively accelerate that process of putting in a single version of the truth into place. So by automatically discovering the data, discovering what's dubica? What's redundant? Uh, that means you can consolidate it down to a single trusted version much more quickly. We've seen many customers have tried to do this manually, and it's literally taken years using manual methods to cover even a small percentage of their I T estates. With our tire, you could do it really very quickly on you can have tangible results within weeks and months on Ben, you can apply controls to the data based on context. So who's the user? What's the content? What's the use case? Things like data quality validations or access permissions on. Then, once you've done there. Your applications and your enterprise are much more secure, much more resilient. As a result, you've got to do these things whilst retaining agility, though. So coming full circle. This is where the partnership with happiest minds really comes in as well. You've got to be agile. You've gotta have controls. Um, on you've got a drug toward the business outcomes. Uh, and it's doing those three things together that really deliver for the customer. >>Thank you. Use f. I mean you and I. In previous episodes, we've looked in detail at the business case. You were just talking about the manual labor involved. We know that you can't scale, but also there's that compression of time. Thio get to the next step in terms of ultimately getting to the outcome. And we talked to a number of customers in the Cube, and the conclusion is, it's really consistent that if you could accelerate the time to value, that's the key driver reducing complexity, automating and getting to insights faster. That's where you see telephone numbers in terms of business impact. So my question is, where should customers start? I mean, how can they take advantage of some of these opportunities that we've discussed today. >>Well, we've tried to make that easy for customers. So with our Tahoe and happiest minds, you can very quickly do what we call a data health check. Um, this is a is a 2 to 3 week process, uh, to really quickly start to understand on deliver value from your data. Um, so, iota, who deploys into the customer environment? Data doesn't go anywhere. Um, we would look at a few data sources on a sample of data. Onda. We can very rapidly demonstrate how they discovery those catalog e on understanding Jupiter data and redundant data can be done. Um, using machine learning, um, on how those problems can be solved. Um, And so what we tend to find is that we can very quickly, as I say in the matter of a few weeks, show a customer how they could get toe, um, or Brazilian outcome on then how they can scale that up, take it into production on, then really understand their data state? Better on build. Um, Brasiliense into the enterprise. >>Excellent. There you have it. We'll leave it right there. Guys, great conversation. Thanks so much for coming on the program. Best of luck to you and the partnership Be well, >>Thank you, David Suresh. Thank you. Thank >>you for watching everybody, This is Dave Volonte for the Cuban are ongoing Siris on data automation without >>Tahoe, digital resilience, automated compliance, privacy and security for your multi cloud. Congratulations. You're on the journey. You have successfully transformed your organization by moving to a cloud based platform to ensure business continuity in these challenging times. But as you scale your digital activities, there is an inevitable influx of users that outpaces traditional methods of cybersecurity, exposing your data toe underlying threats on making your company susceptible toe ever greater risk to become digitally resilient. Have you applied controls your data continuously throughout the data lifecycle? What are you doing to keep your customer on supply data private and secure? I owe Tahoe's automated sensitive data. Discovery is pre programmed with over 300 existing policies that meet government mandated risk and compliance standards. Thes automate the process of applying policies and controls to your data. Our algorithm driven recommendation engine alerts you to risk exposure at the data level and suggests the appropriate next steps to remain compliant on ensure sensitive data is secure. Unsure about where your organization stands in terms of digital resilience. Sign up for our minimal cost commitment. Free data health check. Let us run our sensitive data discovery on key unmapped data silos and sources to give you a clear understanding of what's in your environment. Book time within Iot. Tahoe Engineer. Now. >>Okay, now we're >>gonna go into the demo. We want to get a better understanding of how you can leverage open shift. And I owe Tahoe to facilitate faster application deployment. Let me pass the mic to Sabetta. Take it away. >>Uh, thanks, Dave. Happy to be here again, Guys, uh, they've mentioned names to be the Davis. I'm the enterprise account executive here. Toyota ho eso Today we just wanted to give you guys a general overview of how we're using open shift. Yeah. Hey, I'm Noah Iota host data operations engineer, working with open ship. And I've been learning the Internets of open shift for, like, the past few months, and I'm here to share. What a plan. Okay, so So before we begin, I'm sure everybody wants to know. Noel, what are the benefits of using open shift. Well, there's five that I can think of a faster time, the operation simplicity, automation control and digital resilience. Okay, so that that's really interesting, because there's an exact same benefits that we had a Tahoe delivered to our customers. But let's start with faster time the operation by running iota. Who on open shift? Is it faster than, let's say, using kubernetes and other platforms >>are >>objective iota. Who is to be accessible across multiple cloud platforms, right? And so by hosting our application and containers were able to achieve this. So to answer your question, it's faster to create and use your application images using container tools like kubernetes with open shift as compared to, like kubernetes with docker cry over container D. Okay, so we got a bit technical there. Can you explain that in a bit more detail? Yeah, there's a bit of vocabulary involved, uh, so basically, containers are used in developing things like databases, Web servers or applications such as I have top. What's great about containers is that they split the workload so developers can select the libraries without breaking anything. And since Hammond's can update the host without interrupting the programmers. Uh, now, open shift works hand in hand with kubernetes to provide a way to build those containers for applications. Okay, got It s basically containers make life easier for developers and system happens. How does open shift differ from other platforms? Well, this kind of leads into the second benefit I want to talk about, which is simplicity. Basically, there's a lot of steps involved with when using kubernetes with docker. But open shift simplifies this with their source to image process that takes the source code and turns it into a container image. But that's not all. Open shift has a lot of automation and features that simplify working with containers, an important one being its Web console. Here. I've set up a light version of open ship called Code Ready Containers, and I was able to set up her application right from the Web console. And I was able to set up this entire thing in Windows, Mac and Lennox. So its environment agnostic in that sense. Okay, so I think I've seen the top left that this is a developers view. What would a systems admin view look like? It's a good question. So here's the administrator view and this kind of ties into the benefit of control. Um, this view gives insights into each one of the applications and containers that are running, and you could make changes without affecting deployment. Andi can also, within this view, set up each layer of security, and there's multiple that you can prop up. But I haven't fully messed around with it because with my luck, I'd probably locked myself out. So that seems pretty secure. Is there a single point security such as you use a log in? Or are there multiple layers of security? Yeah, there are multiple layers of security. There's your user login security groups and general role based access controls. Um, but there's also a ton of layers of security surrounding like the containers themselves. But for the sake of time, I won't get too far into it. Okay, eso you mentioned simplicity In time. The operation is being two of the benefits. You also briefly mention automation. And as you know, automation is the backbone of our platform here, Toyota Ho. So that's certainly grabbed my attention. Can you go a bit more in depth in terms of automation? Open shift provides extensive automation that speeds up that time the operation. Right. So the latest versions of open should come with a built in cryo container engine, which basically means that you get to skip that container engine insulation step and you don't have to, like, log into each individual container host and configure networking, configure registry servers, storage, etcetera. So I'd say, uh, it automates the more boring kind of tedious process is Okay, so I see the iota ho template there. What does it allow me to do? Um, in terms of automation in application development. So we've created an open shift template which contains our application. This allows developers thio instantly, like set up our product within that template. So, Noah Last question. Speaking of vocabulary, you mentioned earlier digital resilience of the term we're hearing, especially in the banking and finance world. Um, it seems from what you described, industries like banking and finance would be more resilient using open shift, Correct. Yeah, In terms of digital resilience, open shift will give you better control over the consumption of resource is each container is using. In addition, the benefit of containers is that, like I mentioned earlier since Hammond's can troubleshoot servers about bringing down the application and if the application does go down is easy to bring it back up using templates and, like the other automation features that open ship provides. Okay, so thanks so much. Know us? So any final thoughts you want to share? Yeah. I just want to give a quick recap with, like, the five benefits that you gained by using open shift. Uh, the five are timeto operation automation, control, security and simplicity. You could deploy applications faster. You could simplify the workload you could automate. A lot of the otherwise tedious processes can maintain full control over your workflow. And you could assert digital resilience within your environment. Guys, >>Thanks for that. Appreciate the demo. Um, I wonder you guys have been talking about the combination of a Iot Tahoe and red hat. Can you tie that in subito Digital resilience >>Specifically? Yeah, sure, Dave eso when we speak to the benefits of security controls in terms of digital resilience at Io Tahoe, we automated detection and apply controls at the data level, so this would provide for more enhanced security. >>Okay, But so if you were trying to do all these things manually. I mean, what what does that do? How much time can I compress? What's the time to value? >>So with our latest versions, Biota we're taking advantage of faster deployment time associated with container ization and kubernetes. So this kind of speeds up the time it takes for customers. Start using our software as they be ableto quickly spin up io towel on their own on premise environment are otherwise in their own cloud environment, like including aws. Assure or call GP on IBM Cloud a quick start templates allow flexibility deploy into multi cloud environments all just using, like, a few clicks. Okay, so so now just quickly add So what we've done iota, Who here is We've really moved our customers away from the whole idea of needing a team of engineers to apply controls to data as compared to other manually driven work flows. Eso with templates, automation, previous policies and data controls. One person can be fully operational within a few hours and achieve results straight out of the box on any cloud. >>Yeah, we've been talking about this theme of abstracting the complexity. That's really what we're seeing is a major trend in in this coming decade. Okay, great. Thanks, Sabina. Noah, How could people get more information or if they have any follow up questions? Where should they go? >>Yeah, sure. They've. I mean, if you guys are interested in learning more, you know, reach out to us at info at iata ho dot com to speak with one of our sales engineers. I mean, we love to hear from you, so book a meeting as soon as you can. All >>right. Thanks, guys. Keep it right there from or cube content with.
SUMMARY :
Always good to see you again. Great to be back. Good to see you. Thank you very much. I wonder if you could explain to us how you think about what is a hybrid cloud and So the hybrid cloud is a 90 architecture that incorporates some degree off And it is that interconnectivity that allows the workloads workers to be moved So in the early days of Cloud that turned private Cloud was thrown a lot to manage and orchestrate thes applications with platforms like Is that the ability to leverage things like containers? And what do you put in the cloud? One of the big problems that virtually every companies face is data fragmentation. the way in which you do that is machine learning. And that's one of the big themes and we've talked about this on earlier episodes. And that type of strategy can help you to improve the security on Hey, Any color you could add to this conversation? is there being able to assess it to say who should have access to it. Yeah, sure. the difference between, you know, cultivating an automated enterprise or automation caress. What can you add really? bond or in as you mentioned, a W s. They had G cps well, So what are the differences that you can share from your experience is running on a sort of either And from a rate of perspective, we provide tools that enable you to do this. A j. You're sharp, you know, financial background. know, our survey data shows that security it's at the top of the spending priority list, Um, that decoupled from the data source that Well, and the people cost to a swell when you think about yes, the copy creep. Gone are the days where you needed thio have a dozen people governing managing to get people to click on ads and a J. Those examples that you just gave of, you know, to give you a clear understanding of what's in your environment. Great to have you in the Cube. Trust you guys talk about happiest minds. We have Bean ranked among the mission on the culture. Now you said you had up data services for Iot Tahoe. What you focused To the stakeholders within those businesses on dis is of the partnership with happiest minds, you know? So when you combine our emphasis on automation with the emphasis And maybe you could talk about some of the challenges that they faced along the way. So one of the key things putting data at the core, I like to say, but so I wonder if you could talk a little bit more about maybe for the first step is to identify the critical data. off the digital ourselves and then as 1/5 step, we work as a back up plan So you mentioned compliance and governance when when your digital business, you're, as you say, So identifying the data across the various no heterogeneous environment is well, but maybe you could pick it up from there, So you don't have the same need thio to build and to manage multiple copies of the data. and the conclusion is, it's really consistent that if you could accelerate the time to value, to really quickly start to understand on deliver value from your data. Best of luck to you and the partnership Be well, Thank you, David Suresh. to give you a clear understanding of what's in your environment. Let me pass the mic to And I've been learning the Internets of open shift for, like, the past few months, and I'm here to share. into each one of the applications and containers that are running, and you could make changes without affecting Um, I wonder you guys have been talking about the combination of apply controls at the data level, so this would provide for more enhanced security. What's the time to value? a team of engineers to apply controls to data as compared to other manually driven work That's really what we're seeing I mean, if you guys are interested in learning more, you know, reach out to us at info at iata Keep it right there from or cube content with.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
David | PERSON | 0.99+ |
Jeff Hammer | PERSON | 0.99+ |
John | PERSON | 0.99+ |
Eva Hora | PERSON | 0.99+ |
David Suresh | PERSON | 0.99+ |
Sabina | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Yusuf Khan | PERSON | 0.99+ |
Europe | LOCATION | 0.99+ |
London | LOCATION | 0.99+ |
2021 | DATE | 0.99+ |
two | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Dave Volonte | PERSON | 0.99+ |
Siri | TITLE | 0.99+ |
ORGANIZATION | 0.99+ | |
Fozzie | PERSON | 0.99+ |
2 | QUANTITY | 0.99+ |
five | QUANTITY | 0.99+ |
David Pleasure | PERSON | 0.99+ |
iata ho dot com | ORGANIZATION | 0.99+ |
Jay | PERSON | 0.99+ |
Five | QUANTITY | 0.99+ |
six step | QUANTITY | 0.99+ |
five benefits | QUANTITY | 0.99+ |
15 people | QUANTITY | 0.99+ |
Yousef | PERSON | 0.99+ |
$10 million | QUANTITY | 0.99+ |
This year | DATE | 0.99+ |
first step | QUANTITY | 0.99+ |
Ideo Tahoe | ORGANIZATION | 0.99+ |
last year | DATE | 0.99+ |
Andre | PERSON | 0.99+ |
hundreds | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
one cloud | QUANTITY | 0.99+ |
2011 | DATE | 0.99+ |
Tahoe | ORGANIZATION | 0.99+ |
Today | DATE | 0.99+ |
Noel | PERSON | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
Prem | ORGANIZATION | 0.99+ |
today | DATE | 0.99+ |
tonight | DATE | 0.99+ |
Io Tahoe | ORGANIZATION | 0.99+ |
second benefit | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
Iota A J. | ORGANIZATION | 0.99+ |
one step | QUANTITY | 0.99+ |
both | QUANTITY | 0.98+ |
third one | QUANTITY | 0.98+ |
Siris | TITLE | 0.98+ |
Aziz | PERSON | 0.98+ |
red hat | ORGANIZATION | 0.98+ |
each layer | QUANTITY | 0.98+ |
both businesses | QUANTITY | 0.98+ |
fourth idea | QUANTITY | 0.98+ |
apple | ORGANIZATION | 0.98+ |
1/5 step | QUANTITY | 0.98+ |
Toyota Ho | ORGANIZATION | 0.98+ |
first challenge | QUANTITY | 0.98+ |
41 | QUANTITY | 0.98+ |
azure | ORGANIZATION | 0.98+ |
Io Tahoe | PERSON | 0.98+ |
One person | QUANTITY | 0.98+ |
one location | QUANTITY | 0.98+ |
single | QUANTITY | 0.98+ |
Noah | PERSON | 0.98+ |
over 300 existing policies | QUANTITY | 0.98+ |
Iot Tahoe | ORGANIZATION | 0.98+ |
Thio | PERSON | 0.98+ |
Lenox | ORGANIZATION | 0.98+ |
two years ago | DATE | 0.98+ |
A. J A. Z. | PERSON | 0.98+ |
single point | QUANTITY | 0.98+ |
first thing | QUANTITY | 0.97+ |
Yussef | PERSON | 0.97+ |
Jupiter | LOCATION | 0.97+ |
second thing | QUANTITY | 0.97+ |
three things | QUANTITY | 0.97+ |
about 20 years | QUANTITY | 0.97+ |
single cloud | QUANTITY | 0.97+ |
First | QUANTITY | 0.97+ |
Suresh | PERSON | 0.97+ |
3 week | QUANTITY | 0.97+ |
each container | QUANTITY | 0.97+ |
each cloud platform | QUANTITY | 0.97+ |
Noah Fields and Sabita Davis | Io-Tahoe Enterprise Digital Resilience on Hybrid & Multicloud
>> Narrator: From around the globe, it's theCUBE presenting enterprise digital resilience on hybrid and multicloud brought to you by Io-Tahoe. >> Okay, now we're going to go into the demo and we want to get a better understanding of how you can leverage OpenShift and Io-Tahoe to facilitate faster application deployment. Let me pass the mic to Sabita, take it away. >> Thanks, Dave. Happy to be here again. >> Guys as Dave mentioned my name's Sabita Davis. I'm the Enterprise Account Executive here at Io-Tahoe. So today we just wanted to give you guys a general overview of how we're using OpenShift. >> Yeah, hey, I'm Noah, Io-Tahoe's Data Operations Engineer working with OpenShift and I've been learning the ins and outs of OpenShift for like the past few months. And I'm here to share what I've learned. >> Okay so before we begin I'm sure everybody wants to know Noah. What are the benefits of using OpenShift? >> Well, there's five that I can think of, faster time to operations, simplicity, automation, control and digital resilience. >> Okay, so that's really interesting because those are the exact same benefits that we at Io-Tahoe deliver to our customers. But let's start with faster time to operation, by running Io-Tahoe on OpenShift is it faster than let's say using Kubernetes and other platforms? >> Well, our objective at Io-Tahoe is to be accessible across multiple cloud platforms, right? And so by hosting our application in containers we're able to achieve this. So to answer your question it's faster to create end user application images using container tools like Kubernetes with OpenShift as compared to like Kubernetes with Docker, Kryo >> or Containerd. >> Okay, so we got a bit technical there. Can you explain that in a bit more detail? >> Yeah, there's a bit of vocabulary involved. So basically containers are used in developing things like databases, web servers or applications such as Io-Tahoe. What's great about containers is that they split the workload. So developers can select the libraries without breaking anything. And CIS admins can update the host without interrupting the programmers. Now OpenShift works hand-in-hand with Kubernetes to provide a way to build those containers for applications. >> Okay, got it. So basically containers make life easier for developers and system admins. So how does OpenShift differ from other platforms? >> Well, this kind of leads into the second benefit I want to talk about which is simplicity. Basically there's a lot of steps involved with when using Kubernetes with Docker but OpenShift simplifies this with their source to image process that takes the source code and turns it into a container image but that's not all. OpenShift has a lot of automation and features that simplify working with containers an important one being its web console. So here I've set up a light version of OpenShift called CodeReady Containers. And I was able to set up for application right from the web console. And I was able to set up this entire thing in Windows, Mac and Linux. So it's environment agnostic in that sense. >> Okay, so I think I see in the top left that this is a developer's view. What would a systems admin view look like? >> That's a good question. So here's the administrator view and this kind of ties into the benefit of control. This view gives insights into each one of the applications and containers that are running and you can make changes without affecting deployment. And you can also within this view set up each layer of security and there's multiple that you can prop up but I haven't fully messed around with it because since with my luck, I'd probably lock myself out. >> Okay, so that seems pretty secure. Is there a single point security such as you user login or are there multiple layers of security? >> Yeah, there are multiple layers of security. There's your user login, security groups and general role based access controls but there's also a ton of layers of security surrounding like the containers themselves. But for the sake of time, I won't get too far into it. >> Okay, so you mentioned simplicity and time to operation as being two of the benefits. You also briefly mentioned automation and as you know automation is the backbone of our platform here at Io-Tahoe. So that certainly grabbed my attention. Can you go a bit more in depth in terms of automation? >> OpenShift provides extensive automation that speeds up that time to operation, right? So the latest versions of OpenShift come with a built-in cryo container engine which basically means that you get to skip that container engine installation step. And you don't have to like log into each individual container hosts and configure networking, configure registry servers, storage, et cetera. So I'd say it automates the more boring kind of tedious processes. >> Okay, so I see the Io-Tahoe template there. What does it allow me to do? >> In terms of automation in application development. So we've created an OpenShift template which contains our application. This allows developers to instantly like set up a product within that template or within that, yeah. >> Okay, so Noah, last question. Speaking of vocabulary, you mentioned earlier digital resilience is a term we're hearing especially in the banking and finance world. It seems from what you described industries like banking and finance would be more resilient using OpenShift, correct? >> Yeah, in terms of digital resilience, OpenShift will give you better control over the consumption of resources each container is using. In addition, the benefit of containers is that like I mentioned earlier sysadmins can troubleshoot the servers without bringing down the application. And if the application does go down it's easy to bring it back up using the templates and like the other automation features that OpenShift provides. >> Okay, so thanks so much Noah. So any final thoughts you want to share? >> Yeah, I just want to give a quick recap of like the five benefits that you gain by using OpenShift. The five are time to operation, automation, control, security and simplicity. You can deploy applications faster, you can simplify the workload, you can automate a lot of the otherwise tedious processes, and maintain full control over your workflow and you can assert digital resilience within your environment. >> So guys, thanks for that appreciate the demo. I wonder you guys have been talking about the combination of Io-Tahoe and Red Hat. Can you tie that in Sabita to digital resilience specifically? >> Yeah, sure Dave. So when we speak to the benefits of security controls in terms of digital resilience at Io-Tahoe we automated detection and apply controls at the data level. So this would provide for more enhanced security. >> Okay, but so if you were to try to do all these things manually I mean, what does that do? How much time can I compress? What's the time to value? >> So with our latest versions of Io-Tahoe we're taking advantage of faster deployment time associated with containerization and Kubernetes. So this kind of speeds up the time it takes for customers start using our softwares. They'd be able to quickly spin up Io-Tahoe in their own on-premise environment or otherwise in their own cloud environment like including AWS, Azure, Oracle GCP and IBM cloud. Our quick start templates allow flexibility to deploy into multicloud environments all just using like a few clicks. >> Okay, so now I'll just quickly add, so what we've done Io-Tahoe here is we've really moved our customers away from the whole idea of needing a team of engineers to apply controls to data as compared to other manually driven workflows. So with templates, automation, pre-built policies and data controls one person can be fully operational within a few hours and achieve results straight out of the box on any cloud. >> Yeah, we've been talking about this theme of abstracting the complexity that's really what we're seeing is a major trend in this coming decade. Okay, great. Thanks Sabita, Noah. How can people get more information or if they have any follow up questions, where should they go? >> Yeah, sure Dave I mean if you guys are interested in learning more reach out to us @infoatiotahoe.com to speak with one of our sales engineers. I mean, we'd love to hear from you. So book a meeting as soon as you can. >> All right, thanks guys. Keep it right there for more cube content with Io-Tahoe. (gentle music)
SUMMARY :
brought to you by Io-Tahoe. Let me pass the mic to Happy to be here again. I'm the Enterprise Account and I've been learning the What are the benefits of using OpenShift? faster time to operations, simplicity, faster time to operation, So to answer your question Okay, so we got a bit technical there. So developers can select the libraries So basically containers make life easier that takes the source code Okay, so I think I see in the top left and there's multiple that you can prop up Okay, so that seems pretty secure. But for the sake of time, I and time to operation as So the latest versions of OpenShift Okay, so I see the This allows developers to instantly like especially in the banking And if the application does go down So any final thoughts you want to share? and you can assert digital resilience that appreciate the demo. controls at the data level. So with our latest versions of Io-Tahoe So with templates, automation, of abstracting the So book a meeting as soon as you can. cube content with Io-Tahoe.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave | PERSON | 0.99+ |
Sabita | PERSON | 0.99+ |
Noah | PERSON | 0.99+ |
Sabita Davis | PERSON | 0.99+ |
Io-Tahoe | ORGANIZATION | 0.99+ |
five | QUANTITY | 0.99+ |
five benefits | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
OpenShift | TITLE | 0.99+ |
each layer | QUANTITY | 0.99+ |
Noah Fields | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
each container | QUANTITY | 0.99+ |
Kubernetes | TITLE | 0.99+ |
Kryo | TITLE | 0.98+ |
Red Hat | ORGANIZATION | 0.98+ |
single point | QUANTITY | 0.98+ |
today | DATE | 0.97+ |
Oracle | ORGANIZATION | 0.97+ |
second benefit | QUANTITY | 0.97+ |
each one | QUANTITY | 0.97+ |
one | QUANTITY | 0.96+ |
Windows | TITLE | 0.95+ |
Docker | TITLE | 0.94+ |
CodeReady | TITLE | 0.93+ |
Io-Tahoe | TITLE | 0.92+ |
Linux | TITLE | 0.91+ |
this coming decade | DATE | 0.91+ |
each individual | QUANTITY | 0.9+ |
theCUBE | ORGANIZATION | 0.89+ |
one person | QUANTITY | 0.82+ |
Azure | ORGANIZATION | 0.79+ |
OpenShift | ORGANIZATION | 0.79+ |
Containerd | TITLE | 0.73+ |
@infoatiotahoe.com | OTHER | 0.71+ |
layers | QUANTITY | 0.7+ |
past | DATE | 0.6+ |
Mac | TITLE | 0.52+ |
Sabita | ORGANIZATION | 0.52+ |
Tahoe | TITLE | 0.51+ |
GCP | TITLE | 0.5+ |
Io | TITLE | 0.5+ |
Yusef Khan & Suresh Kanniappan | Io Tahoe Enterprise Digital Resilience on Hybrid & Multicloud
>>from around the globe. It's the Cube presenting enterprise, Digital resilience on hybrid and multi cloud Brought to You by Iota Ho. Okay, let's now get into the next segment where we'll explore data automation. But from the angle of digital resilience within and as a service consumption model, we're now joined by Yusuf Khan, who heads data services for Iota Ho and Shirish County. Up in Who's the vice president and head of U. S. Sales at happiest Minds. Gents, welcome to the program. Great to have you in the Cube. >>Thank you, David. >>Stretch. You guys talk about happiest minds. This notion of born digital, foreign agile. I like that. But talk about your mission at the company. >>Sure. A former in 2011 Happiest minds Up Born digital born a child company. >>The >>reason is that we are focused on customers. Our customer centric approach on delivering digitals and seamless solutions have helped us be in the race. Along with the Tier one providers, our mission, happiest people, happiest customers is focused to enable customer happiness through people happiness. We have Bean ranked among the top 25 I t services company in the great places to work serving hour glass to ratings off 4.1 against the rating off five is among the job in the Indian nineties services company that >>shows the >>mission on the culture. What we have built on the values, right sharing, mindful, integrity, learning and social on social responsibilities are the core values off our company on. That's where the entire culture of the company has been built. >>That's great. That sounds like a happy place to be. Now you have you head up data services for Iot Tahoe. We've talked in the past. Of course you're out of London. What do you what's your day to day focus with customers and partners? What you focused on? >>Well, David, my team work daily with customers and partners to help them better understand their data, improve their data quality, their data governance on help them make that data more accessible in a self service kind of way. To the stakeholders within those businesses on dis is all a key part of digital resilience that will will come on to talk about but later. You're >>right, e mean, that self service theme is something that we're gonna we're gonna really accelerate this decade, Yussef and so. But I wonder before we get into that, maybe you could talk about the nature of the partnership with happiest minds. You know, why do you guys choose toe work closely together? >>Very good question. Um, we see Io Tahoe on Happiest minds as a great mutual fit. A Suresh has said happiest minds are very agile organization. Um, I think that's one of the key things that attracts their customers on Io. Tahoe is all about automation. We're using machine learning algorithms to make data discovery data cataloging, understanding, data, redundancy, uh, much easier on. We're enabling customers and partners to do it much more quickly. So when you combine our emphasis on automation with the emphasis on agility, the happiest minds have that. That's a really nice combination. Work works very well together, very powerful. I think the other things that a key are both businesses, a serious have said are really innovative digital native type type companies. Um, very focused on newer technologies, the cloud etcetera, uh, on. Then finally, I think that both challenger brands Andi happiest minds have a really positive, fresh ethical approach to people and customers that really resonates with us that I have tied to its >>great thank you for that. So Russia, Let's get into the whole notion of digital resilience. I wanna I wanna sort of set it up with what I see. And maybe you can comment be prior to the pandemic. A lot of customers that kind of equated disaster recovery with their business continuance or business resilient strategy, and that's changed almost overnight. How have you seen your clients respond to that? What? I sometimes called the forced march to become a digital business. And maybe you could talk about some of the challenges that they faced along the way. >>Absolutely. So, uh, especially during this pandemic times when you see Dave customers have been having tough times managing their business. So happiest minds. Being a digital Brazilian company, we were able to react much faster in the industry, apart from the other services company. So one of the key things is the organizations trying to adopt onto the digital technologies right there has bean lot off data which has been to managed by these customers on. There have been lot off threats and risk, which has been to manage by the CEO Seo's so happiest minds digital resilient technology fight the where we're bringing the data complaints as a service, we were ableto manage the resilience much ahead off other competitors in the market. We were ableto bring in our business community processes from day one, where we were ableto deliver our services without any interruption to the services what we were delivering to our customers. >>So >>that is where the digital resilience with business community process enabled was very helpful for us who enable our customers continue there business without any interruptions during pandemics. >>So, I mean, some of the challenges that that customers tell me they obviously had to figure out how to get laptops to remote workers and that that whole remote, you know, work from home pivot figure out how to secure the end points. And, you know, those were kind of looking back there kind of table stakes, but it sounds like you've got a digital business means a data business putting data at the core, I like to say, but so I wonder if you could talk a little bit more about maybe the philosophy you have toward digital resilience in the specific approach you take with clients? >>Absolutely. They seen any organization data becomes. The key on this for the first step is to identify the critical data. Right. So we this is 1/6 process. What we following happiest minds. First of all, we take stock off the current state, though the customers think that they have a clear visibility off their data. How are we do more often assessment from an external point off view on See how critical their data is? Then we help the customers to strategies that right the most important thing is to identify the most important critical herself. Data being the most critical assault for any organization. Identification off the data's key for the customers. Then we help in building a viable operating model to ensure these identified critical assets are secure on monitor dearly so that they are consumed well as well as protected from external threats. Then, as 1/4 step, we try to bring in awareness, toe the people we train them at all levels in the organization. That is a P for people to understand the importance off the residual our cells. And then as 1/5 step, we work as a back up plan in terms of bringing in a very comprehensive and the holistic testing approach on people process as well as in technology. We'll see how the organization can withstand during a crisis time. And finally we do a continuous governance off this data, which is a key right. It is not just a one step process. We set up the environment. We do the initial analysis and set up the strategy on continuously govern this data to ensure that they are not only know managed will secure as well as they also have to meet the compliance requirements off the organization's right. That is where we help organizations toe secure on Meet the regulations off the organizations. As for the privacy laws, >>so >>this is a constant process. It's not on one time effort. We do a constant process because every organization goes towards the digital journey on. They have to face all these as part off the evolving environment on digital journey, and that's where they should be kept ready in terms off. No recovering, rebounding on moving forward if things goes wrong. >>So let's stick on that for a minute, and then I wanna bring yourself into the conversation. So you mentioned compliance and governance. When? When your digital business. Here, as you say, you're a data business. So that brings up issues. Data sovereignty. Uh, there's governance, this compliance. There's things like right to be forgotten. There's data privacy, so many things. These were often kind of afterthoughts for businesses that bolted on, if you will. I know a lot of executives are very much concerned that these air built in on, and it's not a one shot deal. So do you have solutions around compliance and governance? Can you deliver that as a service? Maybe you could talk about some of the specifics there, >>so some of way have offered multiple services. Tow our customers on digital race against. On one of the key service is the data complaints. As a service here we help organizations toe map the key data against the data compliance requirements. Some of the features includes in terms off the continuous discovery off data right, because organizations keep adding on data when they move more digital on helping the helping and understanding the actual data in terms off the residents of data, it could be a heterogeneous data sources. It could be on data basis or it could be even on the data lakes. Or it could be or no even on compromise, all the cloud environment. So identifying the data across the various no heterogeneous environment is very key. Feature off our solution. Once we identify, classify this sensitive data, the data privacy regulations on the traveling laws have to be map based on the business rules. So we define those rules on help map those data so that organizations know how critical their digital assets are. Then we work on a continuous marching off data for anomalies because that's one of the key teachers off the solution, which needs to be implemented on the day to day operational basis. So we're helping monitoring those anomalies off data for data quality management on an ongoing basis. And finally we also bringing the automatic data governance where we can manage the sensory data policies on their data relationships in terms off, mapping on manage their business rules on we drive reputations toe also suggest appropriate actions to the customers. Take on those specific data sets. >>Great. Thank you, Yousef. Thanks for being patient. I want to bring in Iota ho thio discussion and understand where your customers and happiest minds can leverage your data automation capability that you and I have talked about in the past. And I'm gonna be great if you had an example is well, but maybe you could pick it up from there. >>Sure. I mean, at a high level, assertions are clearly articulated. Really? Um, Iota, who delivers business agility. So that's by, um, accelerating the time to operationalize data, automating, putting in place controls and ultimately putting, helping put in place digital resilience. I mean, way if we step back a little bit in time, um, traditional resilience in relation to data are often met manually, making multiple copies of the same data. So you have a DB A. They would copy the data to various different places on business. Users would access it in those functional style owes. And of course, what happened was you ended up with lots of different copies off the same data around the enterprise. Very inefficient. Onda course ultimately, uh, increases your risk profile. Your risk of a data breach. Um, it's very hard to know where everything is, and I realized that expression they used David, the idea of the forced march to digital. So with enterprises that are going on this forced march, what they're finding is they don't have a single version of the truth, and almost nobody has an accurate view of where their critical data is. Then you have containers bond with containers that enables a big leap forward so you could break applications down into micro services. Updates are available via a P I s. And so you don't have the same need to build and to manage multiple copies of the data. So you have an opportunity to just have a single version of the truth. Then your challenge is, how do you deal with these large legacy data states that the service has been referring Thio, where you you have toe consolidate, and that's really where I Tahoe comes in. Um, we massively accelerate that process of putting in a single version of the truth into place. So by automatically discovering the data, um, discovering what's duplicate what's redundant, that means you can consolidate it down to a single trusted version much more quickly. We've seen many customers have tried to do this manually, and it's literally taken years using manual methods to cover even a small percentage of their I T estates with a tire. You could do it really very quickly on you can have tangible results within weeks and months. Um, and then you can apply controls to the data based on context. So who's the user? What's the content? What's the use case? Things like data quality validations or access permissions on. Then once you've done there, your applications and your enterprise are much more secure, much more resilient. As a result, you've got to do these things whilst retaining agility, though. So coming full circle. This is where the partnership with happiest minds really comes in as well. You've got to be agile. You've gotta have controls, um, on you've got a drug towards the business outcomes and it's doing those three things together that really deliver for the customer. Thank >>you. Use f. I mean you and I. In previous episodes, we've looked in detail at the business case. You were just talking about the manual labor involved. We know that you can't scale, but also there's that compression of time. Thio get to the next step in terms of ultimately getting to the outcome and we talked to a number of customers in the Cube. And the conclusion is really consistent that if you could accelerate the time to value, that's the key driver reducing complexity, automating and getting to insights faster. That's where you see telephone numbers in terms of business impact. So my question is, where should customers start? I mean, how can they take advantage of some of these opportunities that we've discussed >>today? Well, we've tried to make that easy for customers. So with our Tahoe and happiest minds, you can very quickly do what we call a data health check on. Dis is a is a 2 to 3 weeks process are two Really quickly start to understand and deliver value from your data. Um, so, iota, who deploys into the customer environment? Data doesn't go anywhere. Um, we would look at a few data sources on a sample of data Onda. We can very rapidly demonstrate how date discovery those catalog e understanding Jupiter data and redundant data can be done. Um, using machine learning, um, on how those problems can be solved. Um, and so what we tend to find is that we can very quickly as I say in a matter of a few weeks, show a customer how they could get toe, um, or Brazilian outcome on. Then how they can scale that up, take it into production on, then really understand their data state Better on build resilience into the enterprise. >>Excellent. There you have it. We'll leave it right there. Guys. Great conversation. Thanks so much for coming on the program. Best of luck to you in the partnership. Be well. >>Thank you, David. Sorry. Thank you. Thank >>you for watching everybody, This is Dave Volonte for the Cuban Are ongoing Siris on data Automation without Tahoe.
SUMMARY :
Great to have you in the Cube. But talk about your mission at the company. digital born a child company. I t services company in the great places to work serving hour glass to ratings mission on the culture. What do you what's your day to day focus To the stakeholders within those businesses on dis is all a key part of digital of the partnership with happiest minds. So when you combine our emphasis I sometimes called the forced march to become a digital business. So one of the key things that is where the digital resilience with business community process enabled was very putting data at the core, I like to say, but so I wonder if you could talk a little bit more about maybe for the first step is to identify the critical data. They have to face all these as part off the evolving environment So do you have solutions around compliance and governance? So identifying the data across the various no heterogeneous is well, but maybe you could pick it up from there. So by automatically discovering the data, um, And the conclusion is really consistent that if you could accelerate the time to value, So with our Tahoe and happiest minds, you can very quickly do what we call Best of luck to you in the partnership. Thank you. you for watching everybody, This is Dave Volonte for the Cuban Are ongoing Siris on data Automation without
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
David | PERSON | 0.99+ |
Yusuf Khan | PERSON | 0.99+ |
Yusef Khan | PERSON | 0.99+ |
2 | QUANTITY | 0.99+ |
London | LOCATION | 0.99+ |
Suresh Kanniappan | PERSON | 0.99+ |
Yousef | PERSON | 0.99+ |
one step | QUANTITY | 0.99+ |
Dave Volonte | PERSON | 0.99+ |
first step | QUANTITY | 0.99+ |
2011 | DATE | 0.99+ |
1/5 step | QUANTITY | 0.99+ |
4.1 | QUANTITY | 0.99+ |
Yussef | PERSON | 0.99+ |
Iot Tahoe | ORGANIZATION | 0.99+ |
both | QUANTITY | 0.99+ |
both businesses | QUANTITY | 0.98+ |
one | QUANTITY | 0.98+ |
two | QUANTITY | 0.98+ |
five | QUANTITY | 0.98+ |
single | QUANTITY | 0.98+ |
Dave | PERSON | 0.98+ |
1/6 | QUANTITY | 0.98+ |
today | DATE | 0.97+ |
3 weeks | QUANTITY | 0.97+ |
Suresh | PERSON | 0.97+ |
Jupiter | LOCATION | 0.96+ |
Io Tahoe | ORGANIZATION | 0.96+ |
one shot | QUANTITY | 0.96+ |
single version | QUANTITY | 0.96+ |
Russia | LOCATION | 0.96+ |
1/4 step | QUANTITY | 0.96+ |
First | QUANTITY | 0.96+ |
Siris | TITLE | 0.96+ |
Tahoe | PERSON | 0.94+ |
Cube | ORGANIZATION | 0.93+ |
Iota | ORGANIZATION | 0.92+ |
day one | QUANTITY | 0.9+ |
one time | QUANTITY | 0.88+ |
Iota Ho | ORGANIZATION | 0.87+ |
three things | QUANTITY | 0.85+ |
Brazilian | OTHER | 0.84+ |
Tier one | QUANTITY | 0.84+ |
forced | EVENT | 0.82+ |
Shirish County | LOCATION | 0.81+ |
Seo | PERSON | 0.81+ |
Cuban | OTHER | 0.81+ |
Tahoe | ORGANIZATION | 0.73+ |
Bean | PERSON | 0.72+ |
Iota | TITLE | 0.69+ |
pandemic | EVENT | 0.67+ |
U. S. Sales | ORGANIZATION | 0.66+ |
top 25 I t | QUANTITY | 0.64+ |
Thio | PERSON | 0.61+ |
Io | ORGANIZATION | 0.57+ |
Indian | OTHER | 0.55+ |
teachers | QUANTITY | 0.55+ |
Andi | PERSON | 0.54+ |
minute | QUANTITY | 0.53+ |
CEO | PERSON | 0.52+ |
Onda | LOCATION | 0.51+ |
Cube | COMMERCIAL_ITEM | 0.45+ |
service | QUANTITY | 0.45+ |
march | EVENT | 0.44+ |
nineties | DATE | 0.41+ |
Fadzi Ushewokunze and Ajay Vohora | Io Tahoe Enterprise Digital Resilience on Hybrid and Multicloud
>> Announcer: From around the globe, it's theCUBE presenting Enterprise Digital Resilience on Hybrid and multicloud brought to you by io/tahoe >> Hello everyone, and welcome to our continuing series covering data automation brought to you by io/tahoe. Today we're going to look at how to ensure enterprise resilience for hybrid and multicloud, let's welcome in Ajay Vohora who's the CEO of io/tahoe Ajay, always good to see you again, thanks for coming on. >> Great to be back David, pleasure. >> And he's joined by Fadzi Ushewokunze, who is a global principal architect for financial services, the vertical of financial services at Red Hat. He's got deep experiences in that sector. Welcome Fadzi, good to see you. >> Thank you very much. Happy to be here. >> Fadzi, let's start with you. Look, there are a lot of views on cloud and what it is. I wonder if you could explain to us how you think about what is a hybrid cloud and how it works. >> Sure, Yeah. So, a hybrid cloud is an IT architecture that incorporates some degree of workload portability, orchestration and management across multiple clouds. Those clouds could be private clouds or public clouds or even your own data centers. And how does it all work? It's all about secure interconnectivity and on demand allocation of resources across clouds. And separate clouds can become hybrid when you're seamlessly interconnected. And it is that interconnectivity that allows the workloads to be moved and how management can be unified and orchestration can work. And how well you have these interconnections has a direct impact of how well your hybrid cloud will work. >> Okay, so well Fadzi, staying with you for a minute. So, in the early days of cloud that term private cloud was thrown around a lot. But it often just meant virtualization of an on-prem system and a network connection to the public cloud. Let's bring it forward. What, in your view does a modern hybrid cloud architecture look like? >> Sure, so, for modern hybrid clouds we see that teams or organizations need to focus on the portability of applications across clouds. That's very important, right. And when organizations build applications they need to build and deploy these applications as a small collections of independently loosely coupled services. And then have those things run on the same operating system, which means in other words, running it all Linux everywhere and building cloud native applications and being able to manage it and orchestrate these applications with platforms like Kubernetes or Red Hat OpenShift, for example. >> Okay, so, Fadzi that's definitely different from building a monolithic application that's fossilized and doesn't move. So, what are the challenges for customers, you know, to get to that modern cloud is as you've just described it as it skillsets, is it the ability to leverage things like containers? What's your View there? >> So, I mean, from what we've seen around the industry especially around financial services where I spend most of my time. We see that the first thing that we see is management, right. Now, because you have all these clouds, you know, all these applications. You have a massive array of connections, of interconnections. You also have massive array of integrations portability and resource allocation as well. And then orchestrating all those different moving pieces things like storage networks. Those are really difficult to manage, right? So, management is the first challenge. The second one is workload placement. Where do you place this cloud? How do you place these cloud native operations? Do you, what do you keep on site on prem and what do you put in the cloud? That is the other challenge. The major one, the third one is security. Security now becomes the key challenge and concern for most customers. And we're going to talk about how to address that. >> Yeah, we're definitely going to dig into that. Let's bring Ajay into the conversation. Ajay, you know, you and I have talked about this in the past. One of the big problems that virtually every company face is data fragmentation. Talk a little bit about how io/tahoe, unifies data across both traditional systems, legacy systems and it connects to these modern IT environments. >> Yeah, sure Dave. I mean, a Fadzi just nailed it there. It used to be about data, the volume of data and the different types of data, but as applications become more connected and interconnected the location of that data really matters. How we serve that data up to those apps. So, working with Red Hat and our partnership with Red Hat. Being able to inject our data discovery machine learning into these multiple different locations. whether it be an AWS or an IBM cloud or a GCP or on prem. Being able to automate that discovery and pulling that single view of where is all my data, then allows the CIO to manage cost. They can do things like, one, I keep the data where it is, on premise or in my Oracle cloud or in my IBM cloud and connect the application that needs to feed off that data. And the way in which we do that is machine learning that learns over time as it recognizes different types of data, applies policies to classify that data and brings it all together with automation. >> Right, and one of the big themes that we've talked about this on earlier episodes is really simplification, really abstracting a lot of that heavy lifting away. So, we can focus on things Ajay, as you just mentioned. I mean, Fadzi, one of the big challenges that of course we all talk about is governance across these disparate data sets. I'm curious as your thoughts how does Red Hat really think about helping customers adhere to corporate edicts and compliance regulations? Which of course are particularly acute within financial services. >> Oh yeah, yes. So, for banks and payment providers like you've just mentioned there. Insurers and many other financial services firms, you know they have to adhere to a standard such as say a PCI DSS. And in Europe you've got the GDPR, which requires stringent tracking, reporting, documentation and, you know for them to, to remain in compliance. And the way we recommend our customers to address these challenges is by having an automation strategy, right. And that type of strategy can help you to improve the security on compliance of of your organization and reduce the risk out of the business, right. And we help organizations build security and compliance from the start with our consulting services, residencies. We also offer courses that help customers to understand how to address some of these challenges. And there's also, we help organizations build security into their applications with our open source middleware offerings and even using a platform like OpenShift, because it allows you to run legacy applications and also containerized applications in a unified platform. Right, and also that provides you with, you know with the automation and the tooling that you need to continuously monitor, manage and automate the systems for security and compliance purposes. >> Ajay, anything, any color you could add to this conversation? >> Yeah, I'm pleased Fadzi brought up OpenShift. I mean we're using OpenShift to be able to take that security application of controls to the data level and it's all about context. So, understanding what data is there, being able to assess it to say, who should have access to it, which application permission should be applied to it. That's a great combination of Red Hat and io/tahoe. >> Fadzi, what about multi-cloud? Doesn't that complicate the situation even further, maybe you could talk about some of the best practices to apply automation across not only hybrid cloud, but multi-cloud as well. >> Yeah, sure, yeah. So, the right automation solution, you know can be the difference between, you know cultivating an automated enterprise or automation carries. And some of the recommendations we give our clients is to look for an automation platform that can offer the first thing is complete support. So, that means have an automation solution that provides, you know, promotes IT availability and reliability with your platform so that, you can provide enterprise grade support, including security and testing integration and clear roadmaps. The second thing is vendor interoperability in that, you are going to be integrating multiple clouds. So, you're going to need a solution that can connect to multiple clouds seamlessly, right? And with that comes the challenge of maintainability. So, you're going to need to look into a automation solution that is easy to learn or has an easy learning curve. And then, the fourth idea that we tell our customers is scalability. In the hybrid cloud space, scale is the big, big deal here. And you need to deploy an automation solution that can span across the whole enterprise in a consistent manner, right. And then also that allows you finally to integrate the multiple data centers that you have. >> So, Ajay, I mean, this is a complicated situation for if a customer has to make sure things work on AWS or Azure or Google. They're going to spend all their time doing that. What can you add to really just simplify that multi-cloud and hybrid cloud equation. >> Yeah, I can give a few customer examples here. One being a manufacturer that we've worked with to drive that simplification. And the real bonuses for them has been a reduction in cost. We worked with them late last year to bring the cost spend down by $10 million in 2021. So, they could hit that reduced budget. And, what we brought to that was the ability to deploy using OpenShift templates into their different environments, whether it was on premise or in, as you mentioned, AWS. They had GCP as well for their marketing team and across those different platforms, being able to use a template, use prebuilt scripts to get up and running and catalog and discover that data within minutes. It takes away the legacy of having teams of people having to jump on workshop calls. And I know we're all on a lot of teams zoom calls. And in these current times. They're just simply using enough hours of the day to manually perform all of this. So, yeah, working with Red Hat, applying machine learning into those templates, those little recipes that we can put that automation to work regardless which location the data's in allows us to pull that unified view together. >> Great, thank you. Fadzi, I want to come back to you. So, the early days of cloud you're in the Big Apple, you know financial services really well. Cloud was like an evil word and within financial services, and obviously that's changed, it's evolved. We talk about the pandemic has even accelerated that. And when you really dug into it, when you talk to customers about their experiences with security in the cloud, it was not that it wasn't good, it was great, whatever, but it was different. And there's always this issue of skill, lack of skills and multiple tools, set up teams. are really overburdened. But in the cloud requires, you know, new thinking you've got the shared responsibility model. You've got to obviously have specific corporate, you know requirements and compliance. So, this is even more complicated when you introduce multiple clouds. So, what are the differences that you can share from your experiences running on a sort of either on prem or on a mono cloud or, you know, versus across clouds? What, do you suggest there? >> Sure, you know, because of these complexities that you have explained here mixed configurations and the inadequate change control are the top security threats. So, human error is what we want to avoid, because as you know, as your clouds grow with complexity then you put humans in the mix. Then the rate of errors is going to increase and that is going to expose you to security threats. So, this is where automation comes in, because automation will streamline and increase the consistency of your infrastructure management also application development and even security operations to improve in your protection compliance and change control. So, you want to consistently configure resources according to a pre-approved, you know, pre-approved policies and you want to proactively maintain them in a repeatable fashion over the whole life cycle. And then, you also want to rapidly the identify system that require patches and reconfiguration and automate that process of patching and reconfiguring. So that, you don't have humans doing this type of thing, And you want to be able to easily apply patches and change assistance settings according to a pre-defined base like I explained before, you know with the pre-approved policies. And also you want ease of auditing and troubleshooting, right. And from a Red Hat perspective we provide tools that enable you to do this. We have, for example a tool called Ansible that enables you to automate data center operations and security and also deployment of applications. And also OpenShift itself, it automates most of these things and obstruct the human beings from putting their fingers and causing, you know potentially introducing errors, right. Now, in looking into the new world of multiple clouds and so forth. The differences that we're seeing here between running a single cloud or on prem is three main areas, which is control, security and compliance, right. Control here, it means if you're on premise or you have one cloud you know, in most cases you have control over your data and your applications, especially if you're on prem. However, if you're in the public cloud, there is a difference that the ownership it is still yours, but your resources are running on somebody else's or the public clouds, EWS and so forth infrastructure. So, people that are going to do these need to really, especially banks and governments need to be aware of the regulatory constraints of running those applications in the public cloud. And we also help customers rationalize some of these choices. And also on security, you will see that if you're running on premises or in a single cloud you have more control, especially if you're on prem. You can control the sensitive information that you have. However, in the cloud, that's a different situation especially from personal information of employees and things like that. You need to be really careful with that. And also again, we help you rationalize some of those choices. And then, the last one is compliance. As well, you see that if you're running on prem on single cloud, regulations come into play again, right? And if you're running on prem, you have control over that. You can document everything, you have access to everything that you need, but if you're going to go to the public cloud again, you need to think about that. We have automation and we have standards that can help you you know, address some of these challenges. >> So, that's really strong insights, Fadzi. I mean, first of all Ansible has a lot of market momentum, you know, Red Hat's done a really good job with that acquisition. Your point about repeatability is critical, because you can't scale otherwise. And then, that idea you're putting forth about control, security and compliance. It's so true, I called it the shared responsibility model. And there was a lot of misunderstanding in the early days of cloud. I mean, yeah, maybe AWS is going to physically secure the you know, the S3, but in the bucket but we saw so many misconfigurations early on. And so it's key to have partners that really understand this stuff and can share the experiences of other clients. So, this all sounds great. Ajay, you're sharp, financial background. What about the economics? You know, our survey data shows that security it's at the top of the spending priority list, but budgets are stretched thin. I mean, especially when you think about the work from home pivot and all the areas that they had to, the holes that they had to fill there, whether it was laptops, you know, new security models, et cetera. So, how to organizations pay for this? What's the business case look like in terms of maybe reducing infrastructure costs, so I can pay it forward or there's a there's a risk reduction angle. What can you share there? >> Yeah, I mean, that perspective I'd like to give here is not being multi-cloud as multi copies of an application or data. When I think back 20 years, a lot of the work in financial services I was looking at was managing copies of data that were feeding different pipelines, different applications. Now, what we're seeing at io/tahoe a lot of the work that we're doing is reducing the number of copies of that data. So that, if I've got a product lifecycle management set of data, if I'm a manufacturer I'm just going to keep that at one location. But across my different clouds, I'm going to have best of breed applications developed in-house, third parties in collaboration with my supply chain, connecting securely to that single version of the truth. What I'm not going to do is to copy that data. So, a lot of what we're seeing now is that interconnectivity using applications built on Kubernetes that are decoupled from the data source. That allows us to reduce those copies of data within that you're gaining from a security capability and resilience, because you're not leaving yourself open to those multiple copies of data. And with that come cost of storage and a cost to compute. So, what we're saying is using multi-cloud to leverage the best of what each cloud platform has to offer. And that goes all the way to Snowflake and Heroku on a cloud managed databases too. >> Well and the people cost too as well. When you think about, yes, the copy creep. But then, you know, when something goes wrong a human has to come in and figure it out. You know, you brought up Snowflake, I get this vision of the data cloud, which is, you know data. I think we're going to be rethinking Ajay, data architectures in the coming decade where data stays where it belongs, it's distributed and you're providing access. Like you said, you're separating the data from the applications. Applications as we talked about with Fadzi, much more portable. So, it's really the last 10 years it'd be different than the next 10 years ago Ajay. >> Definitely, I think the people cost reduction is used. Gone are the days where you needed to have a dozen people governing, managing byte policies to data. A lot of that repetitive work, those tasks can be in part automated. We're seen examples in insurance where reduced teams of 15 people working in the back office, trying to apply security controls, compliance down to just a couple of people who are looking at the exceptions that don't fit. And that's really important because maybe two years ago the emphasis was on regulatory compliance of data with policies such as GDPR and CCPA. Last year, very much the economic effect to reduce head counts and enterprises running lean looking to reduce that cost. This year, we can see that already some of the more proactive companies are looking at initiatives, such as net zero emissions. How they use data to understand how they can become more, have a better social impact and using data to drive that. And that's across all of their operations and supply chain. So, those regulatory compliance issues that might have been external. We see similar patterns emerging for internal initiatives that are benefiting that environment, social impact, and of course costs. >> Great perspectives. Jeff Hammerbacher once famously said, the best minds of my generation are trying to get people to click on ads and Ajay those examples that you just gave of, you know social good and moving things forward are really critical. And I think that's where data is going to have the biggest societal impact. Okay guys, great conversation. Thanks so much for coming to the program. Really appreciate your time. >> Thank you. >> Thank you so much, Dave. >> Keep it right there, for more insight and conversation around creating a resilient digital business model. You're watching theCube. (soft music)
SUMMARY :
Ajay, always good to see you for financial services, the vertical Thank you very much. explain to us how you think And how well you have So, in the early days of cloud and being able to manage it and is it the ability to leverage We see that the first thing that we see One of the big problems that virtually And the way in which we do that is Right, and one of the And that type of strategy can help you to being able to assess it to say, some of the best practices can be the difference between, you know What can you add to really just simplify enough hours of the day that you can share to everything that you need, that security it's at the top And that goes all the way to Snowflake of the data cloud, you needed to have a dozen just gave of, you know Keep it right there, for
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Fadzi | PERSON | 0.99+ |
Jeff Hammerbacher | PERSON | 0.99+ |
Ajay Vohora | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
David | PERSON | 0.99+ |
Europe | LOCATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Last year | DATE | 0.99+ |
Ajay | PERSON | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
Fadzi Ushewokunze | PERSON | 0.99+ |
15 people | QUANTITY | 0.99+ |
2021 | DATE | 0.99+ |
This year | DATE | 0.99+ |
ORGANIZATION | 0.99+ | |
One | QUANTITY | 0.99+ |
GDPR | TITLE | 0.99+ |
$10 million | QUANTITY | 0.99+ |
fourth idea | QUANTITY | 0.99+ |
second thing | QUANTITY | 0.99+ |
OpenShift | TITLE | 0.99+ |
Ansible | ORGANIZATION | 0.99+ |
Linux | TITLE | 0.98+ |
two years ago | DATE | 0.98+ |
single | QUANTITY | 0.98+ |
third one | QUANTITY | 0.98+ |
io | ORGANIZATION | 0.98+ |
second one | QUANTITY | 0.98+ |
first challenge | QUANTITY | 0.98+ |
first thing | QUANTITY | 0.98+ |
EWS | ORGANIZATION | 0.97+ |
both | QUANTITY | 0.97+ |
next 10 years ago | DATE | 0.97+ |
Today | DATE | 0.97+ |
one | QUANTITY | 0.96+ |
one cloud | QUANTITY | 0.95+ |
single cloud | QUANTITY | 0.95+ |
late last year | DATE | 0.94+ |
pandemic | EVENT | 0.93+ |
each cloud platform | QUANTITY | 0.93+ |
Red Hat OpenShift | TITLE | 0.91+ |
a minute | QUANTITY | 0.91+ |
one location | QUANTITY | 0.91+ |
theCUBE | ORGANIZATION | 0.89+ |
Kubernetes | TITLE | 0.88+ |
io/tahoe | ORGANIZATION | 0.87+ |
three main areas | QUANTITY | 0.87+ |
Ansible | TITLE | 0.86+ |
CCPA | TITLE | 0.85+ |
zero emissions | QUANTITY | 0.83+ |
tahoe | ORGANIZATION | 0.81+ |
IBM | ORGANIZATION | 0.81+ |
a dozen people | QUANTITY | 0.79+ |
Snowflake | TITLE | 0.78+ |
Io Tahoe | PERSON | 0.75+ |
Azure | ORGANIZATION | 0.75+ |
last 10 years | DATE | 0.74+ |
20 years | QUANTITY | 0.74+ |
IBM cloud | ORGANIZATION | 0.72+ |
single version | QUANTITY | 0.71+ |
Red Hat | TITLE | 0.71+ |
S3 | COMMERCIAL_ITEM | 0.71+ |
Noah Fields and Sabita Davis | Io-Tahoe
>>From around the globe. It's the cube presenting enterprise digital resilience on hybrid and multicloud brought to you by IO Tahoe. Okay. Now we're going to go into the demo and we want to get a better understanding of how you can leverage OpenShift and IO Tahoe to facilitate faster application deployment. Let me pass the mic to Savita, take it away. >>Uh, thanks Dave. Happy to be here again. Um, guys, as they've mentioned, my name is to be the Davis. I'm the enterprise account executive here at IO Tahoe. Uh, so today we just wanted to give you guys a general overview of how we're using open shift. >>Yeah. Hey, I'm Noah IO. Tahoe's data operations engineer working with OpenShift, and I've been learning the ins and outs of OpenShift for like the past few months and I'm here to share it up line. >>Okay. So, so before we begin, I'm sure everybody wants to know Noah. What are the benefits of using OpenShift? >>Well, um, there's five that I can think of a faster time to operations, simplicity, automation control, and digital resilience. >>Okay. So, so that, that's really interesting because those are the exact same benefits that we at Aja Tahoe delivered to our customers. But, uh, let's start with faster time to operation by running IO Tahoe on OpenShift. Is it faster than let's say using Kubernetes and other platforms? >>Well, um, our objective at IO Tahoe has to be accessible across multiple cloud platforms, right? And so by hosting our application and containers, uh, we're able to achieve this. So to answer your question, it's faster to create end user application images, using container tools like Kubernetes with OpenShift as compared to like Kubernetes with Docker cryo or container D. >>Okay. So, so we got a bit technical there. Um, can you explain that in a bit more detail? >>Yeah, there's a bit of vocabulary involved. Uh, so basically containers are used in developing things like databases, web servers, or applications such as I've taught. What's great about containers is that they split the workload. So developers can select a libraries without breaking anything. And CIS admins can update the host without interrupting the programmers. Uh, now OpenShift works hand-in-hand with Kubernetes to provide a way to build those containers for applications. >>Okay, got it. Uh, so basically containers make life easier for developers and system admins. So how does OpenShift differ from other platforms? >>Um, well this kind of leads into the second benefit I want to talk about, which is simplicity. Basically. There's a lot of steps involved with when you're using Kubernetes with a Docker, but OpenShift simplifies this with their source to image process that takes the source code and turns it into a container image, but that's not all, uh, OpenShift has a lot of automation and features that simplify working with containers and important one being its web console. Um, so here I've set up a light version of OpenShift code ready containers. And I was able to set up our application right from the web console. And I was able to set up this entire thing in windows, Mac, and Linux. So it's environment agnostic in that sense. >>Okay. So I think I seen the top left. This is a developer's view. What would a systems admin view look like? >>That's a good question. So, uh, here's the, uh, administrator view and this kind of ties into the benefit of control. Um, this view gives insights into each one of the applications and containers that are running and you can make changes without affecting deployment. Um, and you can also within this view, set up each layer of security and there's multiple that you can prop up, but I haven't fully messed around with it because since with my look, I'd probably locked myself out. >>Okay. Um, so, so that seems pretty secure. Um, is there a single point security such as you use a login or are there multiple layers of security? Yeah. >>Um, there are multiple layers of security. There's your user login security groups and general role based access controls. Um, but there's also a ton of layers of security surrounding like the containers themselves. But for the sake of time, I won't get too far into it. >>Okay. Uh, so you mentioned simplicity and time to operation as being two of the benefits. You also briefly mentioned automation and as you know, automation is the backbone of our platform here at IO Tahoe. So that's certainly grabbed my attention. Can you go a bit more in depth in terms of automation? >>Yeah, sure. I'd say that automation is important benefit. Uh, OpenShift provides extensive automation that speeds up that time to operation, right? So the latest versions of open should come with a built-in cryo container engine, which basically means that you get to skip that container engine installation step. And you don't have to like log into each individual container hosts and configure networking, configure the registered servers, storage, et cetera. So I'd say, uh, it automates the more boring kind of tedious processes. >>Okay. So I see the iota template there. What does it allow me to do >>In terms of automation in application development? So we've created an OpenShift template, which contains our application. This allows developers to instantly like, um, set up a product within that template or within that. Yeah. >>Okay. Um, so Noah, last question. Speaking of vocabulary, you mentioned earlier digital resilience is a term we're hearing, especially in the banking and finance world. Um, it seems from what you described industries like banking and finance would be more resilient using OpenShift, correct? >>Yeah. In terms of digital resilience, OpenShift will give you better control over the consumption of resources each container is using. In addition, the benefit of containers is that, uh, like I mentioned earlier, CIS admins can troubleshoot the servers about like bringing down the application. And if the application does go down, it's easy to bring it back up using the templates and like the other automation features that OpenShift provides. >>Okay. So thanks so much. So any final thoughts you want to share? >>Yeah. Just want to give a quick recap of like the five benefits that you gain by using OpenShift. Uh, the five are time to operation automation, control, security and simplicity. Uh, you can deploy applications faster. You can simplify the workload. You can automate a lot of the otherwise tedious processes can maintain full control over your workflow and you can assert digital resilience within your environment. >>So guys, thanks for that. Appreciate the demo. Um, I wonder you guys have been talking about the combination of IO Tahoe and red hat. Can you tie that in Sabita to digital resilience specifically? >>Yeah, sure. Dave, um, so why don't we speak to the benefits of security controls in terms of digital resilience at Iowa hope? Uh, we automated detection and applied controls at the data level. So this would provide for more enhanced security. >>Okay. But so if you were to try to do all these things manually, I mean, what's, what does that do? How, how much time can I compress? What's the time to value? >>So, um, with our latest versions via Tahoe, we're taking advantage of faster deployment time, um, associated with containerization and Kubernetes. So this kind of speeds up the time it takes for customers to start using our software as they be able to quickly spin up a hotel and their own on-premise environment or otherwise in their own cloud environment, like including AWS or shore Oracle GCP and IBM cloud. Um, our quick start templates allow flexibility to deploy into multicloud environments, all just using like a few clicks. >>Okay. Um, so, so now I'll just quickly add, so what we've done, I Tahoe here is we've really moved our customers away from the whole idea of needing a team of engineers to apply controls to data as compared to other manually driven workflows. Uh, so with templates, automation, pre-built policies and data controls, one person can be fully operational within a few hours and achieve results straight out of the box, uh, on any cloud. >>Yeah. We've been talking about this theme of abstracting, the complexity that's really what we're seeing is a major trend in this coming decade. Okay, great. Thanks Savita Noah. Uh, ho how can people get more information or if they have any follow-up questions, where should they go? >>Yeah, sure. They've I mean, if you guys are interested in learning more, you know, reach out to us at info at dot com to speak with one of our sales engineers. I mean, we'd love to hear from you. So book a meeting as soon as you can. >>All right. Thanks guys. Keep it right there for more cube content with IO Tahoe.
SUMMARY :
resilience on hybrid and multicloud brought to you by IO Tahoe. so today we just wanted to give you guys a general overview of how we're using open shift. and I've been learning the ins and outs of OpenShift for like the past few months and I'm here to share it up line. What are the benefits of using OpenShift? Well, um, there's five that I can think of a faster time to operations, at Aja Tahoe delivered to our customers. So to answer your question, it's faster to create end user application Um, can you explain that in a bit more detail? Uh, so basically containers are used in Uh, so basically containers make life easier for developers and system Um, so here I've set up a light version of OpenShift code ready containers. This is a developer's view. Um, and you can also within this view, set up each layer of security and there's multiple that you can prop you use a login or are there multiple layers of security? But for the sake of time, I won't get too far into it. You also briefly mentioned automation and as you know, automation is the backbone of our platform here at IO Tahoe. So the latest versions of open should come with a built-in cryo container engine, What does it allow me to do This allows developers to instantly like, Um, it seems from what you described industries like banking and finance would be more resilient go down, it's easy to bring it back up using the templates and like the other automation features that OpenShift provides. So any final thoughts you want to share? Uh, the five are time to operation automation, Um, I wonder you guys have been talking about the combination So this would provide for more enhanced security. What's the time to value? So this kind of speeds up the time it takes for Uh, so with templates, Uh, ho how can people get more information or if they have any follow-up questions, where should they go? So book a meeting as soon as you can. Keep it right there for more cube content with IO Tahoe.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Aja Tahoe | ORGANIZATION | 0.99+ |
Dave | PERSON | 0.99+ |
IO Tahoe | ORGANIZATION | 0.99+ |
Sabita Davis | PERSON | 0.99+ |
two | QUANTITY | 0.99+ |
five | QUANTITY | 0.99+ |
five benefits | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
OpenShift | TITLE | 0.99+ |
each layer | QUANTITY | 0.99+ |
Kubernetes | TITLE | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
each container | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
second benefit | QUANTITY | 0.98+ |
Noah | PERSON | 0.98+ |
one | QUANTITY | 0.98+ |
Savita Noah | PERSON | 0.98+ |
Linux | TITLE | 0.98+ |
Noah Fields | PERSON | 0.98+ |
windows | TITLE | 0.97+ |
Oracle | ORGANIZATION | 0.97+ |
single point | QUANTITY | 0.97+ |
one person | QUANTITY | 0.96+ |
Iowa | LOCATION | 0.96+ |
Tahoe | ORGANIZATION | 0.96+ |
dot com | ORGANIZATION | 0.96+ |
Io-Tahoe | PERSON | 0.96+ |
CIS | ORGANIZATION | 0.95+ |
Savita | PERSON | 0.94+ |
iota | TITLE | 0.94+ |
IO | ORGANIZATION | 0.93+ |
each one | QUANTITY | 0.91+ |
Davis | PERSON | 0.89+ |
each individual container | QUANTITY | 0.8+ |
Docker cryo | TITLE | 0.79+ |
this coming decade | DATE | 0.78+ |
open | TITLE | 0.76+ |
OpenShift | ORGANIZATION | 0.73+ |
Tahoe | TITLE | 0.72+ |
GCP | TITLE | 0.67+ |
past few months | DATE | 0.64+ |
Mac | COMMERCIAL_ITEM | 0.59+ |
Tahoe | PERSON | 0.58+ |
container | TITLE | 0.58+ |
ton | QUANTITY | 0.56+ |
red | ORGANIZATION | 0.51+ |
benefits | QUANTITY | 0.43+ |
Sabita | COMMERCIAL_ITEM | 0.36+ |
Ajay Vohora and Lester Waters, Io-Tahoe | Io-Tahoe Adaptive Data Governance
>> Narrator: From around the globe its "theCUBE" presenting Adaptive Data Governance, brought to you by Io-Tahoe. >> And we're back with the Data Automation series. In this episode we're going to learn more about what Io-Tahoe is doing in the field of adaptive data governance, how can help achieve business outcomes and mitigate data security risks. I'm Lisa Martin and I'm joined by Ajay Vohora the CEO of Io-Tahoe, and Lester Waters the CTO of Io-Tahoe. Gentlemen it's great to have you on the program. >> Thank you Lisa is good to be back. >> Great to see you Lisa. >> Likewise, very seriously this isn't cautious as we are. Lester were going to start with you, what's going on at Io-Tahoe, what's new? >> Well, I've been with Io-Tahoe for a little over the year, and one thing I've learned is every customer needs are just a bit different. So we've been working on our next major release of the Io-Tahoe product and to really try to address these customer concerns because we want to be flexible enough in order to come in and not just profile the data and not just understand data quality and lineage, but also to address the unique needs of each and every customer that we have. And so that required a platform rewrite of our product so that we could extend the product without building a new version of the product, we wanted to be able to have pluggable modules. We are also focused a lot on performance, that's very important with the bulk of data that we deal with and we're able to pass through that data in a single pass and do the analytics that are needed whether it's a lineage data quality or just identifying the underlying data. And we're incorporating all that we've learned, we're tuning up our machine learning, we're analyzing on more dimensions than we've ever done before, we're able to do data quality without doing an initial reggie expert for example, just out of the box. So I think it's all of these things are coming together to form our next version of our product and We're really excited about. >> Sounds exciting, Ajay from the CEOs level what's going on? >> Wow, I think just building on that, what Lester just mentioned now it's we're growing pretty quickly with our partners, and today here with Oracle we're excited to explain how that's shaping up lots of collaboration already with Oracle, and government in insurance and in banking. And we're excited because we get to have an impact, it's really satisfying to see how we're able to help businesses transform and redefine what's possible with their data. And having Oracle there as a partner to lean in with is definitely helping. >> Excellent, we're going to dig into that a little bit later. Lester let's go back over to you, explain adaptive data governance, help us understand that. >> Really adaptive data governance is about achieving business outcomes through automation. It's really also about establishing a data-driven culture and pushing what's traditionally managed in IT out to the business. And to do that, you've got to enable an environment where people can actually access and look at the information about the data, not necessarily access the underlying data because we've got privacy concern system, but they need to understand what kind of data they have, what shape it's in, what's dependent on it upstream and downstream, and so that they can make their educated decisions on what they need to do to achieve those business outcomes. A lot of frameworks these days are hardwired, so you can set up a set of business rules, and that set of business rules works for a very specific database and a specific schema. But imagine a world where you could just say, you know, (tapping) the start date of a loan must always be before the end date of a loan, and having that generic rule regardless of the underlying database, and applying it even when a new database comes online and having those rules applied, that's what adaptive data governance about. I like to think of it as the intersection of three circles, really it's the technical metadata coming together with policies and rules, and coming together with the business ontologies that are unique to that particular business. And bringing this all together allows you to enable rapid change in your environment, so, it's a mouthful adaptive data governance, but that's what it kind of comes down to. >> So Ajay help me understand this, is this what enterprise companies are doing now or are they not quite there yet? >> Well, you know Lisa I think every organization is going at his pace, but markets are changing economy and the speed at which some of the changes in the economy happening is compelling more businesses to look at being more digital in how they serve their own customers. So what we're saying is a number of trends here from heads of data, chief data officers, CIO stepping back from a one size fits all approach because they've tried that before and it just hasn't worked. They've spent millions of dollars on IT programs trying to drive value from that data, and they've ended up with large teams of manual processing around data to try and hard-wire these policies to fit with the context and each line of business, and that hasn't worked. So, the trends that we're seeing emerge really relate to how do I as a chief data officer, as a CIO, inject more automation and to allow these common tasks. And we've been able to see that impact, I think the news here is if you're trying to create a knowledge graph, a data catalog, or a business glossary, and you're trying to do that manually, well stop, you don't have to do that manual anymore. I think best example I can give is Lester and I we like Chinese food and Japanese food, and if you were sitting there with your chopsticks you wouldn't eat a bowl of rice with the chopsticks one grain at a time, what you'd want to do is to find a more productive way to enjoy that meal before it gets cold. And that's similar to how we're able to help organizations to digest their data is to get through it faster, enjoy the benefits of putting that data to work. >> And if it was me eating that food with you guys I would be not using chopsticks I would be using a fork and probably a spoon. So Lester how then does Io-Tahoe go about doing this and enabling customers to achieve this? >> Let me show you a little story here. So if you take a look at the challenges that most customers have they're very similar, but every customer is on a different data journey, so, but it all starts with what data do I have, what shape is that data in, how is it structured, what's dependent on it upstream and downstream, what insights can I derive from that data, and how can I answer all of those questions automatically? So if you look at the challenges for these data professionals, you know, they're either on a journey to the cloud, maybe they're doing a migration to Oracle, maybe they're doing some data governance changes, and it's about enabling this. So if you look at these challenges, I'm going to take you through a story here, and I want to introduce Amanda. Amanda is not Latin like anyone in any large organizations, she is looking around and she just sees stacks of data, I mean, different databases the one she knows about, the ones she doesn't know about but should know about, various different kinds of databases, and Amanda is this tasking with understanding all of this so that they can embark on her data journey program. So Amanda goes through and she's great, (snaps finger) "I've got some handy tools, I can start looking at these databases and getting an idea of what we've got." But when she digs into the databases she starts to see that not everything is as clear as she might've hoped it would be. Property names or column names have ambiguous names like Attribute one and Attribute two, or maybe Date one and Date two, so Amanda is starting to struggle even though she's got tools to visualize and look at these databases, she's still knows she's got a long road ahead, and with 2000 databases in her large enterprise, yes it's going to be a long journey. But Amanda is smart, so she pulls out her trusty spreadsheet to track all of her findings, and what she doesn't know about she raises a ticket or maybe tries to track down in order to find what that data means, and she's tracking all this information, but clearly this doesn't scale that well for Amanda. So maybe the organization will get 10 Amanda's to sort of divide and conquer that work. But even that doesn't work that well 'cause there's still ambiguities in the data. With Io-Tahoe what we do is we actually profile the underlying data. By looking at the underlying data, we can quickly see that Attribute one looks very much like a US social security number, and Attribute two looks like a ICD 10 medical code. And we do this by using ontologies, and dictionaries, and algorithms to help identify the underlying data and then tag it. Key to doing this automation is really being able to normalize things across different databases so that where there's differences in column names, I know that in fact they contain the same data. And by going through this exercise with Io-Tahoe, not only can we identify the data, but we also can gain insights about the data. So for example, we can see that 97% of that time, that column named Attribute one that's got US social security numbers, has something that looks like a social security number. But 3% of the time it doesn't quite look right, maybe there's a dash missing, maybe there's a digit dropped, or maybe there's even characters embedded in it, that may be indicative of a data quality issues, so we try to find those kinds of things. Going a step further, we also try to identify data quality relationships. So for example we have two columns, one date one date two, through observation we can see the date one 99% of the time is less than date two, 1% of the time it's not, probably indicative of the data quality issue, but going a step further we can also build a business rule that says date one is actually than date two, and so then when it pops up again we can quickly identify and remediate that problem. So these are the kinds of things that we can do with Io-Tahoe. Going even a step further, we can take your favorite data science solution, productionize it, and incorporate it into our next version as what we call a worker process to do your own bespoke analytics. >> Bespoke analytics, excellent, Lester thank you. So Ajay, talk us through some examples of where you're putting this to use, and also what is some of the feedback from some customers. >> Yeah, what I'm thinking how do you bring into life a little bit Lisa lets just talk through a case study. We put something together, I know it's available for download, but in a well-known telecommunications media company, they have a lot of the issues that lasted just spoke about lots of teams of Amanda's, super bright data practitioners, and are maybe looking to get more productivity out of their day, and deliver a good result for their own customers, for cell phone subscribers and broadband users. So, there are so many examples that we can see here is how we went about auto generating a lot of that old understanding of that data within hours. So, Amanda had her data catalog populated automatically, a business glossary built up, and maybe I would start to say, "Okay, where do I want to apply some policies to the data to set in place some controls, whether I want to adapt how different lines of business maybe tasks versus customer operations have different access or permissions to that data." And what we've been able to do that is to build up that picture to see how does data move across the entire organization, across the state, and monitor that over time for improvement. So we've taken it from being like reactive, let's do something to fix something to now more proactive. We can see what's happening with our data, who's using it, who's accessing it, how it's being used, how it's being combined, and from there taking a proactive approach is a real smart use of the tanons in that telco organization and the folks that work there with data. >> Okay Ajay, so digging into that a little bit deeper, and one of the things I was thinking when you were talking through some of those outcomes that you're helping customers achieve is ROI. How do customers measure ROI, What are they seeing with Io-Tahoe solution? >> Yeah, right now the big ticket item is time to value. And I think in data a lot of the upfront investment costs are quite expensive, they happen today with a lot of the larger vendors and technologies. Well, a CIO, an economic buyer really needs to be certain about this, how quickly can I get that ROI? And I think we've got something that we can show just pull up a before and after, and it really comes down to hours, days, and weeks where we've been able to have that impact. And in this playbook that we put together the before and after picture really shows those savings that committed a bit through providing data into some actionable form within hours and days to drive agility. But at the same time being able to enforce the controls to protect the use of that data and who has access to it, so atleast the number one thing I'd have to say is time, and we can see that on the graphic that we've just pulled up here. >> Excellent, so ostensible measurable outcomes that time to value. We talk about achieving adaptive data governance. Lester, you guys talk about automation, you talk about machine learning, how are you seeing those technologies being a facilitator of organizations adopting adaptive data governance? >> Well, as we see the manual date, the days of manual effort are out, so I think this is a multi-step process, but the very first step is understanding what you have in normalizing that across your data estate. So, you couple this with the ontologies that are unique to your business and algorithms, and you basically go across it and you identify and tag that data, that allows for the next steps to happen. So now I can write business rules not in terms of named columns, but I can write them in terms of the tags. Using that automated pattern recognition where we observed the loan starts should be before the loan (indistinct), being able to automate that is a huge time saver, and the fact that we can suggest that as a rule rather than waiting for a person to come along and say, "Oh wow, okay, I need this rule, I need this rule." These are steps that increase, or I should say decrease that time to value that Ajay talked about. And then lastly, a couple of machine learning, because even with great automation and being able to profile all your data and getting a good understanding, that brings you to a certain point, but there's still ambiguity in the data. So for example I might have two columns date one and date two, I may have even observed that date one should be less than date two, but I don't really know what date one and date two are other than a date. So, this is where it comes in and I'm like, "As the user said, can you help me identify what date one and day two are in this table?" It turns out they're a start date and an end date for a loan, that gets remembered, cycled into machine learning step by step to see this pattern of date one date two. Elsewhere I'm going to say, "Is it start date and end date?" Bringing all these things together with all this automation is really what's key to enable this data database, your data governance program. >> Great, thanks Lester. And Ajay I do want to wrap things up with something that you mentioned in the beginning about what you guys are doing with Oracle, take us out by telling us what you're doing there, how are you guys working together? >> Yeah, I think those of us who worked in IT for many years we've learned to trust Oracle's technology that they're shifting now to a hybrid on-prem cloud generation 2 platform which is exciting, and their existing customers and new customers moving to Oracle are on a journey. So Oracle came to us and said, "Now, we can see how quickly you're able to help us change mindsets," and as mindsets are locked in a way of thinking around operating models of IT that are maybe not agile or more siloed, and they're wanting to break free of that and adopt a more agile API driven approach with their data. So, a lot of the work that we're doing with Oracle is around accelerating what customers can do with understanding their data and to build digital apps by identifying the underlying data that has value. And the time we're able to do that in hours, days, and weeks, rather than many months is opening up the eyes to chief data officers, CIO is to say, "Well, maybe we can do this whole digital transformation this year, maybe we can bring that forward and transform who we are as a company." And that's driving innovation which we're excited about, and I know Oracle keen to drive through. >> And helping businesses transform digitally is so incredibly important in this time as we look to things changing in 2021. Ajay and Lester thank you so much for joining me on this segment, explaining adaptive data governance, how organizations can use it, benefit from it, and achieve ROI, thanks so much guys. >> Thanks you. >> Thanks again Lisa. (bright music)
SUMMARY :
brought to you by Io-Tahoe. going to learn more about this isn't cautious as we are. and do the analytics that are needed to lean in with is definitely helping. Lester let's go back over to you, and so that they can make and to allow these common tasks. and enabling customers to achieve this? that we can do with Io-Tahoe. and also what is some of the in that telco organization and the folks and one of the things I was thinking and we can see that that time to value. that allows for the next steps to happen. that you mentioned in the beginning and I know Oracle keen to drive through. Ajay and Lester thank you Thanks again Lisa.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Ajay Vohora | PERSON | 0.99+ |
Amanda | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Ajay | PERSON | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
2021 | DATE | 0.99+ |
Lisa | PERSON | 0.99+ |
Lester | PERSON | 0.99+ |
one | QUANTITY | 0.99+ |
two columns | QUANTITY | 0.99+ |
97% | QUANTITY | 0.99+ |
Io-Tahoe | ORGANIZATION | 0.99+ |
Lester Waters | PERSON | 0.99+ |
3% | QUANTITY | 0.99+ |
Lester | ORGANIZATION | 0.99+ |
each line | QUANTITY | 0.99+ |
first step | QUANTITY | 0.98+ |
two columns | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
two | QUANTITY | 0.98+ |
millions of dollars | QUANTITY | 0.98+ |
1% | QUANTITY | 0.98+ |
telco | ORGANIZATION | 0.98+ |
one grain | QUANTITY | 0.97+ |
2000 databases | QUANTITY | 0.96+ |
US | LOCATION | 0.95+ |
ICD 10 | OTHER | 0.95+ |
one thing | QUANTITY | 0.94+ |
three circles | QUANTITY | 0.92+ |
Attribute two | OTHER | 0.92+ |
this year | DATE | 0.91+ |
single | QUANTITY | 0.91+ |
isa | PERSON | 0.9+ |
Latin | OTHER | 0.9+ |
each | QUANTITY | 0.87+ |
10 | QUANTITY | 0.85+ |
day | QUANTITY | 0.84+ |
Attribute one | OTHER | 0.84+ |
99% | QUANTITY | 0.83+ |
Io-Tahoe | PERSON | 0.79+ |
Data Automation | TITLE | 0.77+ |
Io- | ORGANIZATION | 0.74+ |
Lester Waters | PERSON | 0.72+ |
one size | QUANTITY | 0.71+ |
Io | ORGANIZATION | 0.69+ |
Sabita Davis and Patrick Zeimet | Io-Tahoe Adaptive Data Governance
>>from around the globe. It's the Cube presenting adaptive data governance brought >>to you by >>Iota Ho. In this next segment, we're gonna be talking to you about getting to know your data. And specifically you're gonna hear from two folks at Io Tahoe. We've got enterprise account execs Evita Davis here, as well as Enterprise Data engineer Patrick Simon. They're gonna be sharing insights and tips and tricks for how you can get to know your data and quickly on. We also want to encourage you to engage with Sabina and Patrick. Use the chat feature to the right, send comments, questions or feedback so you can participate. All right, Patrick Sabetta, take it away. All right. >>Thanks, Lisa. Great to be here as Lisa mentioned guys. I'm the enterprise account executive here in Ohio. Tahoe you Pat? >>Yeah. Hey, everyone so great to be here. A said My name's Patrick Samit. I'm the enterprise data engineer here at Iota Ho. And we're so excited to be here and talk about this topic as one thing we're really trying to perpetuate is that data is everyone's business. >>I couldn't agree more, Pat. So, guys, what patent? I patent. I've actually had multiple discussions with clients from different organizations with different roles. So we spoke with both your technical and your non technical audience. So while they were interested in different aspects of our platform, we found that what they had in common was they wanted to make data easy to understand and usable. So that comes back. The pats point off being everybody's business because no matter your role, we're all dependent on data. So what Pan I wanted to do today was wanted toe walk. You guys through some of those client questions, slash pain points that we're hearing from different industries and different roles and demo how our platform here, like Tahoe, is used for automating those, uh, automating Dozier related tasks. So with that said, are you ready for the first one, Pat? >>Yeah, Let's do it. >>Great. So I'm gonna put my technical hat on for this one, So I'm a data practitioner. I just started my job. ABC Bank. I have over 100 different data sources. So I have data kept in Data Lakes, legacy data, sources, even the cloud. So my issue is I don't know what those data sources hold. I don't know what data sensitive, and I don't even understand how that data is connected. So how can I talk to help? >>Yeah, I think that's a very common experience many are facing and definitely something I've encountered in my past. Typically, the first step is to catalog the data and then start mapping the relationships between your various data stores. Now, more often than not, this has tackled through numerous meetings and a combination of Excel and something similar to video, which are too great tools in their own part. But they're very difficult to maintain. Just due to the rate that we are creating data in the modern world. It starts to beg for an idea that can scale with your business needs. And this is where a platform like Io Tahoe becomes so appealing. You can see here visualization of the data relationships created by the I Ho Tahoe service. Now, what is fantastic about this is it's not only laid out in a very human and digestible format in the same action of creating this view, the data catalog was constructed. >>Um, So is the data catalog automatically populated? Correct. Okay, so So what? I'm using iota. Hope at what I'm getting is this complete, unified automated platform without the added cost, of course. >>Exactly. And that's at the heart of Iota Ho. A great feature with that data catalog is that Iota Ho will also profile your data as it creates the catalog, assigning some meaning to those pesky column Underscore ones and custom variable underscore tents that are always such a joy to deal with. Uh, now, by leveraging this interface, we can start to answer the first part of your question and understand where the core relationships within our data exists. Personally, I'm a big fan of this >>view, >>as it really just helps the i b naturally John to these focal points that coincide with these key columns following that train of thought. Let's examine the customer I D column that seems to be at the center of a lot of these relationships. We can see that it's a fairly important column as it's maintaining the relationship between at least three other tables. Now you notice all the connectors are in this blue color. This means that their system defined relationships. But I hope Tahoe goes that extra mile and actually creates thes orange colored connectors as well. These air ones that are machine learning algorithms have predicted to be relationships. Uh, and you can leverage to try and make new and powerful relationships within your data. So I hope that answers the first part of your question. >>Eso So this is really cool. And I can see how this could be leverage quickly. Now. What if I added new data sources or your multiple data sources and needed toe? Identify what data sensitive. Can I Oh, Tahoe, Detect that. >>Yeah, definitely. Within the i o ta platform. There already over 300 pre defined policies such as HIPAA, ferpa, C, c, p, a and the like. One can choose which of these policies to run against their data along for flexibility and efficiency and running the policies that affect organization. >>Okay, so so 300 is an exceptional number. I'll give you that. But what about internal policies that apply to my organization? Is there any ability for me to write custom policies? >>Yeah, that's no issue. And is something that clients leverage fairly often to utilize this function when simply has to write a rejects that our team has helped many deploy. After that, the custom policy is stored for future use to profile sensitive data. One then selects the data sources they're interested in and select the policies that meet your particular needs. The interface will automatically take your data according to the policies of detects, after which you can review the discoveries confirming or rejecting the tagging. All of these insights are easily exported through the interface, so one can work these into the action items within your project management systems. And I think this lends to the collaboration as a team can work through the discovery simultaneously. And as each item is confirmed or rejected, they can see it ni instantaneously. All this translates to a confidence that with iota how you can be sure you're in compliance. >>Um, so I'm glad you mentioned compliance because that's extremely important to my organization. >>So >>what you're saying when I use the eye a Tahoe automated platform, we'd be 90% more compliant that before were other than if you were going to be using a human. >>Yeah, definitely. The collaboration and documentation that the iota ho interface lends itself to can really help you build that confidence that your compliance is sound. >>Does >>that answer your question about sense of data? >>Definitely so. So path. I have the next question for you. So we're planning on migration on guy. Have a set of reports I need to migrate. But what I need to know is that well, what what data sources? Those report those reports are dependent on and what's feeding those tables? >>Yeah, it's a fantastic questions to be toe identifying critical data elements, and the interdependencies within the various databases could be a time consuming but vital process and the migration initiative. Luckily, Iota Ho does have an answer, and again, it's presented in a very visual format. >>So what I'm looking at here is my entire day landscape. >>Yes, exactly. >>So let's say I add another data source. I can still see that Unified 3 60 view. >>Yeah, One feature that is particularly helpful is the ability to add data sources after the data lineage. Discovery has finished along for the flexibility and scope necessary for any data migration project. If you only need need to select a few databases or your entirety, this service will provide the answers. You're looking for this visual representation of the connectivity makes the identification of critical data elements a simple matter. The connections air driven by both system defined flows as well as those predicted by our algorithms, the confidence of which, uh can actually be customized to make sure that they're meeting the needs of the initiative that you have in place. Now, this also provides tabular output in case you need it for your own internal documentation or for your action items, which we can see right here. Uh, in this interface, you can actually also confirm or deny the pair rejection the pair directions along to make sure that the data is as accurate as possible. Does that help with your data lineage needs? >>Definitely. So So, Pat, My next big question here is So now I know a little bit about my data. How do I know I can trust it? So what I'm interested in knowing really is is it in a fit state for Meteo use it? Is it accurate? Does it conform to the right format? >>Yeah, that's a great question. I think that is a pain point felt across the board, be it by data practitioners or data consumers alike. another service that iota hope provides is the ability to write custom data quality rules and understand how well the data pertains to these rules. This dashboard gives a unified view of the strength of these rules, and your dad is overall quality. >>Okay, so Pat s o on on the accuracy scores there. So if my marketing team needs to run, a campaign can read dependent those accuracy scores to know what what tables have quality data to use for our marketing campaign. >>Yeah, this view would allow you to understand your overall accuracy as well as dive into the minutia to see which data elements are of the highest quality. So for that marketing campaign, if you need everything in a strong form, you'll be able to see very quickly with these high level numbers. But if you're only dependent on a few columns to get that information out the door, you can find that within this view, uh, >>so you >>no longer have to rely on reports about reports, but instead just come to this one platform to help drive conversations between stakeholders and data practitioners. I hope that helps answer your questions about that quality. >>Oh, definitely. So I have another one for you here. Path. So I get now the value of IATA who brings by automatically captured all those technical metadata from sources. But how do we match that with the business glossary? >>Yeah, within the same data quality service that we just reviewed. One can actually add business rules detailing the definitions and the business domains that these fall into. What's more is that the data quality rules were just looking at can then be tied into these definitions, allowing insight into the strength of these business rules. It is this service that empowers stakeholders across the business to be involved with the data life cycle and take ownership over the rules that fall within their domain. >>Okay, so those custom rules can I apply that across data sources? >>Yeah. You can bring in as many data sources as you need, so long as you could tie them to that unified definition. >>Okay, great. Thanks so much bad. And we just want to quickly say to everyone working in data, we understand your pain, so please feel free to reach out >>to us. We >>are website the chapel. Oh, Arlington. And let's get a conversation started on how iota Who can help you guys automate all those manual task to help save you time and money. Thank you. Thank >>you. Erin. >>Impact. If I could ask you one quick question, how do you advise customers? You just walk in this great example This banking example that you and city to talk through. How do you advise customers get started? >>Yeah, I think the number one thing that customers could do to get started with our platform is to just run the tag discovery and build up that data catalog. It lends itself very quickly to the other needs you might have, such as thes quality rules as well as identifying those kind of tricky columns that might exist in your data. Those custom variable underscore tens I mentioned before >>last questions to be to anything to add to what Pat just described as a starting place. >>Um, no, I think actually passed something that pretty well, I mean, just just by automating all those manual tasks, I mean, it definitely can save your company a lot of time and money, so we we encourage you just reach out to us. Let's get that conversation started. >>Excellent. Savita and Pat, Thank you so much. We hope you have learned a lot from these folks about how to get to know your data. Make sure that it's quality so that you can maximize the value of it. Thanks for watching.
SUMMARY :
from around the globe. for how you can get to know your data and quickly on. I'm the enterprise account executive here in Ohio. I'm the enterprise data engineer here at Iota Ho. So we spoke with both your technical and your non technical So I have data kept in Data Lakes, legacy data, sources, even the cloud. Typically, the first step is to catalog the data and then start mapping the relationships Um, So is the data catalog automatically populated? Uh, now, by leveraging this interface, we can start to answer the first part of your question So I hope that answers the first part of your question. And I can see how this could be leverage quickly. to run against their data along for flexibility and efficiency and running the policies that affect organization. policies that apply to my organization? And I think this lends to the collaboration as a team can work through the discovery that before were other than if you were going to be using a human. interface lends itself to can really help you build that confidence that your compliance is I have the next question for you. Yeah, it's a fantastic questions to be toe identifying critical data elements, and the interdependencies within I can still see that Unified 3 60 view. Yeah, One feature that is particularly helpful is the ability to add data sources after the data Does it conform to the right format? hope provides is the ability to write custom data quality rules and understand how well the data needs to run, a campaign can read dependent those accuracy scores to know what what tables have quality Yeah, this view would allow you to understand your overall accuracy as well as dive into the minutia I hope that helps answer your questions about that quality. So I have another one for you here. to be involved with the data life cycle and take ownership over the rules that fall within their domain. so long as you could tie them to that unified definition. we understand your pain, so please feel free to reach out to us. help you guys automate all those manual task to help save you time and money. you. This banking example that you and city to talk through. Yeah, I think the number one thing that customers could do to get started with our platform so we we encourage you just reach out to us. Make sure that it's quality so that you can maximize the value of it.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Sabina | PERSON | 0.99+ |
Savita | PERSON | 0.99+ |
Pat | PERSON | 0.99+ |
Patrick | PERSON | 0.99+ |
Patrick Zeimet | PERSON | 0.99+ |
Patrick Simon | PERSON | 0.99+ |
Evita Davis | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
Ohio | LOCATION | 0.99+ |
ABC Bank | ORGANIZATION | 0.99+ |
Patrick Sabetta | PERSON | 0.99+ |
Sabita Davis | PERSON | 0.99+ |
I Ho Tahoe | ORGANIZATION | 0.99+ |
Patrick Samit | PERSON | 0.99+ |
90% | QUANTITY | 0.99+ |
Erin | PERSON | 0.99+ |
Excel | TITLE | 0.99+ |
each item | QUANTITY | 0.99+ |
first step | QUANTITY | 0.99+ |
two folks | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
Io Tahoe | ORGANIZATION | 0.98+ |
both | QUANTITY | 0.98+ |
first part | QUANTITY | 0.98+ |
John | PERSON | 0.98+ |
HIPAA | TITLE | 0.98+ |
first one | QUANTITY | 0.97+ |
iota | TITLE | 0.95+ |
one quick question | QUANTITY | 0.94+ |
ferpa | TITLE | 0.93+ |
Iota Ho | TITLE | 0.93+ |
Cube | ORGANIZATION | 0.93+ |
One feature | QUANTITY | 0.92+ |
IATA | ORGANIZATION | 0.92+ |
over 100 different data sources | QUANTITY | 0.9+ |
one | QUANTITY | 0.89+ |
one platform | QUANTITY | 0.88+ |
three other tables | QUANTITY | 0.86+ |
Pan | PERSON | 0.85+ |
Tahoe | ORGANIZATION | 0.84+ |
Iota Ho | TITLE | 0.84+ |
one thing | QUANTITY | 0.82+ |
Tahoe | PERSON | 0.82+ |
Iota Ho | ORGANIZATION | 0.75+ |
over 300 | QUANTITY | 0.74+ |
C | TITLE | 0.74+ |
both system | QUANTITY | 0.72+ |
at least | QUANTITY | 0.68+ |
Data Lakes | LOCATION | 0.68+ |
Meteo | ORGANIZATION | 0.64+ |
One | QUANTITY | 0.58+ |
Io-Tahoe | ORGANIZATION | 0.56+ |
Dozier | ORGANIZATION | 0.56+ |
p | TITLE | 0.52+ |
300 | OTHER | 0.48+ |
Arlington | PERSON | 0.41+ |
Tahoe | LOCATION | 0.4+ |
3 60 | OTHER | 0.38+ |
Santiago Castro, Gudron van der Wal and Yusef Khan | Io-Tahoe Adaptive Data Governance
>> Presenter: From around the globe, it's theCUBE. Presenting Adaptive Data Governance, brought to you by Io-Tahoe. >> Our next segment here is an interesting panel, you're going to hear from three gentlemen, about Adaptive Data Governance. We're going to talk a lot about that. Please welcome Yusef Khan, the global director of data services for Io-Tahoe. We also have Santiago Castor, the chief data officer at the First Bank of Nigeria, and Gudron Van Der Wal, Oracle's senior manager of digital transformation and industries. Gentlemen, it's great to have you joining us in this panel. (indistinct) >> All right, Santiago, we're going to start with you. Can you talk to the audience a little bit about the First Bank of Nigeria and its scale? This is beyond Nigeria, talk to us about that. >> Yes. So First Bank of Nigeria was created 125 years ago, it's one of the oldest, if not the oldest bank in Africa. And because of the history, it grew, everywhere in the region, and beyond the region. I'm currently based in London, where it's kind of the European headquarters. And it really promotes trade finance, institutional banking, corporate banking, private banking around the world, in particular in relationship to Africa. We are also in Asia, in the Middle East. And yes, and is a very kind of active bank in all these regions. >> So Santiago, talk to me about what adaptive data governance means to you, and how does it helps the First Bank of Nigeria to be able to innovate faster with the data that you have. >> Yes I like that concept of adaptive data governance, because it's kind of, I would say, an approach that can really happen today with the new technology before it was much more difficult to implement. So just to give you a little bit of context, I used to work in consulting for 16-17 years before joining the First Bank of Nigeria. And I saw many organizations trying to apply different type of approaches in data governance. And the beginning, early days was really kind of (indistinct), where you top down approach, where data governance was seen as implement a set of rules, policies and procedures, but really from the top down. And is important, it's important to have the battle of your sea level, of your director, whatever is, so just that way it fails, you really need to have a complimentary approach, I often say both amount, and actually, as a CEO I'm really trying to decentralized data governance, really instead of imposing a framework that some people in the business don't understand or don't care about it. It really needs to come from them. So what I'm trying to say is that, data basically support business objectives. And what you need to do is every business area needs information on particular decisions to actually be able to be more efficient, create value, et cetera. Now, depending on the business questions they have to show, they will need certain data sets. So they need actually to be able to have data quality for their own, 'çause now when they understand that, they become the stewards naturally of their own data sets. And that is where my bottom line is meeting my top down. You can guide them from the top, but they need themselves to be also in power and be actually in a way flexible to adapt the different questions that they have in order to be able to respond to the business needs. And I think that is where these adaptive data governance starts. Because if you want, I'll give you an example. In the bank, we work, imagine a Venn diagram. So we have information that is provided to finance, and all information to risk, or information for business development. And in this Venn diagram, there is going to be part of that every circle that are going to kind of intersect with each other. So what you want as a data governance is to help providing what is in common, and then let them do their own analysis to what is really related to their own area as an example, nationality. You will say in a bank that will open an account is the nationality of your customer, that's fine for final when they want to do a balance sheet an accounting or a P&L, but for risk, they want that type of analysis plus the net nationality of exposure, meaning where you are actually exposed as a risk, you can have a customer that are on hold in the UK, but then trade with Africa, and in Africa they're exposing their credit. So what I'm trying to say is they have these pieces in common and pieces that are different. Now I cannot impose a definition for everyone. I need them to adapt and to bring their answers to their own business questions. That is adaptive data governance. And all that is possible because we have and I was saying at the very beginning, just to finalize the point, we have new technologies that allow you to do these metadata classification in a very sophisticated way that you can actually create analytics of your metadata. You can understand your different data sources, in order to be able to create those classifications like nationalities and way of classifying your customers, your products, et cetera. But you will need to understand which areas need, what type nationality or classification, which others will mean that all the time. And the more you create that understanding, that intelligence about how your people are using your data you create in a way building blocks like a label, if you want. Where you provide them with those definitions, those catalogs you understand how they are used or you let them compose like Lego. They would play their way to build their analysis. And they will be adaptive. And I think the new technologies are allowing that. And this is a real game changer. I will say that over and over. >> So one of the things that you just said Santiago kind of struck me in to enable the users to be adaptive, they probably don't want to be logging in support ticket. So how do you support that sort of self service to meet the demand of the user so that they can be adaptive? >> Yeah, that's a really good question. And that goes along with that type of approach. I was saying in a way more and more business users want autonomy, and they want to basically be able to grab the data and answers their question. Now, when you have that, that's great, because then you have demand. The business is asking for data. They're asking for the insight. So how do you actually support that? I will say there is a changing culture that is happening more and more. I would say even the current pandemic has helped a lot into that because you have had, in a way, of course, technology is one of the biggest winners without technology we couldn't have been working remotely. Without this technology, where people can actually log in from their homes and still have a market data marketplaces where they self serve their information. But even beyond that, data is a big winner. Data because the pandemic has shown us that crisis happened, but we cannot predict everything and that we are actually facing a new kind of situation out of our comfort zone, where we need to explore and we need to adapt and we need to be flexible. How do we do that? With data. As a good example this, every country, every government, is publishing everyday data stats of what's happening in the countries with the COVID and the pandemic so they can understand how to react because this is new. So you need facts in order to learn and adapt. Now, the companies that are the same. Every single company either saw the revenue going down, or the revenue going very up for those companies that are very digital already now, it changed the reality. So they needed to adapt, but for that they needed information in order to think and innovate and try to create responses. So that type of self service of data, (indistinct) for data in order to be able to understand what's happening when the construct is changing, is something that is becoming more of the topic today because of the pandemic, because of the new capabilities of technologies that allow that. And then, you then are allowed to basically help, your data citizens, I call them in organization. People that know their business and can actually start playing and answer their own questions. So these technologies that gives more accessibility to the data, that gives some cataloging so we can understand where to go or where to find lineage and relationships. All this is basically the new type of platforms or tools that allow you to create what I call a data marketplace. So once you create that marketplace, they can play with it. And I was talking about new culture. And I'm going to finish with that idea. I think these new tools are really strong because they are now allowing for people that are not technology or IT people to be able to play with data because it comes in the digital world they are useful. I'll give you an example with all your stuff where you have a very interesting search functionality, where you want to find your data and you want to self serve, you go there in that search, and you actually go and look for your data. Everybody knows how to search in Google, everybody searching the internet. So this is part of the data culture, the digital culture, they know how to use those tools. Now similarly, that data marketplace is in Io-Tahoe, you can for example, see which data sources are mostly used. So when I'm doing an analysis, I see that police in my area are also using these sources so I trust those sources. We are a little bit like Amazon, when you might suggest you what next to buy, again this is the digital kind of culture where people very easily will understand. Similarly, you can actually like some type of data sets that are working, that's Facebook. So what I'm trying to say is you have some very easy user friendly technologies that allows you to understand how to interact with them. And then within the type of digital knowledge that you have, be able to self serve, play, collaborate with your peers, collaborate with the data query analysis. So its really enabling very easily that transition to become a data savvy without actually needing too much knowledge of IT, or coding, et cetera, et cetera. And I think that is a game changer as well. >> And enabling that speed that we're all demanding today during these unprecedented times. Gudron I wanted to go to you, as we talk about in the spirit of evolution, technology's changing. Talk to us a little bit about Oracle Digital. What are you guys doing there? >> Yeah, thank you. Well, Oracle Digital is a business unit at Oracle EMEA. And we focus on emerging countries, as well as low end enterprises in the mid market in more developed countries. And four years ago, they started with the idea to engage digital with our customers via central hubs across EMEA. That means engaging with video having conference calls, having a wall, agreeing wall, where we stand in front and engage with our customers. No one at that time could have foreseen how this is the situation today. And this helps us to engage with our customers in the way we're already doing. And then about my team. The focus of my team is to have early stage conversations with our customers on digital transformation and innovation. And we also have a team of industry experts who engage with our customers and share expertise across EMEA. And we we inspire our customers. The outcome of these conversations for Oracle is a deep understanding of our customer needs, which is very important. So we can help the customer and for the customer means that we will help them with our technology and our resources to achieve their goals. >> It's all about outcomes. Right Gudron? So in terms of automation, what are some of the things Oracle is doing there to help your clients leverage automation to improve agility so that they can innovate faster? Which on these interesting times it's demanding. >> Yeah. Thank you. Well, traditionally, Oracle is known for their databases, which has been innovated year over year since the first launch. And the latest innovation is the autonomous database and autonomous data warehouse. For our customers, this means a reduction in operational costs by 90%, with a multimodal converged database, and machine learning based automation for full lifecycle management. Our database is self driving. This means we automate database provisioning, tuning and scaling. The database is self securing. This means ultimate data protection and security and itself repairing the ultimate failure detection, failover and repair. And the question is for our customers, what does it mean? It means they can focus on their business instead of maintaining their infrastructure and their operations. >> That's absolutely critical. Yusef, I want to go over to you now. Some of the things that we've talked about, just the massive progression and technology, the evolution of that, but we know that whether we're talking about data management, or digital transformation. A one size fits all approach doesn't work to address the challenges that the business has. That the IT folks have. As you are looking to the industry, with what Santiago told us about First Bank of Nigeria, what are some of the changes that you're seeing that Io-Tahoe has seen throughout the industry? >> Well, Lisa, I think the first way I'd characterize it is to say, the traditional kind of top down approach to data, where you have almost a data policeman who tells you what you can and cannot do just doesn't work anymore. It's too slow, it's too result intensive. Data Management, data governance, digital transformation itself, it has to be collaborative. And it has to be an element of personalization today to users. In the environment we find ourselves in now, it has to be about enabling self service as well. A one size fits all model when it comes to those things around data doesn't work. As Santiago was saying, it needs to be adaptive to how the data is used and who is using it. And in order to do this, companies, enterprises, organizations really need to know their data. They need to understand what data they hold, where it is, and what the sensitivity of it is. They can then in a more agile way, apply appropriate controls and access so that people themselves are in groups within businesses are agile and can innovate. Otherwise, everything grinds to a halt, and you risk falling behind your competitors. >> Yet a one size fits all terms doesn't apply when you're talking about adaptive and agility. So we heard from Santiago about some of the impact that they're making with First Bank of Nigeria. Yusef, talk to us about some of the business outcomes that you're seeing other customers make leveraging automation that they could not do before. >> I guess one of the key ones is around. Just it's automatically being able to classify terabytes of data or even petabytes of data across different sources to find duplicates, which you can then remediate and delete. Now, with the capabilities that Io-Tahoe offers, and Oracle offers, you can do things not just with a five times or 10 times improvement, but it actually enables you to do project for stock that otherwise would fail, or you would just not be able to do. Classifying multi terabyte and multi petabyte estates across different sources, formats, very large volumes of data. In many scenarios, you just can't do that manually. We've worked with government departments. And the issues there as you'd expect are the result of fragmented data. There's a lot of different sources, there's a lot of different formats. And without these newer technologies to address it, with automation and machine learning, the project isn't doable. But now it is. And that could lead to a revolution in some of these businesses organizations. >> To enable that revolution now, there's got to be the right cultural mindset. And one, when Santiago was talking about those really kind of adopting that and I think, I always call that getting comfortably uncomfortable. But that's hard for organizations to do. The technology is here to enable that. But when you're talking with customers, how do you help them build the trust and the confidence that the new technologies and a new approaches can deliver what they need? How do you help drive that kind of attack in the culture? >> It's really good question, because it can be quite scary. I think the first thing we'd start with is to say, look, the technology is here, with businesses like Io-Tahoe, unlike Oracle, it's already arrived. What you need to be comfortable doing is experimenting, being agile around it and trying new ways of doing things. If you don't want to get left behind. And Santiago, and the team at FBN, are a great example of embracing it, testing it on a small scale and then scaling up. At Io-Tahoe we offer what we call a data health check, which can actually be done very quickly in a matter of a few weeks. So we'll work with the customer, pick a use case, install the application, analyze data, drive out some some quick wins. So we worked in the last few weeks of a large energy supplier. And in about 20 days, we were able to give them an accurate understanding of their critical data elements, help them apply data protection policies, minimize copies of the data, and work out what data they needed to delete to reduce their infrastructure spend. So it's about experimenting on that small scale, being agile, and then scaling up in a in a kind of very modern way. >> Great advice. Santiago, I'd like to go back to you. Is we kind of look at, again, that topic of culture, and the need to get that mindset there to facilitate these rapid changes. I want to understand kind of last question for you about how you're doing that. From a digital transformation perspective, we know everything is accelerating in 2020. So how are you building resilience into your data architecture and also driving that cultural change that can help everyone in this shift to remote working and a lot of the the digital challenges that we're all going through? >> That's a really interesting transition, I would say. I was mentioning, just going back to some of the points before to transition these I said that the new technologies allowed us to discover the data in a new way to blog and see very quickly information, to have new models of (indistinct) data, we are talking about data (indistinct), and giving autonomy to our different data units. Well, from that autonomy, they can then compose and innovate their own ways. So for me now we're talking about resilience. Because, in a way autonomy and flexibility in our organization, in our data structure, we platform gives you resilience. The organizations and the business units that I have experienced in the pandemic, are working well, are those that actually, because they're not physically present anymore in the office, you need to give them their autonomy and let them actually engage on their own side and do their own job and trust them in a away. And as you give them that they start innovating, and they start having a really interesting idea. So autonomy and flexibility, I think, is a key component of the new infrastructure where you get the new reality that pandemic shows that yes, we used to be very kind of structure, policies, procedures, as they're important, but now we learn flexibility and adaptability at the same site. Now, when you have that, a key other components of resiliency is speed, of course, people want to access the data and access it fast and decide fast, especially changes are changing so quickly nowadays, that you need to be able to, interact and iterate with your information to answer your questions quickly. And coming back maybe to where Yusef was saying, I completely agree is about experimenting, and iterate. You will not get it right the first time, especially that the world is changing too fast. And we don't have answers already set for everything. So we need to just go play and have ideas fail, fail fast, and then learn and then go for the next. So, technology that allows you to be flexible, iterate, and in a very fast agile way continue will allow you to actually be resilient in the way because you're flexible, you adapt, you are agile and you continue answering questions as they come without having everything said in a stroke that is too hard. Now coming back to your idea about the culture is changing in employees and in customers. Our employees, our customers are more and more digital service. And in a way the pandemic has accelerated that. We had many branches of the bank that people used to go to ask for things now they cannot go. You need to, here in Europe with the lockdown you physically cannot be going to the branches and the shops that have been closed. So they had to use our mobile apps. And we have to go into the internet banking, which is great, because that was the acceleration we wanted. Similarly, our employees needed to work remotely. So they needed to engage with a digital platform. Now what that means, and this is, I think the really strong point for the cultural change for resilience is that more and more we have two type of connectivity that is happening with data. And I call it employees connecting to data. The session we're talking about, employees connecting with each other, the collaboration that Yusef was talking about, which is allowing people to share ideas, learn and innovate. Because the more you have platforms where people can actually find themselves and play with the data, they can bring ideas to the analysis. And then employees actually connecting to algorithms. And this is the other part that is really interesting. We also are a partner of Oracle. And Oracle (indistinct) is great, they have embedded within the transactional system, many algorithms that are allowing us to calculate as the transactions happen. What happened there is that when our customers engage with algorithms, and again, with Io-Tahoe as well, the machine learning that is there for speeding the automation of how you find your data allows you to create an alliance with the machine. The machine is there to actually in a way be your best friend, to actually have more volume of data calculated faster in a way to discover more variety. And then, we couldn't cope without being connected to these algorithms. And then, we'll finally get to the last connection I was saying is, the customers themselves engaging with the connecting with the data. I was saying they're more and more engaging with our app and our website and they're digitally serving. The expectation of the customer has changed. I work in a bank where the industry is completely challenged. You used to have people going to a branch, as I was saying, they cannot not only not go there, but they're even going from branch to digital to ask to now even wanting to have business services actually in every single app that they are using. So the data becomes a service for them. The data they want to see how they spend their money and the data of their transactions will tell them what is actually their spending is going well with their lifestyle. For example, we talk about a normal healthy person. I want to see that I'm standing, eating good food and the right handle, healthy environment where I'm mentally engaged. Now all these is metadata is knowing how to classify your data according to my values, my lifestyle, is algorithms I'm coming back to my three connections, is the algorithms that allow me to very quickly analyze that metadata. And actually my staff in the background, creating those understanding of the customer journey to give them service that they expect on a digital channel, which is actually allowing them to understand how they are engaging with financial services. >> Engagement is absolutely critical Santiago. Thank you for sharing that. I do want to wrap really quickly. Gudron one last question for you. Santiago talked about Oracle, you've talked about it a little bit. As we look at digital resilience, talk to us a little bit in the last minute about the evolution of Oracle, what you guys are doing there to help your customers get the resilience that they have to have to be. To not just survive, but thrive. >> Yeah. Well, Oracle has a cloud offering for infrastructure, database, platform service, and the complete solutions offered at SaaS. And as Santiago also mentioned, we are using AI across our entire portfolio, and by this will help our customers to focus on their business innovation and capitalize on data by enabling your business models. And Oracle has a global coverage with our cloud regions. It's massively investing in innovating and expanding their cloud. And by offering cloud as public cloud in our data centers, and also as private clouds with clouded customer, we can meet every sovereignty and security requirement. And then this way, we help people to see data in new ways. We discover insights and unlock endless possibilities. And maybe one one of my takeaways is, if I speak with customers, I always tell them, you better start collecting your data now. We enable this by this like Io-Tahoe help us as well. If you collect your data now you are ready for tomorrow. You can never collect your data backwards. So that is my takeaway for today. >> You can't collect your data backwards. Excellent Gudron. Gentlemen, thank you for sharing all of your insights, very informative conversation. All right. This is theCUBE, the leader in live digital tech coverage. (upbeat music)
SUMMARY :
brought to you by Io-Tahoe. Gentlemen, it's great to have going to start with you. And because of the history, it grew, So Santiago, talk to me about So just to give you a that you just said Santiago And I'm going to finish with that idea. And enabling that speed and for the customer means to help your clients leverage automation and itself repairing the that the business has. And in order to do this, of the business outcomes And that could lead to a revolution and the confidence that And Santiago, and the team and the need to get that of the customer journey to give them they have to have to be. and the complete the leader in live digital tech coverage.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Yusef Khan | PERSON | 0.99+ |
Yusef | PERSON | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
London | LOCATION | 0.99+ |
Europe | LOCATION | 0.99+ |
Io-Tahoe | ORGANIZATION | 0.99+ |
Gudron Van Der Wal | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
Asia | LOCATION | 0.99+ |
10 times | QUANTITY | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Africa | LOCATION | 0.99+ |
five times | QUANTITY | 0.99+ |
Santiago Castro | PERSON | 0.99+ |
Santiago Castor | PERSON | 0.99+ |
2020 | DATE | 0.99+ |
UK | LOCATION | 0.99+ |
FBN | ORGANIZATION | 0.99+ |
Oracle Digital | ORGANIZATION | 0.99+ |
Lego | ORGANIZATION | 0.99+ |
Santiago | PERSON | 0.99+ |
Middle East | LOCATION | 0.99+ |
90% | QUANTITY | 0.99+ |
Gudron | PERSON | 0.99+ |
Gudron van der Wal | PERSON | 0.99+ |
first | QUANTITY | 0.99+ |
First Bank of Nigeria | ORGANIZATION | 0.99+ |
tomorrow | DATE | 0.99+ |
ORGANIZATION | 0.99+ | |
both | QUANTITY | 0.99+ |
First Bank of Nigeria | ORGANIZATION | 0.98+ |
COVID | EVENT | 0.98+ |
one | QUANTITY | 0.98+ |
Io-Tahoe | LOCATION | 0.98+ |
four years ago | DATE | 0.98+ |
today | DATE | 0.98+ |
two type | QUANTITY | 0.97+ |
pandemic | EVENT | 0.97+ |
about 20 days | QUANTITY | 0.97+ |
16-17 years | QUANTITY | 0.97+ |
first time | QUANTITY | 0.97+ |
first launch | QUANTITY | 0.97+ |
first way | QUANTITY | 0.96+ |
three gentlemen | QUANTITY | 0.96+ |
ORGANIZATION | 0.95+ | |
125 years ago | DATE | 0.95+ |
Nigeria | LOCATION | 0.95+ |
one size | QUANTITY | 0.94+ |
IO TAHOE EPISODE 4 DATA GOVERNANCE V2
>>from around the globe. It's the Cube presenting adaptive data governance brought to you by Iota Ho. >>And we're back with the data automation. Siri's. In this episode, we're gonna learn more about what I owe Tahoe is doing in the field of adaptive data governance how it can help achieve business outcomes and mitigate data security risks. I'm Lisa Martin, and I'm joined by a J. Bihar on the CEO of Iot Tahoe and Lester Waters, the CEO of Bio Tahoe. Gentlemen, it's great to have you on the program. >>Thank you. Lisa is good to be back. >>Great. Staley's >>likewise very socially distant. Of course as we are. Listen, we're gonna start with you. What's going on? And I am Tahoe. What's name? Well, >>I've been with Iot Tahoe for a little over the year, and one thing I've learned is every customer needs air just a bit different. So we've been working on our next major release of the I O. Tahoe product. But to really try to address these customer concerns because, you know, we wanna we wanna be flexible enough in order to come in and not just profile the date and not just understand data quality and lineage, but also to address the unique needs of each and every customer that we have. And so that required a platform rewrite of our product so that we could, uh, extend the product without building a new version of the product. We wanted to be able to have plausible modules. We also focused a lot on performance. That's very important with the bulk of data that we deal with that we're able to pass through that data in a single pass and do the analytics that are needed, whether it's, uh, lineage, data quality or just identifying the underlying data. And we're incorporating all that we've learned. We're tuning up our machine learning we're analyzing on MAWR dimensions than we've ever done before. We're able to do data quality without doing a Nen initial rejects for, for example, just out of the box. So I think it's all of these things were coming together to form our next version of our product. We're really excited by it, >>So it's exciting a J from the CEO's level. What's going on? >>Wow, I think just building on that. But let's still just mentioned there. It's were growing pretty quickly with our partners. And today, here with Oracle are excited. Thio explain how that shaping up lots of collaboration already with Oracle in government, in insurance, on in banking and we're excited because we get to have an impact. It's real satisfying to see how we're able. Thio. Help businesses transform, Redefine what's possible with their data on bond. Having I recall there is a partner, uh, to lean in with is definitely helping. >>Excellent. We're gonna dig into that a little bit later. Let's let's go back over to you. Explain adaptive data governance. Help us understand that >>really adaptive data governance is about achieving business outcomes through automation. It's really also about establishing a data driven culture and pushing what's traditionally managed in I t out to the business. And to do that, you've got to you've got Thio. You've got to enable an environment where people can actually access and look at the information about the data, not necessarily access the underlying data because we've got privacy concerns itself. But they need to understand what kind of data they have, what shape it's in what's dependent on it upstream and downstream, and so that they could make their educated decisions on on what they need to do to achieve those business outcomes. >>Ah, >>lot of a lot of frameworks these days are hardwired, so you can set up a set of business rules, and that set of business rules works for a very specific database and a specific schema. But imagine a world where you could just >>say, you >>know, the start date of alone must always be before the end date of alone and having that generic rule, regardless of the underlying database and applying it even when a new database comes online and having those rules applied. That's what adaptive data governance about I like to think of. It is the intersection of three circles, Really. It's the technical metadata coming together with policies and rules and coming together with the business ontology ease that are that are unique to that particular business. And this all of this. Bringing this all together allows you to enable rapid change in your environment. So it's a mouthful, adaptive data governance. But that's what it kind of comes down to. >>So, Angie, help me understand this. Is this book enterprise companies are doing now? Are they not quite there yet. >>Well, you know, Lisa, I think every organization is is going at its pace. But, you know, markets are changing the economy and the speed at which, um, some of the changes in the economy happening is is compelling more businesses to look at being more digital in how they serve their own customers. Eh? So what we're seeing is a number of trends here from heads of data Chief Data Officers, CEO, stepping back from, ah, one size fits all approach because they've tried that before, and it it just hasn't worked. They've spent millions of dollars on I T programs China Dr Value from that data on Bennett. And they've ended up with large teams of manual processing around data to try and hardwire these policies to fit with the context and each line of business and on that hasn't worked. So the trends that we're seeing emerge really relate. Thio, How do I There's a chief data officer as a CEO. Inject more automation into a lot of these common tax. Andi, you know, we've been able toc that impact. I think the news here is you know, if you're trying to create a knowledge graph a data catalog or Ah, business glossary. And you're trying to do that manually will stop you. You don't have to do that manually anymore. I think best example I can give is Lester and I We we like Chinese food and Japanese food on. If you were sitting there with your chopsticks, you wouldn't eat the bowl of rice with the chopsticks, one grain at a time. What you'd want to do is to find a more productive way to to enjoy that meal before it gets cold. Andi, that's similar to how we're able to help the organizations to digest their data is to get through it faster, enjoy the benefits of putting that data to work. >>And if it was me eating that food with you guys, I would be not using chopsticks. I would be using a fork and probably a spoon. So eso Lester, how then does iota who go about doing this and enabling customers to achieve this? >>Let me, uh, let me show you a little story have here. So if you take a look at the challenges the most customers have, they're very similar, but every customers on a different data journey, so but it all starts with what data do I have? What questions or what shape is that data in? Uh, how is it structured? What's dependent on it? Upstream and downstream. Um, what insights can I derive from that data? And how can I answer all of those questions automatically? So if you look at the challenges for these data professionals, you know, they're either on a journey to the cloud. Maybe they're doing a migration oracle. Maybe they're doing some data governance changes on bits about enabling this. So if you look at these challenges and I'm gonna take you through a >>story here, E, >>I want to introduce Amanda. Man does not live like, uh, anyone in any large organization. She's looking around and she just sees stacks of data. I mean, different databases, the one she knows about, the one she doesn't know about what should know about various different kinds of databases. And a man is just tasking with understanding all of this so that they can embark on her data journey program. So So a man who goes through and she's great. I've got some handy tools. I can start looking at these databases and getting an idea of what we've got. Well, as she digs into the databases, she starts to see that not everything is as clear as she might have hoped it would be. You know, property names or column names, or have ambiguous names like Attribute one and attribute to or maybe date one and date to s Oh, man is starting to struggle, even though she's get tools to visualize. And look what look at these databases. She still No, she's got a long road ahead. And with 2000 databases in her large enterprise, yes, it's gonna be a long turkey but Amanda Smart. So she pulls out her trusty spreadsheet to track all of her findings on what she doesn't know about. She raises a ticket or maybe tries to track down the owner to find what the data means. And she's tracking all this information. Clearly, this doesn't scale that well for Amanda, you know? So maybe organization will get 10 Amanda's to sort of divide and conquer that work. But even that doesn't work that well because they're still ambiguities in the data with Iota ho. What we do is we actually profile the underlying data. By looking at the underlying data, we can quickly see that attribute. One looks very much like a U. S. Social Security number and attribute to looks like a I c D 10 medical code. And we do this by using anthologies and dictionaries and algorithms to help identify the underlying data and then tag it. Key Thio Doing, uh, this automation is really being able to normalize things across different databases, so that where there's differences in column names, I know that in fact, they contain contain the same data. And by going through this exercise with a Tahoe, not only can we identify the data, but we also could gain insights about the data. So, for example, we can see that 97% of that time that column named Attribute one that's got us Social Security numbers has something that looks like a Social Security number. But 3% of the time, it doesn't quite look right. Maybe there's a dash missing. Maybe there's a digit dropped. Or maybe there's even characters embedded in it. So there may be that may be indicative of a data quality issues, so we try to find those kind of things going a step further. We also try to identify data quality relationships. So, for example, we have two columns, one date, one date to through Ah, observation. We can see that date 1 99% of the time is less than date, too. 1% of the time. It's not probably indicative of a data quality issue, but going a step further, we can also build a business rule that says Day one is less than date to. And so then when it pops up again, we can quickly identify and re mediate that problem. So these are the kinds of things that we could do with with iota going even a step further. You could take your your favorite data science solution production ISAT and incorporated into our next version a zey what we call a worker process to do your own bespoke analytics. >>We spoke analytics. Excellent, Lester. Thank you. So a J talk us through some examples of where you're putting this to use. And also what is some of the feedback from >>some customers? But I think it helped do this Bring it to life a little bit. Lisa is just to talk through a case study way. Pull something together. I know it's available for download, but in ah, well known telecommunications media company, they had a lot of the issues that lasted. You spoke about lots of teams of Amanda's, um, super bright data practitioners, um, on baby looking to to get more productivity out of their day on, deliver a good result for their own customers for cell phone subscribers, Um, on broadband users. So you know that some of the examples that we can see here is how we went about auto generating a lot of that understanding off that data within hours. So Amanda had her data catalog populated automatically. A business class three built up on it. Really? Then start to see. Okay, where do I want Thio? Apply some policies to the data to to set in place some controls where they want to adapt, how different lines of business, maybe tax versus customer operations have different access or permissions to that data on What we've been able to do there is, is to build up that picture to see how does data move across the entire organization across the state. Andi on monitor that overtime for improvement, so have taken it from being a reactive. Let's do something Thio. Fix something. Thio, Now more proactive. We can see what's happening with our data. Who's using it? Who's accessing it, how it's being used, how it's being combined. Um, on from there. Taking a proactive approach is a real smart use of of the talents in in that telco organization Onda folks that worked there with data. >>Okay, Jason, dig into that a little bit deeper. And one of the things I was thinking when you were talking through some of those outcomes that you're helping customers achieve is our ally. How do customers measure are? Why? What are they seeing with iota host >>solution? Yeah, right now that the big ticket item is time to value on. And I think in data, a lot of the upfront investment cause quite expensive. They have been today with a lot of the larger vendors and technologies. So what a CEO and economic bio really needs to be certain of is how quickly can I get that are away. I think we've got something we can show. Just pull up a before and after, and it really comes down to hours, days and weeks. Um, where we've been able Thio have that impact on in this playbook that we pulled together before and after picture really shows. You know, those savings that committed a bit through providing data into some actionable form within hours and days to to drive agility, but at the same time being out and forced the controls to protect the use of that data who has access to it. So these are the number one thing I'd have to say. It's time on. We can see that on the the graphic that we've just pulled up here. >>We talk about achieving adaptive data governance. Lester, you guys talk about automation. You talk about machine learning. How are you seeing those technologies being a facilitator of organizations adopting adaptive data governance? Well, >>Azaz, we see Mitt Emmanuel day. The days of manual effort are so I think you know this >>is a >>multi step process. But the very first step is understanding what you have in normalizing that across your data estate. So you couple this with the ontology, that air unique to your business. There is no algorithms, and you basically go across and you identify and tag tag that data that allows for the next steps toe happen. So now I can write business rules not in terms of columns named columns, but I could write him in terms of the tags being able to automate. That is a huge time saver and the fact that we can suggest that as a rule, rather than waiting for a person to come along and say, Oh, wow. Okay, I need this rule. I need this will thes air steps that increased that are, I should say, decrease that time to value that A. J talked about and then, lastly, a couple of machine learning because even with even with great automation and being able to profile all of your data and getting a good understanding, that brings you to a certain point. But there's still ambiguities in the data. So, for example, I might have to columns date one and date to. I may have even observed the date. One should be less than day two, but I don't really know what date one and date to our other than a date. So this is where it comes in, and I might ask the user said, >>Can >>you help me identify what date? One and date You are in this in this table. Turns out they're a start date and an end date for alone That gets remembered, cycled into the machine learning. So if I start to see this pattern of date one day to elsewhere, I'm going to say, Is it start dating and date? And these Bringing all these things together with this all this automation is really what's key to enabling this This'll data governance. Yeah, >>great. Thanks. Lester and a j wanna wrap things up with something that you mentioned in the beginning about what you guys were doing with Oracle. Take us out by telling us what you're doing there. How are you guys working together? >>Yeah, I think those of us who worked in i t for many years we've We've learned Thio trust articles technology that they're shifting now to ah, hybrid on Prohm Cloud Generation to platform, which is exciting. Andi on their existing customers and new customers moving to article on a journey. So? So Oracle came to us and said, you know, we can see how quickly you're able to help us change mindsets Ondas mindsets are locked in a way of thinking around operating models of I t. That there may be no agile and what siloed on day wanting to break free of that and adopt a more agile A p I at driven approach. A lot of the work that we're doing with our recall no is around, uh, accelerating what customers conduce with understanding their data and to build digital APS by identifying the the underlying data that has value. Onda at the time were able to do that in in in hours, days and weeks. Rather many months. Is opening up the eyes to Chief Data Officers CEO to say, Well, maybe we can do this whole digital transformation this year. Maybe we can bring that forward and and transform who we are as a company on that's driving innovation, which we're excited about it. I know Oracle, a keen Thio to drive through and >>helping businesses transformed digitally is so incredibly important in this time as we look Thio things changing in 2021 a. J. Lester thank you so much for joining me on this segment explaining adaptive data governance, how organizations can use it benefit from it and achieve our Oi. Thanks so much, guys. >>Thank you. Thanks again, Lisa. >>In a moment, we'll look a adaptive data governance in banking. This is the Cube, your global leader in high tech coverage. >>Innovation, impact influence. Welcome to the Cube. Disruptors. Developers and practitioners learn from the voices of leaders who share their personal insights from the hottest digital events around the globe. Enjoy the best this community has to offer on the Cube, your global leader in high tech digital coverage. >>Our next segment here is an interesting panel you're gonna hear from three gentlemen about adaptive data. Governments want to talk a lot about that. Please welcome Yusuf Khan, the global director of data services for Iot Tahoe. We also have Santiago Castor, the chief data officer at the First Bank of Nigeria, and good John Vander Wal, Oracle's senior manager of digital transformation and industries. Gentlemen, it's great to have you joining us in this in this panel. Great >>to be >>tried for me. >>Alright, Santiago, we're going to start with you. Can you talk to the audience a little bit about the first Bank of Nigeria and its scale? This is beyond Nigeria. Talk to us about that. >>Yes, eso First Bank of Nigeria was created 125 years ago. One of the oldest ignored the old in Africa because of the history he grew everywhere in the region on beyond the region. I am calling based in London, where it's kind of the headquarters and it really promotes trade, finance, institutional banking, corporate banking, private banking around the world in particular, in relationship to Africa. We are also in Asia in in the Middle East. >>So, Sanjay, go talk to me about what adaptive data governance means to you. And how does it help the first Bank of Nigeria to be able to innovate faster with the data that you have? >>Yes, I like that concept off adaptive data governor, because it's kind of Ah, I would say an approach that can really happen today with the new technologies before it was much more difficult to implement. So just to give you a little bit of context, I I used to work in consulting for 16, 17 years before joining the president of Nigeria, and I saw many organizations trying to apply different type of approaches in the governance on by the beginning early days was really kind of a year. A Chicago A. A top down approach where data governance was seeing as implement a set of rules, policies and procedures. But really, from the top down on is important. It's important to have the battle off your sea level of your of your director. Whatever I saw, just the way it fails, you really need to have a complimentary approach. You can say bottom are actually as a CEO are really trying to decentralize the governor's. Really, Instead of imposing a framework that some people in the business don't understand or don't care about it, it really needs to come from them. So what I'm trying to say is that data basically support business objectives on what you need to do is every business area needs information on the detector decisions toe actually be able to be more efficient or create value etcetera. Now, depending on the business questions they have to solve, they will need certain data set. So they need actually to be ableto have data quality for their own. For us now, when they understand that they become the stores naturally on their own data sets. And that is where my bottom line is meeting my top down. You can guide them from the top, but they need themselves to be also empower and be actually, in a way flexible to adapt the different questions that they have in orderto be able to respond to the business needs. Now I cannot impose at the finish for everyone. I need them to adapt and to bring their answers toe their own business questions. That is adaptive data governor and all That is possible because we have. And I was saying at the very beginning just to finalize the point, we have new technologies that allow you to do this method data classifications, uh, in a very sophisticated way that you can actually create analitico of your metadata. You can understand your different data sources in order to be able to create those classifications like nationalities, a way of classifying your customers, your products, etcetera. >>So one of the things that you just said Santa kind of struck me to enable the users to be adaptive. They probably don't want to be logging in support ticket. So how do you support that sort of self service to meet the demand of the users so that they can be adaptive. >>More and more business users wants autonomy, and they want to basically be ableto grab the data and answer their own question. Now when you have, that is great, because then you have demand of businesses asking for data. They're asking for the insight. Eso How do you actually support that? I would say there is a changing culture that is happening more and more. I would say even the current pandemic has helped a lot into that because you have had, in a way, off course, technology is one of the biggest winners without technology. We couldn't have been working remotely without these technologies where people can actually looking from their homes and still have a market data marketplaces where they self serve their their information. But even beyond that data is a big winner. Data because the pandemic has shown us that crisis happened, that we cannot predict everything and that we are actually facing a new kind of situation out of our comfort zone, where we need to explore that we need to adapt and we need to be flexible. How do we do that with data. Every single company either saw the revenue going down or the revenue going very up For those companies that are very digital already. Now it changed the reality, so they needed to adapt. But for that they needed information. In order to think on innovate, try toe, create responses So that type of, uh, self service off data Haider for data in order to be able to understand what's happening when the prospect is changing is something that is becoming more, uh, the topic today because off the condemning because of the new abilities, the technologies that allow that and then you then are allowed to basically help your data. Citizens that call them in the organization people that no other business and can actually start playing and an answer their own questions. Eso so these technologies that gives more accessibility to the data that is some cataloging so they can understand where to go or what to find lineage and relationships. All this is is basically the new type of platforms and tools that allow you to create what are called a data marketplace. I think these new tools are really strong because they are now allowing for people that are not technology or I t people to be able to play with data because it comes in the digital world There. Used to a given example without your who You have a very interesting search functionality. Where if you want to find your data you want to sell, Sir, you go there in that search and you actually go on book for your data. Everybody knows how to search in Google, everybody's searching Internet. So this is part of the data culture, the digital culture. They know how to use those schools. Now, similarly, that data marketplace is, uh, in you can, for example, see which data sources they're mostly used >>and enabling that speed that we're all demanding today during these unprecedented times. Goodwin, I wanted to go to you as we talk about in the spirit of evolution, technology is changing. Talk to us a little bit about Oracle Digital. What are you guys doing there? >>Yeah, Thank you. Um, well, Oracle Digital is a business unit that Oracle EMEA on. We focus on emerging countries as well as low and enterprises in the mid market, in more developed countries and four years ago. This started with the idea to engage digital with our customers. Fear Central helps across EMEA. That means engaging with video, having conference calls, having a wall, a green wall where we stand in front and engage with our customers. No one at that time could have foreseen how this is the situation today, and this helps us to engage with our customers in the way we were already doing and then about my team. The focus of my team is to have early stage conversations with our with our customers on digital transformation and innovation. And we also have a team off industry experts who engaged with our customers and share expertise across EMEA, and we inspire our customers. The outcome of these conversations for Oracle is a deep understanding of our customer needs, which is very important so we can help the customer and for the customer means that we will help them with our technology and our resource is to achieve their goals. >>It's all about outcomes, right? Good Ron. So in terms of automation, what are some of the things Oracle's doing there to help your clients leverage automation to improve agility? So that they can innovate faster, which in these interesting times it's demanded. >>Yeah, thank you. Well, traditionally, Oracle is known for their databases, which have bean innovated year over year. So here's the first lunch on the latest innovation is the autonomous database and autonomous data warehouse. For our customers, this means a reduction in operational costs by 90% with a multi medal converts, database and machine learning based automation for full life cycle management. Our databases self driving. This means we automate database provisioning, tuning and scaling. The database is self securing. This means ultimate data protection and security, and it's self repairing the automates failure, detection fail over and repair. And then the question is for our customers, What does it mean? It means they can focus on their on their business instead off maintaining their infrastructure and their operations. >>That's absolutely critical use if I want to go over to you now. Some of the things that we've talked about, just the massive progression and technology, the evolution of that. But we know that whether we're talking about beta management or digital transformation, a one size fits all approach doesn't work to address the challenges that the business has, um that the i t folks have, as you're looking through the industry with what Santiago told us about first Bank of Nigeria. What are some of the changes that you're seeing that I owe Tahoe seeing throughout the industry? >>Uh, well, Lisa, I think the first way I'd characterize it is to say, the traditional kind of top down approach to data where you have almost a data Policeman who tells you what you can and can't do, just doesn't work anymore. It's too slow. It's too resource intensive. Uh, data management data, governments, digital transformation itself. It has to be collaborative on. There has to be in a personalization to data users. Um, in the environment we find ourselves in. Now, it has to be about enabling self service as well. Um, a one size fits all model when it comes to those things around. Data doesn't work. As Santiago was saying, it needs to be adapted toe how the data is used. Andi, who is using it on in order to do this cos enterprises organizations really need to know their data. They need to understand what data they hold, where it is on what the sensitivity of it is they can then any more agile way apply appropriate controls on access so that people themselves are and groups within businesses are our job and could innovate. Otherwise, everything grinds to a halt, and you risk falling behind your competitors. >>Yeah, that one size fits all term just doesn't apply when you're talking about adaptive and agility. So we heard from Santiago about some of the impact that they're making with First Bank of Nigeria. Used to talk to us about some of the business outcomes that you're seeing other customers make leveraging automation that they could not do >>before it's it's automatically being able to classify terabytes, terabytes of data or even petabytes of data across different sources to find duplicates, which you can then re mediate on. Deletes now, with the capabilities that iota offers on the Oracle offers, you can do things not just where the five times or 10 times improvement, but it actually enables you to do projects for Stop that otherwise would fail or you would just not be able to dio I mean, uh, classifying multi terrible and multi petabytes states across different sources, formats very large volumes of data in many scenarios. You just can't do that manually. I mean, we've worked with government departments on the issues there is expect are the result of fragmented data. There's a lot of different sources. There's lot of different formats and without these newer technologies to address it with automation on machine learning, the project isn't durable. But now it is on that that could lead to a revolution in some of these businesses organizations >>to enable that revolution that there's got to be the right cultural mindset. And one of the when Santiago was talking about folks really kind of adapted that. The thing I always call that getting comfortably uncomfortable. But that's hard for organizations to. The technology is here to enable that. But well, you're talking with customers use. How do you help them build the trust in the confidence that the new technologies and a new approaches can deliver what they need? How do you help drive the kind of a tech in the culture? >>It's really good question is because it can be quite scary. I think the first thing we'd start with is to say, Look, the technology is here with businesses like I Tahoe. Unlike Oracle, it's already arrived. What you need to be comfortable doing is experimenting being agile around it, Andi trying new ways of doing things. Uh, if you don't wanna get less behind that Santiago on the team that fbn are a great example off embracing it, testing it on a small scale on, then scaling up a Toyota, we offer what we call a data health check, which can actually be done very quickly in a matter of a few weeks. So we'll work with a customer. Picky use case, install the application, uh, analyzed data. Drive out Cem Cem quick winds. So we worked in the last few weeks of a large entity energy supplier, and in about 20 days, we were able to give them an accurate understanding of their critical data. Elements apply. Helping apply data protection policies. Minimize copies of the data on work out what data they needed to delete to reduce their infrastructure. Spend eso. It's about experimenting on that small scale, being agile on, then scaling up in a kind of very modern way. >>Great advice. Uh, Santiago, I'd like to go back to Is we kind of look at again that that topic of culture and the need to get that mindset there to facilitate these rapid changes, I want to understand kind of last question for you about how you're doing that from a digital transformation perspective. We know everything is accelerating in 2020. So how are you building resilience into your data architecture and also driving that cultural change that can help everyone in this shift to remote working and a lot of the the digital challenges and changes that we're all going through? >>The new technologies allowed us to discover the dating anyway. Toe flawed and see very quickly Information toe. Have new models off over in the data on giving autonomy to our different data units. Now, from that autonomy, they can then compose an innovator own ways. So for me now, we're talking about resilience because in a way, autonomy and flexibility in a organization in a data structure with platform gives you resilience. The organizations and the business units that I have experienced in the pandemic are working well. Are those that actually because they're not physically present during more in the office, you need to give them their autonomy and let them actually engaged on their own side that do their own job and trust them in a way on as you give them, that they start innovating and they start having a really interesting ideas. So autonomy and flexibility. I think this is a key component off the new infrastructure. But even the new reality that on then it show us that, yes, we used to be very kind off structure, policies, procedures as very important. But now we learn flexibility and adaptability of the same side. Now, when you have that a key, other components of resiliency speed, because people want, you know, to access the data and access it fast and on the site fast, especially changes are changing so quickly nowadays that you need to be ableto do you know, interact. Reiterate with your information to answer your questions. Pretty, um, so technology that allows you toe be flexible iterating on in a very fast job way continue will allow you toe actually be resilient in that way, because you are flexible, you adapt your job and you continue answering questions as they come without having everything, setting a structure that is too hard. We also are a partner off Oracle and Oracle. Embodies is great. They have embedded within the transactional system many algorithms that are allowing us to calculate as the transactions happened. What happened there is that when our customers engaged with algorithms and again without your powers, well, the machine learning that is there for for speeding the automation of how you find your data allows you to create a new alliance with the machine. The machine is their toe, actually, in a way to your best friend to actually have more volume of data calculated faster. In a way, it's cover more variety. I mean, we couldn't hope without being connected to this algorithm on >>that engagement is absolutely critical. Santiago. Thank you for sharing that. I do wanna rap really quickly. Good On one last question for you, Santiago talked about Oracle. You've talked about a little bit. As we look at digital resilience, talk to us a little bit in the last minute about the evolution of Oracle. What you guys were doing there to help your customers get the resilience that they have toe have to be not just survive but thrive. >>Yeah. Oracle has a cloud offering for infrastructure, database, platform service and a complete solutions offered a South on Daz. As Santiago also mentioned, We are using AI across our entire portfolio and by this will help our customers to focus on their business innovation and capitalize on data by enabling new business models. Um, and Oracle has a global conference with our cloud regions. It's massively investing and innovating and expanding their clouds. And by offering clouds as public cloud in our data centers and also as private cloud with clouded customer, we can meet every sovereignty and security requirements. And in this way we help people to see data in new ways. We discover insights and unlock endless possibilities. And and maybe 11 of my takeaways is if I If I speak with customers, I always tell them you better start collecting your data. Now we enable this partners like Iota help us as well. If you collect your data now, you are ready for tomorrow. You can never collect your data backwards, So that is my take away for today. >>You can't collect your data backwards. Excellently, John. Gentlemen, thank you for sharing all of your insights. Very informative conversation in a moment, we'll address the question. Do you know your data? >>Are you interested in test driving the iota Ho platform kick Start the benefits of data automation for your business through the Iota Ho Data Health check program. Ah, flexible, scalable sandbox environment on the cloud of your choice with set up service and support provided by Iota ho. Look time with a data engineer to learn more and see Io Tahoe in action from around the globe. It's the Cube presenting adaptive data governance brought to you by Iota Ho. >>In this next segment, we're gonna be talking to you about getting to know your data. And specifically you're gonna hear from two folks at Io Tahoe. We've got enterprise account execs to be to Davis here, as well as Enterprise Data engineer Patrick Simon. They're gonna be sharing insights and tips and tricks for how you could get to know your data and quickly on. We also want to encourage you to engage with the media and Patrick, use the chat feature to the right, send comments, questions or feedback so you can participate. All right, Patrick Savita, take it away. Alright. >>Thankfully saw great to be here as Lisa mentioned guys, I'm the enterprise account executive here in Ohio. Tahoe you Pat? >>Yeah. Hey, everyone so great to be here. I said my name is Patrick Samit. I'm the enterprise data engineer here in Ohio Tahoe. And we're so excited to be here and talk about this topic as one thing we're really trying to perpetuate is that data is everyone's business. >>So, guys, what patent I got? I've actually had multiple discussions with clients from different organizations with different roles. So we spoke with both your technical and your non technical audience. So while they were interested in different aspects of our platform, we found that what they had in common was they wanted to make data easy to understand and usable. So that comes back. The pats point off to being everybody's business because no matter your role, we're all dependent on data. So what Pan I wanted to do today was wanted to walk you guys through some of those client questions, slash pain points that we're hearing from different industries and different rules and demo how our platform here, like Tahoe, is used for automating Dozier related tasks. So with that said are you ready for the first one, Pat? >>Yeah, Let's do it. >>Great. So I'm gonna put my technical hat on for this one. So I'm a data practitioner. I just started my job. ABC Bank. I have, like, over 100 different data sources. So I have data kept in Data Lakes, legacy data, sources, even the cloud. So my issue is I don't know what those data sources hold. I don't know what data sensitive, and I don't even understand how that data is connected. So how can I saw who help? >>Yeah, I think that's a very common experience many are facing and definitely something I've encountered in my past. Typically, the first step is to catalog the data and then start mapping the relationships between your various data stores. Now, more often than not, this has tackled through numerous meetings and a combination of excel and something similar to video which are too great tools in their own part. But they're very difficult to maintain. Just due to the rate that we are creating data in the modern world. It starts to beg for an idea that can scale with your business needs. And this is where a platform like Io Tahoe becomes so appealing, you can see here visualization of the data relationships created by the I. O. Tahoe service. Now, what is fantastic about this is it's not only laid out in a very human and digestible format in the same action of creating this view, the data catalog was constructed. >>Um so is the data catalog automatically populated? Correct. Okay, so So what I'm using Iota hope at what I'm getting is this complete, unified automated platform without the added cost? Of course. >>Exactly. And that's at the heart of Iota Ho. A great feature with that data catalog is that Iota Ho will also profile your data as it creates the catalog, assigning some meaning to those pesky column underscore ones and custom variable underscore tents. They're always such a joy to deal with. Now, by leveraging this interface, we can start to answer the first part of your question and understand where the core relationships within our data exists. Uh, personally, I'm a big fan of this view, as it really just helps the i b naturally John to these focal points that coincide with these key columns following that train of thought, Let's examine the customer I D column that seems to be at the center of a lot of these relationships. We can see that it's a fairly important column as it's maintaining the relationship between at least three other tables. >>Now you >>notice all the connectors are in this blue color. This means that their system defined relationships. But I hope Tahoe goes that extra mile and actually creates thes orange colored connectors as well. These air ones that are machine learning algorithms have predicted to be relationships on. You can leverage to try and make new and powerful relationships within your data. >>Eso So this is really cool, and I can see how this could be leverage quickly now. What if I added new data sources or your multiple data sources and need toe identify what data sensitive can iota who detect that? >>Yeah, definitely. Within the hotel platform. There, already over 300 pre defined policies such as hip for C, C, P. A and the like one can choose which of these policies to run against their data along for flexibility and efficiency and running the policies that affect organization. >>Okay, so so 300 is an exceptional number. I'll give you that. But what about internal policies that apply to my organization? Is there any ability for me to write custom policies? >>Yeah, that's no issue. And it's something that clients leverage fairly often to utilize this function when simply has to write a rejects that our team has helped many deploy. After that, the custom policy is stored for future use to profile sensitive data. One then selects the data sources they're interested in and select the policies that meet your particular needs. The interface will automatically take your data according to the policies of detects, after which you can review the discoveries confirming or rejecting the tagging. All of these insights are easily exported through the interface. Someone can work these into the action items within your project management systems, and I think this lends to the collaboration as a team can work through the discovery simultaneously, and as each item is confirmed or rejected, they can see it ni instantaneously. All this translates to a confidence that with iota hope, you can be sure you're in compliance. >>So I'm glad you mentioned compliance because that's extremely important to my organization. So what you're saying when I use the eye a Tahoe automated platform, we'd be 90% more compliant that before were other than if you were going to be using a human. >>Yeah, definitely the collaboration and documentation that the Iot Tahoe interface lends itself to really help you build that confidence that your compliance is sound. >>So we're planning a migration. Andi, I have a set of reports I need to migrate. But what I need to know is, uh well, what what data sources? Those report those reports are dependent on. And what's feeding those tables? >>Yeah, it's a fantastic questions to be toe identifying critical data elements, and the interdependencies within the various databases could be a time consuming but vital process and the migration initiative. Luckily, Iota Ho does have an answer, and again, it's presented in a very visual format. >>Eso So what I'm looking at here is my entire day landscape. >>Yes, exactly. >>Let's say I add another data source. I can still see that unified 3 60 view. >>Yeah, One future that is particularly helpful is the ability to add data sources after the data lineage. Discovery has finished alone for the flexibility and scope necessary for any data migration project. If you only need need to select a few databases or your entirety, this service will provide the answers. You're looking for things. Visual representation of the connectivity makes the identification of critical data elements a simple matter. The connections air driven by both system defined flows as well as those predicted by our algorithms, the confidence of which, uh, can actually be customized to make sure that they're meeting the needs of the initiative that you have in place. This also provides tabular output in case you needed for your own internal documentation or for your action items, which we can see right here. Uh, in this interface, you can actually also confirm or deny the pair rejection the pair directions, allowing to make sure that the data is as accurate as possible. Does that help with your data lineage needs? >>Definitely. So So, Pat, My next big question here is So now I know a little bit about my data. How do I know I can trust >>it? So >>what I'm interested in knowing, really is is it in a fit state for me to use it? Is it accurate? Does it conform to the right format? >>Yeah, that's a great question. And I think that is a pain point felt across the board, be it by data practitioners or data consumers alike. Another service that I owe Tahoe provides is the ability to write custom data quality rules and understand how well the data pertains to these rules. This dashboard gives a unified view of the strength of these rules, and your dad is overall quality. >>Okay, so Pat s o on on the accuracy scores there. So if my marketing team needs to run, a campaign can read dependent those accuracy scores to know what what tables have quality data to use for our marketing campaign. >>Yeah, this view would allow you to understand your overall accuracy as well as dive into the minutia to see which data elements are of the highest quality. So for that marketing campaign, if you need everything in a strong form, you'll be able to see very quickly with these high level numbers. But if you're only dependent on a few columns to get that information out the door, you can find that within this view, eso >>you >>no longer have to rely on reports about reports, but instead just come to this one platform to help drive conversations between stakeholders and data practitioners. >>So I get now the value of IATA who brings by automatically capturing all those technical metadata from sources. But how do we match that with the business glossary? >>Yeah, within the same data quality service that we just reviewed, one can actually add business rules detailing the definitions and the business domains that these fall into. What's more is that the data quality rules were just looking at can then be tied into these definitions. Allowing insight into the strength of these business rules is this service that empowers stakeholders across the business to be involved with the data life cycle and take ownership over the rules that fall within their domain. >>Okay, >>so those custom rules can I apply that across data sources? >>Yeah, you could bring in as many data sources as you need, so long as you could tie them to that unified definition. >>Okay, great. Thanks so much bad. And we just want to quickly say to everyone working in data, we understand your pain, so please feel free to reach out to us. we are Website the chapel. Oh, Arlington. And let's get a conversation started on how iota Who can help you guys automate all those manual task to help save you time and money. Thank you. Thank >>you. Your Honor, >>if I could ask you one quick question, how do you advise customers? You just walk in this great example this banking example that you instantly to talk through. How do you advise customers get started? >>Yeah, I think the number one thing that customers could do to get started with our platform is to just run the tag discovery and build up that data catalog. It lends itself very quickly to the other needs you might have, such as thes quality rules. A swell is identifying those kind of tricky columns that might exist in your data. Those custom variable underscore tens I mentioned before >>last questions to be to anything to add to what Pat just described as a starting place. >>I'm no, I think actually passed something that pretty well, I mean, just just by automating all those manual task. I mean, it definitely can save your company a lot of time and money, so we we encourage you just reach out to us. Let's get that conversation >>started. Excellent. So, Pete and Pat, thank you so much. We hope you have learned a lot from these folks about how to get to know your data. Make sure that it's quality, something you can maximize the value of it. Thanks >>for watching. Thanks again, Lisa, for that very insightful and useful deep dive into the world of adaptive data governance with Iota Ho Oracle First Bank of Nigeria This is Dave a lot You won't wanna mess Iota, whose fifth episode in the data automation Siri's in that we'll talk to experts from Red Hat and Happiest Minds about their best practices for managing data across hybrid cloud Inter Cloud multi Cloud I T environment So market calendar for Wednesday, January 27th That's Episode five. You're watching the Cube Global Leader digital event technique
SUMMARY :
adaptive data governance brought to you by Iota Ho. Gentlemen, it's great to have you on the program. Lisa is good to be back. Great. Listen, we're gonna start with you. But to really try to address these customer concerns because, you know, we wanna we So it's exciting a J from the CEO's level. It's real satisfying to see how we're able. Let's let's go back over to you. But they need to understand what kind of data they have, what shape it's in what's dependent lot of a lot of frameworks these days are hardwired, so you can set up a set It's the technical metadata coming together with policies Is this book enterprise companies are doing now? help the organizations to digest their data is to And if it was me eating that food with you guys, I would be not using chopsticks. So if you look at the challenges for these data professionals, you know, they're either on a journey to the cloud. Well, as she digs into the databases, she starts to see that So a J talk us through some examples of where But I think it helped do this Bring it to life a little bit. And one of the things I was thinking when you were talking through some We can see that on the the graphic that we've just How are you seeing those technologies being think you know this But the very first step is understanding what you have in normalizing that So if I start to see this pattern of date one day to elsewhere, I'm going to say, in the beginning about what you guys were doing with Oracle. So Oracle came to us and said, you know, we can see things changing in 2021 a. J. Lester thank you so much for joining me on this segment Thank you. is the Cube, your global leader in high tech coverage. Enjoy the best this community has to offer on the Cube, Gentlemen, it's great to have you joining us in this in this panel. Can you talk to the audience a little bit about the first Bank of One of the oldest ignored the old in Africa because of the history And how does it help the first Bank of Nigeria to be able to innovate faster with the point, we have new technologies that allow you to do this method data So one of the things that you just said Santa kind of struck me to enable the users to be adaptive. Now it changed the reality, so they needed to adapt. I wanted to go to you as we talk about in the spirit of evolution, technology is changing. customer and for the customer means that we will help them with our technology and our resource is to achieve doing there to help your clients leverage automation to improve agility? So here's the first lunch on the latest innovation Some of the things that we've talked about, Otherwise, everything grinds to a halt, and you risk falling behind your competitors. Used to talk to us about some of the business outcomes that you're seeing other customers make leveraging automation different sources to find duplicates, which you can then re And one of the when Santiago was talking about folks really kind of adapted that. Minimize copies of the data can help everyone in this shift to remote working and a lot of the the and on the site fast, especially changes are changing so quickly nowadays that you need to be What you guys were doing there to help your customers I always tell them you better start collecting your data. Gentlemen, thank you for sharing all of your insights. adaptive data governance brought to you by Iota Ho. In this next segment, we're gonna be talking to you about getting to know your data. Thankfully saw great to be here as Lisa mentioned guys, I'm the enterprise account executive here in Ohio. I'm the enterprise data engineer here in Ohio Tahoe. So with that said are you ready for the first one, Pat? So I have data kept in Data Lakes, legacy data, sources, even the cloud. Typically, the first step is to catalog the data and then start mapping the relationships Um so is the data catalog automatically populated? i b naturally John to these focal points that coincide with these key columns following These air ones that are machine learning algorithms have predicted to be relationships Eso So this is really cool, and I can see how this could be leverage quickly now. such as hip for C, C, P. A and the like one can choose which of these policies policies that apply to my organization? And it's something that clients leverage fairly often to utilize this So I'm glad you mentioned compliance because that's extremely important to my organization. interface lends itself to really help you build that confidence that your compliance is Andi, I have a set of reports I need to migrate. Yeah, it's a fantastic questions to be toe identifying critical data elements, I can still see that unified 3 60 view. Yeah, One future that is particularly helpful is the ability to add data sources after So now I know a little bit about my data. the data pertains to these rules. So if my marketing team needs to run, a campaign can read dependent those accuracy scores to know what the minutia to see which data elements are of the highest quality. no longer have to rely on reports about reports, but instead just come to this one So I get now the value of IATA who brings by automatically capturing all those technical to be involved with the data life cycle and take ownership over the rules that fall within their domain. Yeah, you could bring in as many data sources as you need, so long as you could manual task to help save you time and money. you. this banking example that you instantly to talk through. Yeah, I think the number one thing that customers could do to get started with our so we we encourage you just reach out to us. folks about how to get to know your data. into the world of adaptive data governance with Iota Ho Oracle First Bank of Nigeria
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Amanda | PERSON | 0.99+ |
Jason | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
Patrick Simon | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Santiago | PERSON | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
Yusuf Khan | PERSON | 0.99+ |
Asia | LOCATION | 0.99+ |
16 | QUANTITY | 0.99+ |
Santiago Castor | PERSON | 0.99+ |
Ohio | LOCATION | 0.99+ |
London | LOCATION | 0.99+ |
ABC Bank | ORGANIZATION | 0.99+ |
Patrick Savita | PERSON | 0.99+ |
10 times | QUANTITY | 0.99+ |
Sanjay | PERSON | 0.99+ |
Angie | PERSON | 0.99+ |
Wednesday, January 27th | DATE | 0.99+ |
Africa | LOCATION | 0.99+ |
Thio | PERSON | 0.99+ |
John Vander Wal | PERSON | 0.99+ |
2020 | DATE | 0.99+ |
Patrick | PERSON | 0.99+ |
two columns | QUANTITY | 0.99+ |
90% | QUANTITY | 0.99+ |
Siri | TITLE | 0.99+ |
Toyota | ORGANIZATION | 0.99+ |
Bio Tahoe | ORGANIZATION | 0.99+ |
Azaz | PERSON | 0.99+ |
Pat | PERSON | 0.99+ |
11 | QUANTITY | 0.99+ |
five times | QUANTITY | 0.99+ |
Oracle Digital | ORGANIZATION | 0.99+ |
J. Bihar | PERSON | 0.99+ |
1% | QUANTITY | 0.99+ |
Staley | PERSON | 0.99+ |
Iot Tahoe | ORGANIZATION | 0.99+ |
Iota ho | ORGANIZATION | 0.99+ |
today | DATE | 0.99+ |
Ron | PERSON | 0.99+ |
first | QUANTITY | 0.99+ |
10 | QUANTITY | 0.99+ |
Iota Ho | ORGANIZATION | 0.99+ |
Andi | PERSON | 0.99+ |
Io Tahoe | ORGANIZATION | 0.99+ |
one date | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
excel | TITLE | 0.99+ |
tomorrow | DATE | 0.99+ |
3% | QUANTITY | 0.99+ |
John | PERSON | 0.99+ |
First Bank of Nigeria | ORGANIZATION | 0.99+ |
Middle East | LOCATION | 0.99+ |
Patrick Samit | PERSON | 0.99+ |
I. O. Tahoe | ORGANIZATION | 0.99+ |
first step | QUANTITY | 0.99+ |
97% | QUANTITY | 0.99+ |
Lester | PERSON | 0.99+ |
two folks | QUANTITY | 0.99+ |
Dave | PERSON | 0.99+ |
2021 | DATE | 0.99+ |
fifth episode | QUANTITY | 0.99+ |
one grain | QUANTITY | 0.99+ |
Ajay Vohora, Io-Tahoe | SmartData Marketplaces
>> Narrator: From around the globe, it's theCUBE. With digital coverage of smart data marketplaces. Brought to you by Io-Tahoe. >> Digital transformation has really gone from a buzzword to a mandate, but digital business is a data business. And for the last several months we've been working with Io-Tahoe on an ongoing content series, focused on smart data and automation to drive better insights and outcomes, essentially putting data to work. And today we're going to do a deeper dive on automating data discovery. And one of the thought leaders in this space is Ajay Vohora, who's the CEO of Io-Tahoe. Once again, joining me, Ajay good to see you. Thanks for coming on. >> Great to be here, David, thank you. >> So let's, let's start by talking about some of the business realities and what are the economics that are driving automated data discovery? Why is that so important? >> Yeah, on this one, David it's a number of competing factors. We've got the reality of data which may be sensitive. So there's control. Three other elements wanting to drive value from that data to innovation. You can't really drive a lot of value without exchanging data. So the ability to exchange data and to manage those cost overheads and data discovery is at the root of managing that in an automated way to classify that data and set some policies to put that automation in place. >> Yeah, look, we have a picture of this. If we could bring it up guys, cause I want to, Ajay, help the audience understand kind of where data discovery fits in here. This is, as we talked about, this is a complicated situation for a lot of customers. They've got variety of different tools and you've really laid it out nicely here in this diagram. So, take us through sort of where that piece fits. >> Yeah, I mean, we're at the right hand side of this exchange, you know. We're really now in a data driven economy that is everything's connected through APIs that we consume online through mobile apps. And what's not apparent is the chain of activities and tasks that have to go into serving that data to an API at the outset. They may be many legacy systems, technologies, platforms On-premise, in cloud, hybrid, you name it and across those silos, getting to a unified view is the heavy lifting. I think we've seen some, some great impacts that BI tools, such as Power BI, Tableau, Looker, and so on, and Qlik have had, and they're in our ecosystem on visualizing Data and, you know, CEOs, managers, people that are working in companies day-to-day get a lot of value from saying, "What's the real time activity? "What was the trend over this month versus last month?" The tools to enable that, you know, we hear a lot of good things that we're doing with Snowflake, MongoDB on the public Cloud platforms, GCP Azure about enabling building those pipelines to feed into those analytics. But what often gets hidden is how do you source that data that could be locked into a mainframe, a data warehouse, IOT data, and pull over all of that together. And that is the reality of that is it's a lot of heavy lifting. It's hands on work that can be time consuming. And the issue there is that data may have value. It might have potential to have an impact on the top line for a business, on outcomes for consumers, but you're never really sure unless you've done the investigation, discovered it, unified that, and be able to serve that through to other technologies. >> Guys, if you would bring that picture back up again, because Ajay you made a point and I want to land on that for a second. There's a lot of manual curating. An example would be the data catalog. You know, data scientists complain all the time that they're manually wrangling data. And so you're trying to inject automation into the cycle. And then the other piece that I want you to address is the importance of APIs. You really can't do this without an architecture that allows you to connect things together that sort of enables some of the automation. >> Yep, I mean, I'll take that in two parts, David, the APIs, so virtual machines connected by APIs, business rules, and business logic driven by APIs, applications, so everything across the stack from infrastructure down to the network, hardware is all connected through APIs and the work of serving data through to an API, building those pipelines, is often miscalculated, just how much manual effort that takes and that manual effort, we've got a nice list here of what we automate down at the bottom, those tasks of indexing, labeling, mapping across different legacy systems, all of that takes away from the job of a data scientist or data engineer, looking to produce value, monetize data, and to help that business convey to consumers. >> Yeah, it's that top layer that the business sees, of course, there's a lot of work that has to go into achieving that. I want to talk about some of the key tech trends that you're seeing. And one of the things that we talk about a lot is metadata. The importance of metadata, you know, can't be understated. What are some of the big trends that you're seeing metadata and others? >> Yeah, I'll summarize it as five. There's a trend now look at metadata more holistically across the enterprise. And that really makes sense from trying to look across different data silos and apply a policy to manage that data. So that's the control piece. That's that lever. The other side, sometimes competing with that control around sensitive data around managing the cost of data is innovation. Innovation being able to speculate and experiment and try things out where you don't really know what the outcome is if you're a data scientist and engineer, you've got a hypothesis and therefore you've got that tension between control over data and innovation and driving value from it. So enterprise wide metadata management is really helping to unlock where might that latent value be across that sets of data. The other piece is adaptive data governance. Those controls that stick from the data policemen, data stewards, where they're trying to protect the organization, protect the brand, protect consumers data necessary, but in different use cases, you might want to nuance and apply a different policy to govern that data relevant to the context where you might have data that is less sensitive, that can be used for innovation and adapting the style of governance to fit the context is another trend that we're seeing coming up here. A few others is where we're sitting quite extensively in working with automating data discovery. We're now breaking that down into what can we direct? What do we know is a business outcome is a known upfront objective and direct that data discovery to towards that. And that means applying our algorithms around technology and our tools towards solving a known problem. The other one is autonomous data discovery. And that means, you know, trying to allow background processes to understand what changes are happening with data over time, flagging those anomalies. And the reason that's important is when you look over a length of time to see different spikes, different trends and activity, that's really giving a data ops team the ability to manage and calibrate how they're applying policies and controls the data. And the last two, David, that we're seeing is this huge drive towards self-service. So re-imagining how to apply policy data governance into the hands of a data consumer inside a business, or indeed the consumer themselves, to self-service if they're a banking customer or healthcare customer and the policies and the controls and rules, making sure that those are all in place to adaptively serve those data marketplaces that when are involved in creating. >> I want to ask you about the autonomous data discovering, the adaptive data governance, is the problem we're addressing there one of quality, in other words, machines are better than humans are at doing this? Is it one of scale? That humans just don't don't scale that well? Is it both? Can you add some color to that? >> Yeah, honestly, it's the same equation that existed 10 years ago, 20 years ago, it's being exacerbated, but it's that equation of how do I control all the things that I need to protect? How do I enable innovation where it is going to deliver business value? How do I exchange data between a customer, somebody in my supply chain safely, and do all of that whilst managing the fourth leg, which is cost overheads. There's not an open checkbook here. I've got to figure out if I'm the CIO and CDO, how I do all of this within a fixed budget. So those aspects have always been there, now with more choices, infrastructure in the Cloud, API driven applications, On-premises, and that is expanding the choices that a business has and how they put their data to work. It's also then creating a layer of management and data governance that really has to now manage those four aspects, control, innovation, exchange of data, and the cost overhead. >> That top layer of the first slide that we showed was all about the business value. So, I wonder if we could drill into the business impact a little bit. What are your customers seeing specifically in terms of the impact of all this automation on their business? >> Yeah, so we've had some great results. I think a few of the biggest have been helping customers move away from manually curating their data and their metadata. It used to be a time where if data initiatives or data governance initiatives, there'd be teams of people manually feeding a data catalog. And it's great to have that inventory of classified data to be able to understand single version of the truth, but having 10, 15 people manually process that, keep it up to date, when it's moving feet, the reality of it is what's true about data today, add another few sources and a few months time to your business, start collaborating with new partners, suddenly the landscape has changed. The amount of work has gone up, but what we're finding is through automating, creating that data discovery, feeding our data catalog, that's releasing a lot more time for our customers to spend on innovating and managing their data. A couple of others is around self service data analytics, moving the choices of what data might have business value into the hands of business users and data consumers to have faster cycle times around generating insights. And we're really helping them by automating the creation of those data sets that are needed for that. And the last piece, I'd have to say where we're seeing impacts more recently is in the exchange of data. There are a number of marketplaces out there who are now being compelled to become more digital, to rewire their business processes and everything from an RPA initiative to automation involving digital transformation is having CIOs, chief data officers and enterprise architects rethink how do they, how do they rewire the pipelines for their data to feed that digital transformation? >> Yeah, to me, it comes down to monetization. Now, of course, that's for a for-profit industry. For non-profits, for sure, the cost cutting or in the case of healthcare, which we'll talk about in a moment, I mean, it's patient outcomes, but the job of a Chief Data Officer has gone from data quality and governance and compliance to really figuring out how data can be monetized, not necessarily selling the data, but how it contributes to the monetization of the company. And then really understanding specifically for that organization, how to apply that. And that is a big challenge. We sort of chatted about 10 years ago, the early days of a dupe. And then 1% of the companies had enough engineers to figure it out, but now the tooling is available. The technology is there and the practices are there. And that really, to me is the bottom line, Ajay, is it's show me the money. >> Absolutely. It's definitely is focusing in on the single view of that customer and where we're helping there is to pull together those disparate, siloed sources of data to understand what are the needs of the patient, of the broker of the, if it's insurance? What are the needs of the supply chain manager, if it's manufacturing? And providing that 360 view of data is helping to see, helping that individual unlock the value for the business. So data's providing the lens provided, you know which data it is that can assist in doing that. >> And, you know, you mentioned RPA before, I had an RPA customer tell me she was a Six Sigma expert and she told me, "We would never try to apply Six Sigma "to a business process, "but with RPA we can do so very cheaply." Well, what that means is lower costs. It means better employee satisfaction and really importantly, better customer satisfaction and better customer outcomes. Let's talk about healthcare for a minute because it's a really important industry. It's one that is ripe for disruption and has really been, up until recently, pretty slow to adopt a lot of the major technologies that have been made available. But what are you seeing in terms of this theme we're using a putting data to work in healthcare specifically? >> Yeah, I mean, health care's has had a lot thrown at it. There's been a lot of change in terms of legislation recently, particularly in the U.S. market, in other economies, healthcare is on a path to becoming more digital. And part of that is around transparency of price. So, to be operating effectively as a healthcare marketplace, being able to have that price transparency around what an elective procedure is going to cost before taking that step forward. It's super important to have an informed decision around that. So if we look at the U.S., for example, we've seen that healthcare costs annually have risen to $4 trillion, but even with all of that cost, we have healthcare consumers who are reluctant sometimes to take up healthcare even if they have symptoms. And a lot of that is driven through not knowing what they're opening themselves up to. And, you know, I think David, if you or I were to book travel a holiday, maybe, or trip, we'd want to know what we're in for, what we're paying for upfront. But sometimes in healthcare that choice, the option might be the plan, but the cost that comes with it isn't. So recent legislation in the U.S. is certainly helpful to bring forward that price transparency. The underlying issue there though is the disparate different format types of data that are being used from payers, patients, employers, different healthcare departments to try and make that work. And where we're helping on that aspect in particular related to price transparency is to help make that data machine readable. So, sometimes with data, the beneficiary might be a person, but in a lot of cases, now we're seeing the ability to have different systems interact and exchange data in order to process the workflow to generate online lists of pricing from a provider that's been negotiated with a payer is really an enabling factor. >> So guys, I wonder if you could bring up the next slide, which is kind of the nirvana. So, if you saw the previous slide that the middle there was all different shapes and presumably to disparate data, this is the outcome that you want to get, where everything fits together nicely. And you've got this open exchange. It's not opaque as it is today. It's not bubble gum, band-aids and duct tape, but describe this sort of outcome that you're trying to achieve and maybe a little bit about what it's going to take to get there. >> Ajay: Yeah, that that's the culmination of a number of things. It's making sure that the data is machine readable, making it available to APIs, that could be RPA tools. We're working with technology companies that employ RPA for healthcare, and specifically to manage that patient and payer data to bring that together. In our data discovery, what we're able to do is to classify that data and have it made available to a downstream tool technology or person to apply that, that workflow to the data. So this looks like nirvana, it looks like utopia, but it's, you know, the end objective of a journey that we can see in different economies, that are at different stages of maturity in turning healthcare into a digital service even so that you can consume it from where you live, from home with telemedicine and tele care. >> Yeah, so, and this is not just for healthcare, but you know, you want to achieve that self-service data marketplace in virtually any industry. You're working with TCS, Tata Consulting Services to achieve this. You know, a company like Io-Tahoe has to have partnerships with organizations that have deep industry expertise. Talk about your relationship with TCS and what you guys are doing specifically in this regard. >> Yeah, we've been working with TCS now for a long while and we'll be announcing some of those initiatives here where we're now working together to reach their customers where they've got a brilliant framework of business, 4.0, where they're re-imagining with the clients, how their business can operate with AI, with automation and become more agile and digital. Our technology, now, the reams of patients that we have in our portfolio, being able to apply that at scale, on a global scale across industries, such as banking, insurance and healthcare is really allowing us to see a bigger impact on consumer outcomes, patient outcomes. And the feedback from TCS is that we're really helping in those initiatives remove that friction. They talk a lot about data friction. I think that's a polite term for the image that we just saw with the disparate technologies that the legacy that has built up. So if we want to create a transformation, having that partnership with TCS across industries is giving us that reach and that impact on many different people's day-to-day jobs and lives. >> Let's talk a little bit about the Cloud. It's a topic that we've hit on quite a bit here in this content series. But, but you know, the Cloud companies, the big hyper-scalers, they've put everything into the Cloud, right? But customers are more circumspect than that. But at the same time, machine intelligence, ML, AI, the Cloud is a place to do a lot of that. That's where a lot of the innovation occurs. And so what are your thoughts on getting to the Cloud, putting data to work, if you will, with machine learning, stuff that you're doing with AWS, what's your fit there? >> Yeah, we, David, we work with all of the Cloud platforms, Microsoft Azure, GCP, IBM, but we're expanding our partnership now with AWS. And we're really opening up the ability to work with their Greenfield accounts, where a lot of that data, that technology is in their own data centers at the customer. And that's across banking, healthcare, manufacturing, and insurance. And for good reason, a lot of companies that have taken the time to see what works well for them with the technologies that the Cloud providers are offering, and a lot of cases, testing services or analytics using the Cloud to move workloads to the Cloud to drive data analytics is a real game changer. So there's good reason to maintain a lot of systems On-premise. If that makes sense from a cost, from a liability point of view and the number of clients that we work with that do have, and will keep their mainframe systems when in Cobra is no surprise to us, but equally they want to tap into technologies that AWS has such as SageMaker. The issue is as a Chief Data Officer, I didn't have the budget to move everything to the Cloud they want, I might want to show some results first upfront to my business users and work closely with my Chief Marketing Officer to look at what's happening in terms of customer trends and customer behavior> What are the customer outcomes, patient outcomes and partner outcomes that you can achieve through analytics, data science? So, working with AWS and with clients to manage that hybrid topology of some of that data being in the Cloud, being put to work with AWS SageMaker and Io-Tahoe being used to identify where is the data that needs to be amalgamated and curated to provide the dataset for machine learning, advanced analytics to have an impact for the business. >> So what are the critical attributes of what you're looking at to help customers decide what to move and what the keep if you will? >> Well, one of the quickest outcomes that we help customers achieve is to buy that business glossary, you know, that the items of data, that means something to them across those different silos and pull all of that together into a unified view. Once they've got that data engineer working with a business manager to think through, how do we want to create this application? Now, what is the churn model, the loyalty or the propensity model that we want to put in place here? How do we use predictive analytics to understand what needs for a patient that sort of innovation is what we're unlocking, applying a tools such as SageMaker on AWS to then do the computation and to build those models to deliver that outcome is across that value chain. And it goes back to the first picture that we put up, David, you know, the outcome is that API on the back of it, you've got a machine learning model that's been developed in a tool such as Databricks or Jupiter notebook. That data has to be sourced from somewhere. Somebody has to say that, "Yep, "You've got permission to do what you're trying to do without falling foul "of any compliance around data." And it all goes back to discovering that data, classifying it, indexing it in an automated way to cut those timelines down to hours and days. >> Yeah, it's the innovation part of your data portfolio, if you will, that you're going to put into the Cloud, apply tools like SageMaker and others, your tool Azure. I mean, whatever your favorite tool is, you don't care. The customer's going to choose that. And you know, the Cloud vendors, maybe they want you to use their tool, but they're making their marketplaces available to everybody, but it's that innovation piece, the ones that you, where you want to apply that self-service data marketplace to, and really drive, as I said before, monetization, All right, give us your final thoughts. Ajay, bring us home. >> So final thoughts on this, David, is at the moment, we're seeing a lot of value in helping customers discover their data using automation, automatically curating a data catalog. And that unified view is then being put to work through our API is having an open architecture to plug in whatever tool technology our clients have decided to use. And that open architecture is really feeding into the reality of what CIOs and Chief Data Officers are managing, which is a hybrid On-premise Cloud approach to use best of breed. But business users wanting to use a particular technology to get their business outcome, having the flexibility to do that no matter where your data is sitting On-premise, on Cloud is where self-service comes in so that sales service view of what data I can plug together, jive exchange, monetizing that data is where we're starting to see some real traction with customers. Now accelerating, becoming more digital to serve their own customers. >> Yeah, we really have seen a cultural mind shift going from sort of complacency, and obviously COVID has accelerated this, but the combination of that cultural shift, the Cloud machine intelligence tools give me a lot of hope that the promises of big data will ultimately be lived up to in this next 10 years. So Ajay Vohora, thanks so much for coming back on theCUBE. You're a great guest and appreciate your insights. >> Appreciate it, David. See you next time. >> All right, keep it right there, everybody, right back after this short break. (techno music)
SUMMARY :
Brought to you by Io-Tahoe. and automation to drive So the ability to exchange data help the audience understand and tasks that have to go into serving is the importance of APIs. all of that takes away from the job that has to go into achieving that. And that means, you know, and that is expanding the choices in terms of the impact And the last piece, I'd have to say And that really, to me is the bottom line, of the broker of the, of the major technologies that choice, the option might be the plan, that the middle there Ajay: Yeah, that that's the culmination has to have partnerships that the legacy that has built up. on getting to the Cloud, of some of that data being in the Cloud, that means something to them to apply that self-service having the flexibility to do that that the promises of big data See you next time. right back after this short break.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
David | PERSON | 0.99+ |
Ajay Vohora | PERSON | 0.99+ |
TCS | ORGANIZATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Io-Tahoe | ORGANIZATION | 0.99+ |
$4 trillion | QUANTITY | 0.99+ |
Tata Consulting Services | ORGANIZATION | 0.99+ |
five | QUANTITY | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
two parts | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
first picture | QUANTITY | 0.99+ |
fourth leg | QUANTITY | 0.99+ |
Ajay | PERSON | 0.99+ |
Io-Tahoe | PERSON | 0.99+ |
one | QUANTITY | 0.99+ |
20 years ago | DATE | 0.99+ |
U.S. | LOCATION | 0.99+ |
10 years ago | DATE | 0.98+ |
Three other elements | QUANTITY | 0.98+ |
360 view | QUANTITY | 0.98+ |
1% | QUANTITY | 0.98+ |
last month | DATE | 0.98+ |
first slide | QUANTITY | 0.97+ |
today | DATE | 0.97+ |
Power BI | TITLE | 0.97+ |
Cobra | LOCATION | 0.96+ |
Databricks | ORGANIZATION | 0.96+ |
10, 15 people | QUANTITY | 0.96+ |
single view | QUANTITY | 0.95+ |
Six Sigma | ORGANIZATION | 0.95+ |
GCP Azure | TITLE | 0.95+ |
single version | QUANTITY | 0.94+ |
Cloud | TITLE | 0.94+ |
Tableau | TITLE | 0.92+ |
Azure | TITLE | 0.88+ |
MongoDB | TITLE | 0.86+ |
about 10 years ago | DATE | 0.84+ |
COVID | TITLE | 0.83+ |
first | QUANTITY | 0.82+ |
Snowflake | TITLE | 0.81+ |
GCP | ORGANIZATION | 0.81+ |
two | QUANTITY | 0.81+ |
Looker | TITLE | 0.79+ |
SageMaker | TITLE | 0.78+ |
Greenfield | ORGANIZATION | 0.78+ |
next 10 years | DATE | 0.74+ |
Six Sigma | TITLE | 0.7+ |
this month | DATE | 0.67+ |
Jupiter | ORGANIZATION | 0.63+ |
Qlik | TITLE | 0.62+ |
AWS SageMaker | ORGANIZATION | 0.61+ |
a second | QUANTITY | 0.55+ |
SageMaker | ORGANIZATION | 0.54+ |
IO Tahoe DV Promo V1
>> Narrator: From around the globe, it's the CUBE with digital coverage of Smart Data Marketplaces, brought to you by Io-Tahoe. >> Hello this is Dave Vellante of the CUBE inviting you to join me for a special drill down presentation on the importance of automated data migration. Along with our friends from Io-Tahoe, we're going to explore the recent trends of automated data discovery, adaptive data governance, and just how far we've come from manually curating an enterprise data catalog. Ajay Vahora is the CEO of Io-Tahoe, as well as Stuti Deshpande of AWS and the digital evangelist Ved Sen of TCS, Tata Consultancy Services, will be there as well. Hope you can join us on Thursday, September 17th, at 9:00 a.m. Pacific for Smart Data Marketplaces. For more details, click on theCUBE.net. (upbeat music)
SUMMARY :
the globe, it's the CUBE on the importance of
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Stuti Deshpande | PERSON | 0.99+ |
Ajay Vahora | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Io-Tahoe | ORGANIZATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
CUBE | ORGANIZATION | 0.99+ |
TCS | ORGANIZATION | 0.99+ |
Tata Consultancy Services | ORGANIZATION | 0.99+ |
theCUBE.net | OTHER | 0.98+ |
IO Tahoe | ORGANIZATION | 0.95+ |
Thursday, September 17th, at | DATE | 0.91+ |
Ved Sen | PERSON | 0.83+ |
9:00 a.m. Pacific | DATE | 0.79+ |
Smart Data Marketplaces | EVENT | 0.57+ |
CEO | PERSON | 0.53+ |
DV Promo | COMMERCIAL_ITEM | 0.43+ |
Io-Tahoe Smart Data Lifecycle CrowdChat | Digital
>>from around the globe. It's the Cube with digital coverage of data automated and event. Siri's Brought to You by Iot Tahoe Welcome, everyone to the second episode in our data automated Siri's made possible with support from Iot Tahoe. Today we're gonna drill into the data lifecycle, meaning the sequence of stages that data travels through from creation to consumption to archive. The problem, as we discussed in our last episode, is that data pipelines, they're complicated, They're cumbersome, that disjointed, and they involve highly manual processes. Ah, smart data lifecycle uses automation and metadata to approve agility, performance, data quality and governance and ultimately reduce costs and time to outcomes. Now, in today's session will define the data lifecycle in detail and provide perspectives on what makes a data lifecycle smart and importantly, how to build smarts into your processes. In a moment, we'll be back with Adam Worthington from ethos to kick things off, and then we'll go into an export power panel to dig into the tech behind smart data life cycles, and it will hop into the crowdchat and give you a chance to ask questions. So stay right there. You're watching the cube innovation impact influence. Welcome >>to the Cube disruptors. Developers and practitioners learn from the voices of leaders who share their personal insights from the hottest digital events around the globe. Enjoy the best this community has to offer on the Cube, your global leader. >>High tech digital coverage. Okay, we're back with Adam Worthington. Adam, good to see you. How are things across the pond? >>Thank you, I'm sure. >>Okay, so let's let's set it up. Tell us about yourself. What? Your role is a CTO and >>automatically. As you said, we found a way to have a pretty in company ourselves that we're in our third year on. Do we specialize in emerging disruptive technologies within the infrastructure? That's the kind of cloud space on my phone is the technical lead. So I kind of my job to be an expert in all of the technologies that we work with, which can be a bit of a challenge if you have a huge for phone is one of the reasons, like deliberately focusing on on also kind of pieces a successful validation and evaluation of new technologies. >>So you guys really technology experts, data experts and probably also expert in process and delivering customer outcomes. Right? >>That's a great word there, Dave Outcomes. That's a lot of what I like to speak to customers about. >>Let's talk about smart data, you know, when you when you throw in terms like this is it kind of can feel buzz, wordy. But what are the critical aspects of so called smart data? >>Help to step back a little bit, seen a little bit more in terms of kind of where I can see the types of problems I saw. I'm really an infrastructure solution architect trace on and what I kind of benefit we organically. But over time my personal framework, I focused on three core design principal simplicity, flexibility, inefficient, whatever it was designing. And obviously they need different things, depending on what the technology area is working with. But that's a pretty good. So they're the kind of areas that a smart approach to data will directly address. Reducing silos that comes from simplifying, so moving away from conflict of infrastructure, reducing the amount of copies of data that we have across the infrastructure and reducing the amount of application environments that need different areas so smarter get with data in my eyes anyway, the further we moved away from this. >>But how does it work? I mean, how do you know what's what's involved in injecting smarts into your data lifecycle? >>I think one of my I actually did not ready, but generally one of my favorite quotes from the French lost a mathematician, Blaise Pascal. He said, If I get this right, I have written a short letter, but I didn't have time. But Israel, I love that quite for lots of reasons >>why >>direct application in terms of what we're talking about, it is actually really complicated. These developers technology capabilities to make things simple, more directly meet the needs of the business. So you provide self service capabilities that they just need to stop driving. I mean, making data on infrastructure makes the business users using >>your job. Correct me. If I'm wrong is to kind of put that all together in a solution and then help the customer realize that we talked about earlier that business out. >>Yeah, enough if they said in understanding both sides so that it keeps us on our ability to deliver on exactly what you just said is big experts in the capabilities and new a better way to do things but also having the kind of the business understanding to be able to ask the right questions. That's how new a better price is. Positions another area that I really like his stuff with their platforms. You can do more with less. And that's not just about using data redundancy. That's about creating application environments, that conservative and then the infrastructure to service different requirements that are able to use the random Io thing without getting too kind of low level as well as the sequential. So what that means is you don't necessarily have to move data from application environment a do one thing related, and then move it to the application environment. Be that environment free terms of an analytics on the left Right works. Both keep the data where it is, use it or different different requirements within the infrastructure and again do more with less. And what that does is not just about simplicity and efficiency. It significantly reduces the time to value of that as well. >>Do you have examples that you can share with us even if they're anonymous customers that you work with that are maybe a little further down on the journey. Or maybe not >>looking at the you mentioned data protection earlier. So another organization This is a project which is just kind of hearing confessions moment, huge organization. They're literally petabytes of data that was servicing their back up in archive. And what they have is not just this realization they have combined. I think I different that they have dependent on the what area of infrastructure they were backing up, whether it was virtualization, that was different because they were backing up PC's June 6th. They're backing up another database environment, using something else in the cloud knowledge bases approach that we recommended to work with them on. They were able to significantly reduce complexity and reduce the amount of time that it systems of what they were able to achieve and what this is again. One of the clients have They've gone above the threshold of being able to back up for that. >>Adam, give us the final thoughts, bring us home. In this segment, >>the family built something we didn't particularly such on, that I think it is really barely hidden. It is spoken about as much as I think it is, that agile approaches to infrastructure we're going to be touched on there could be complicated on the lack of it efficient, the impact, a user's ability to be agile. But what you find with traditional approaches and you already touched on some of the kind of benefits new approaches there. It's often very prescriptive, designed for a particular as the infrastructure environment, the way that it served up the users in kind of a packaged. Either way, it means that they need to use it in that whatever wave in data bases, that kind of service of as it comes in from a flexibility standpoint. But for this platform approach, which is the right way to address technology in my eyes enables, it's the infrastructure to be used. Flexible piece of it, the business users of the data users what we find this capability into their innovating in the way they use that on the White House. I bring benefits. This is a platform to prescriptive, and they are able to do that. What you're doing with these new approaches is all of the metrics that we touched on and pass it from a cost standpoint from a visibility standpoint, but what it means is that the innovators in the business want really, is to really understand what they're looking to achieve and now have to to innovate with us. Now, I think I've started to see that with projects season places. If you do it in the right way, you articulate the capability and empower the business users in the right ways. Very significantly. Better position. The advantages on really matching significantly bigger than their competition. Yeah, >>Super Adam in a really exciting space. And we spent the last 10 years gathering all this data, you know, trying to slog through it and figure it out. And now, with the tools that we have and the automation capabilities, it really is a new era of innovation and insights. So, Adam or they didn't thanks so much for coming on the Cube and participating in this program. >>Exciting times with that. Thank you very much Today. >>Now we're going to go into the power panel and go deeper into the technologies that enable smart data life cycles. Stay right there. You're watching the cube. Are >>you interested in test driving? The i o ta ho platform Kickstart the benefits of data automation for your business through the Iot Labs program. Ah, flexible, scalable sandbox environment on the cloud of your choice with set up a service and support provided by Iot. Top. Click on the Link and connect with the data engineer to learn more and see Iot Tahoe in action. >>Welcome back, everybody to the power panel driving business performance with smart data life cycles. Leicester Waters is here. He's the chief technology officer from Iot Tahoe. He's joined by Patrick Smith, who was field CTO from pure storage. And is that data? Who's a system engineering manager at KohI City? Gentlemen, good to see you. Thanks so much for coming on this panel. >>Thank you. >>Let's start with Lester. I wonder if each of you could just give us a quick overview of your role. And what's the number one problem that you're focused on solving for your customers? Let's start with Lester Fleet. >>Yes, I'm Lost Waters, chief technology officer for Iot Tahoe and really the number one problem that we're trying to solve for our customers is to understand, help them understand what they have, because if they don't understand what they have in terms of their data. They can't manage it. They can't control it. The cap monitor. They can't ensure compliance. So really, that's finding all you can about your data that you have. And building a catalog that could be readily consumed by the entire business is what we do. >>Patrick Field, CTO in your title That says to me, You're talking to customers all the time, so you got a good perspective on it. Give us your take on things here. >>Yeah, absolutely. So my patches in here on day talkto customers and prospects in lots of different verticals across the region. And as they look at their environments and their data landscape, they're faced with massive growth in the data that they're trying to analyze and demands to be able to get insight our stuff and to deliver better business value faster than they've ever had to do in the past. So >>got it. And is that of course, Kohi City. You're like the new kid on the block. You guys were really growing rapidly created this whole notion of data management, backup and and beyond. But I'm assistant system engineering manager. What are you seeing from from from customers your role and the number one problem that you're solving. >>Yeah, sure. So the number one problem I see time and again speaking with customers. It's around data fragmentation. So do two things like organic growth, even maybe budgetary limitations. Infrastructure has grown over time very piecemeal, and it's highly distributed internally. And just to be clear, you know, when I say internally, that >>could be >>that it's on multiple platforms or silos within an on Prem infrastructure that it also does extend to the cloud as well. >>Right Cloud is cool. Everybody wants to be in the cloud, right? So you're right, It creates, Ah, maybe unintended consequences. So let's start with the business outcome and kind of try to work backwards to people you know. They want to get more insights from data they want to have. Ah, Mawr efficient data lifecycle. But so let's let me start with you were thinking about like the North Star for creating data driven cultures. You know, what is the North Star or customers >>here? I think the North Star, in a nutshell, is driving value from your data. Without question, I mean way, differentiate ourselves these days by even nuances in our data now, underpinning that, there's a lot of things that have to happen to make that work out. Well, you know, for example, making sure you adequately protect your data, you know? Do you have a good You have a good storage sub system? Do you have a good backup and recovery point objectives? Recovery time objective. How do you Ah, are you fully compliant? Are you ensuring that you're taking all the boxes? There's a lot of regulations these days in terms with respect to compliance, data retention, data, privacy and so forth. Are you taking those boxes? Are you being efficient with your, uh, your your your data? You know, In other words, I think there's a statistic that someone mentioned me the other day that 53% of all businesses have between three and 15 copies of the same data. So you know, finding and eliminating does is it is part of the part of the problem is when you do a chase, >>um, I I like to think of you're right, no doubt, business value and and a lot of that comes from reducing the end in cycle times. But anything that you guys would would add to that. Patrick, Maybe start with Patrick. >>Yeah, I think I think in value from your data really hits on tips on what everyone wants to achieve. But I think there are a couple of key steps in doing that. First of all, is getting access to the data and asked that, Really, it's three big problems, firstly, working out what you've got. Secondly, looking at what? After working on what you've got, how to get access to it? Because it's all very well knowing that you've got some data. But if you can't get access to it either because of privacy reasons, security reasons, then that's a big challenge. And then finally, once you've got access to the data making sure that you can process that data in a timely manner >>for me, you know it would be that an organization has got a really good global view of all of its data. It understands the data flow and dependencies within their infrastructure, understands that precise legal and compliance requirements, and you had the ability to action changes or initiatives within their environment to give the fun. But with a cloud like agility. Um, you know, and that's no easy feat, right? That is hard work. >>Okay, so we've we've talked about. The challenge is in some of the objectives, but there's a lot of blockers out there, and I want to understand how you guys are helping remove them. So So, Lester. But what do you see as some of the big blockers in terms of people really leaning in? So this smart data lifecycle >>yeah, Silos is is probably one of the biggest one I see in business is yes, it's it's my data, not your data. Lots of lots of compartmentalization. Breaking that down is one of the one of the challenges. And having the right tools to help you do that is only part of the solution. There's obviously a lot of cultural things that need to take place Teoh to break down those silos and work together. If you can identify where you have redundant data across your enterprise, you might be able to consolidate those. >>So, Patrick, so one of the blockers that I see is legacy infrastructure, technical debt, sucking all the budget you got. You know, too many people have having to look after, >>as you look at the infrastructure that supports people's data landscapes today for primarily legacy reasons. The infrastructure itself is siloed. So you have different technologies with different underlying hardware and different management methodologies that they're there for good reason, because historically you have to have specific fitness, the purpose for different data requirements. And that's one of the challenges that we tackled head on a pure with with the flash blade technology and the concept of the data, a platform that can deliver in different characteristics for the different workloads. But from a consistent data platform >>now is that I want to go to you because, you know, in the world in your world, which to me goes beyond backup. And one of the challenges is, you know, they say backup is one thing. Recovery is everything, but as well. The the CFO doesn't want to pay for just protection, and one of things that I like about what you guys have done is you. You broadened the perspective to get more value out of your what was once seen as an insurance policy. >>I do see one of the one of the biggest blockers as the fact that the task at hand can, you know, can be overwhelming for customers. But the key here is to remember that it's not an overnight change. It's not, you know, a flick of a switch. It's something that can be tackled in a very piecemeal manner on. Absolutely. Like you said, You know, reduction in TCO and being able to leverage the data for other purposes is a key driver for this. So, you know, this can be this can be resolved. It would be very, you know, pretty straightforward. It can be quite painless as well. Same goes for unstructured data, which is very complex to manage. And, you know, we've all heard the stats from the the analysts. You know, data obviously is growing at an extremely rapid rate, but actually, when you look at that, you know how is actually growing. 80% of that growth is actually in unstructured data, and only 20% of that growth is in unstructured data. S o. You know, these are quick win areas that customers can realize immediate tco improvement and increased agility as well >>paint a picture of this guy that you could bring up the life cycle. You know what you can see here is you've got this this cycle, the data lifecycle and what we're wanting to do is inject intelligence or smarts into this, like like life cycles. You see, you start with ingestion or creation of data. You're you're storing it. You got to put it somewhere, right? You gotta classify it. You got to protect it. And then, of course, you want to reduce the copies, make it, you know, efficient on. And then you want to prepare it so that businesses can actually sumit. And then you've got clients and governance and privacy issues, and I wonder if we could start with you. Lester, this is, you know, the picture of the life cycle. What role does automation play in terms of injecting smarts into the lifecycle? >>Automation is key here, especially from the discover it catalog and classify perspective. I've seen companies where they geo and will take and dump their all of their database scheme is into a spreadsheet so that they can sit down and manually figure out what attributes 37 means for a column names, Uh, and that's that's only the tip of the iceberg. So being able to do automatically detect what you have automatically deduced where what's consuming the data, you know, upstream and downstream. Being able to understand all of the things related to the lifecycle of your data. Back up archive deletion. It is key. And so we're having having good tool. IShares is very >>important. So, Patrick, obviously you participate in the store piece of this picture s I wonder if you could talk more specifically about that. But I'm also interested in how you effect the whole system view the the end end cycle time. >>Yeah, I think Leicester kind of hit the nail on the head in terms of the importance of automation because the data volumes are just just so massive. Now that you can, you can you can effectively manage or understand or catalog your data without automation. Once you understand the data and the value of the data, then that's where you can work out where the data needs to be at any point in >>time, right? So pure and kohi city obviously partner to do that and of course, is that you guys were part of the protect you certainly part of the retain. But Also, you provide data management capabilities and analytics. I wonder if you could add some color there. >>Yeah, absolutely. So, like you said, you know, we focused pretty heavily on data protection. Is just one of our one of our areas on that infrastructure. It is just sitting there, really? Can, you know, with the legacy infrastructure, It's just sitting there, you know, consuming power, space cooling and pretty inefficient. And what, if anything, that protest is a key part of that. If I If I have a modern data platform such as, you know, the cohesive data platform, I can actually do a lot of analytics on that through application. So we have a marketplace for APS. >>I wonder if we could talk about metadata. It's It's increasingly important. Metadata is data about the data, but Leicester maybe explain why it's so important and what role it plays in terms of creating smart data lifecycle. A >>lot of people think it's just about the data itself, but there's a lot of extended characteristics about your data. So so imagine if or my data life cycle I can communicate with the backup system from Kohi City and find out when the last time that data was backed up or where is backed up to. I can communicate exchange data with pure storage and find out what two years? And is the data at the right tier commensurate with its use level pointed out and being able to share that metadata across systems? I think that's the direction that we're going in right now. We're at the stage where just identifying the metadata and trying to bring it together and catalog the next stage will be OK using the AP eyes it that that we have between our systems can't communicate and share that data and build good solutions for customers to use. >>It's a huge point that you just made. I mean, you know, 10 years ago, automating classification was the big problem, and it was machine intelligence, you know, obviously attacking that, But your point about as machines start communicating to each other and you start, it's cloud to cloud. There's all kinds of metadata, uh, kind of new meta data that's being created. I often joke that someday there's gonna be more metadata than data, so that brings us to cloud and that I'd like to start with you. >>You know, I do think, you know, having the cloud is a great thing. And it has got its role to play, and you can have many different permutations and iterations of how you use it on. Um, you know, I may have sort of mentioned previously. You know, I've seen customers go into the cloud very, very quickly, and actually recently, they're starting to remove workloads from the cloud. And the reason why this happens is that, you know, Cloud has got its role to play, but it's not right for absolutely everything, especially in their current form as well. A good analogy I like to use on this may sound a little bit cliche, but you know, when you compare clouds versus on premises data centers, you can use the analogy of houses and hotels. So to give you an idea so you know, when we look at hotels, that's like the equivalent of a cloud, right? I can get everything I need from there. I can get my food, my water, my outdoor facilities. If I need to accommodate more people, I can rent some more rooms. I don't have to maintain the hotel. It's all done for me. When you look at houses the equivalent to on premises infrastructure, I pretty much have to do everything myself, right. So I have to purchase the house. I have to maintain it. I have to buy my own food and water. Eat it. You have to make improvements myself. But then why do we all live in houses? No, in hotels. And the simple answer that I can I can only think of is, is that it's cheaper, right. It's cheaper to do it myself. But that's not to say that hotels haven't got their role to play. Um, you know? So, for example, if I've got loads of visitors coming over for the weekend, I'm not going to go build an extension to my house just for them. I will burst into my hotel into the cloud, um, and use it for, you know, for for things like that. So what I'm really saying is the cloud is great for many things, but it can work out costlier for certain applications, while others are a perfect >>It's an interesting analogy. I hadn't thought of that before, but you're right because I was going to say Well, part of it is you want the cloud experience everywhere, but you don't always want the cloud experience especially, you know, when you're with your family, you want certain privacy that I've not heard that before. He's out. So that's the new perspective s Oh, thank you, but but But Patrick, I do want to come back to that cloud experience because, in fact, that's what's happening. In a lot of cases, organizations are extending the cloud properties of automation on Prem. >>Yeah, I thought, as I thought, a really interesting point and a great analogy for the use of the public cloud. And it really reinforces the importance of the hybrid and multi cloud environment because it gives you the flexibility to choose where is the optimal environment to run your business workloads? And that's what it's all about and the flexibility to change which environment you're running in, either for more months to the next or from one year to the next. Because workloads change and the characteristics that are available in the cloud change, the hybrid cloud is something that we've we've lived with ourselves of pure, So our pure one management technology actually sits in hybrid cloud and what we we started off entirely cloud native. But now we use public cloud for compute. We use our own technology at the end of a high performance network link to support our data platform. So we get the best of both worlds and I think that's where a lot of our customers are trying to get to. >>Alright, I want to come back in a moment there. But before we do, let's see, I wonder if we could talk a little bit about compliance, governance and privacy. I think the Brits hung on. This panel is still in the EU for now, but the you are looking at new rules. New regulations going beyond GDP are where does sort of privacy governance, compliance fit in the data lifecycle, then, is that I want your thoughts on this as well. >>Yeah, this is this is a very important point because the landscape for for compliance, around data privacy and data retention is changing very rapidly. And being able to keep up with those changing regulations in an automated fashion is the only way you're gonna be able to do it. Even I think there's a some sort of Ah, maybe ruling coming out today or tomorrow with the changed in the r. So this is things are all very key points and being able to codify those rules into some software. Whether you know, Iot Tahoe or or your storage system or kohi city, it will help you be compliant is crucial. >>Yeah. Is that anything you can add there? I mean, it's really is your wheelhouse. >>Yeah, absolutely. So, you know, I think anybody who's watching this probably has gotten the message that, you know, less silos is better. And it absolutely it also applies to data in the cloud is where as well. So you know, my aiming Teoh consolidate into fewer platforms, customers can realize a lot better control over their data. And the natural effect of this is that it makes meeting compliance and governance a lot easier. So when it's consolidated, you can start to confidently understand who's accessing your data. How frequently are they accessing the data? You can also do things like, you know, detecting anomalous file access activities and quickly identify potential threats. >>Okay, Patrick, we were talking. You talked earlier about storage optimization. We talked to Adam Worthington about the business case, the numerator, which is the business value, and then the denominator, which is the cost and what's unique about pure in this regard. >>Yeah, and I think there are. There are multiple time dimensions to that. Firstly, if you look at the difference between legacy storage platforms that used to take up racks or aisles of space in the data center, the flash technology that underpins flash blade way effectively switch out racks rack units on. It has a big play in terms of data center footprint, and the environmental is associated with the data center. If you look at extending out storage efficiencies and the benefits it brings, just the performance has a direct effect on start we whether that's, you know, the start from the simplicity that platform so that it's easy and efficient to manage, whether it's the efficiency you get from your data. Scientists who are using the outcomes from the platform, making them more efficient to new. If you look at some of our customers in the financial space there, their time to results are improved by 10 or 20 x by switching to our technology from legacy technologies for their analytics, platforms. >>The guys we've been running, you know, Cube interviews in our studios remotely for the last 120 days is probably the first interview I've done where haven't started off talking about Cove it, Lester. I wonder if you could talk about smart data lifecycle and how it fits into this isolation economy. And hopefully, what will soon be a post isolation economy? >>Yeah, Come. It has dramatically accelerated the data economy. I think. You know, first and foremost, we've all learned to work at home. You know, we've all had that experience where, you know, people would have been all about being able to work at home just a couple days a week. And here we are working five days. That's how to knock on impact to infrastructure, to be able to support that. But going further than that, you know, the data economy is all about how a business can leverage their data to compete in this New World order that we are now in code has really been a forcing function to, you know, it's probably one of the few good things that have come out of government is that we've been forced to adapt and It's a zoo. Been an interesting journey and it continues to be so >>like Lester said, you know, we've We're seeing huge impact here. Working from home has pretty much become the norm. Now, you know, companies have been forced into basically making it work. If you look online retail, that's accelerated dramatically as well. Unified communications and videoconferencing. So really, you know the point here, is that Yes, absolutely. We're you know, we've compressed, you know, in the past, maybe four months. What already would have taken maybe even five years, maybe 10 years or so >>We got to wrap. But Celester Louis, let me ask you to sort of get paint. A picture of the sort of journey the maturity model that people have to take. You know, if they want to get into it, where did they start? And where are they going to give us that view, >>I think, versus knowing what you have. You don't know what you have. You can't manage it. You can't control that. You can't secure what you can't ensure. It's a compliant s so that that's first and foremost. Uh, the second is really, you know, ensuring that your compliance once, once you know what you have. Are you securing it? Are you following the regulatory? The applicable regulations? Are you able to evidence that, uh, how are you storing your data? Are you archiving it? Are you storing it effectively and efficiently? Um, you know, have you Nirvana from my perspective, is really getting to a point where you you've consolidated your data, you've broken down the silos and you have a virtually self service environment by which the business can consume and build upon their data. And really, at the end of the day, as we said at the beginning, it's all about driving value out of your data. And ah, the automation is is key to this, sir. This journey >>that's awesome and you just described is sort of a winning data culture. Lester, Patrick, thanks so much for participating in this power panel. >>Thank you, David. >>Alright, So great overview of the steps in the data lifecycle and how to inject smarts into the process is really to drive business outcomes. Now it's your turn. Hop into the crowd chat, please log in with Twitter or linked in or Facebook. Ask questions, answer questions and engage with the community. Let's crowdchat, right. Yeah, yeah, yeah.
SUMMARY :
behind smart data life cycles, and it will hop into the crowdchat and give you a chance to ask questions. Enjoy the best this community has to offer Adam, good to see you. and So I kind of my job to be an expert in all of the technologies that we work with, So you guys really technology experts, data experts and probably also expert in That's a lot of what I like to speak to customers Let's talk about smart data, you know, when you when you throw in terms like this is it kind of can feel buzz, reducing the amount of copies of data that we have across the infrastructure and reducing I love that quite for lots of reasons So you provide self service capabilities help the customer realize that we talked about earlier that business out. that it keeps us on our ability to deliver on exactly what you just said is big experts Do you have examples that you can share with us even if they're anonymous customers that you work looking at the you mentioned data protection earlier. In this segment, But what you find with traditional approaches and you already touched on some of you know, trying to slog through it and figure it out. Thank you very much Today. Now we're going to go into the power panel and go deeper into the technologies that enable Click on the Link and connect with the data Welcome back, everybody to the power panel driving business performance with smart data life I wonder if each of you could just give us a quick overview of your role. So really, that's finding all you can about your data that you so you got a good perspective on it. to deliver better business value faster than they've ever had to do in the past. What are you seeing from from from And just to be clear, you know, when I say internally, that it also does extend to the cloud as well. So let's start with the business outcome and kind of try to work backwards to people you and eliminating does is it is part of the part of the problem is when you do a chase, But anything that you guys would would add to that. But if you can't get access to it either because of privacy reasons, and you had the ability to action changes or initiatives within their environment to give But what do you see as some of the big blockers in terms of people really If you can identify where you have redundant data across your enterprise, technical debt, sucking all the budget you got. So you have different And one of the challenges is, you know, they say backup is one thing. But the key here is to remember that it's not an overnight the copies, make it, you know, efficient on. what you have automatically deduced where what's consuming the data, this picture s I wonder if you could talk more specifically about that. you can you can effectively manage or understand or catalog your data without automation. is that you guys were part of the protect you certainly part of the retain. Can, you know, with the legacy infrastructure, It's just sitting there, you know, consuming power, the data, but Leicester maybe explain why it's so important and what role it And is the data at the right tier commensurate with its use level pointed out I mean, you know, 10 years ago, automating classification And it has got its role to play, and you can have many different permutations and iterations of how you you know, when you're with your family, you want certain privacy that I've not heard that before. at the end of a high performance network link to support our data platform. This panel is still in the EU for now, but the you are looking at new Whether you know, Iot Tahoe or or your storage system I mean, it's really is your wheelhouse. So you know, my aiming Teoh consolidate into Worthington about the business case, the numerator, which is the business value, to manage, whether it's the efficiency you get from your data. The guys we've been running, you know, Cube interviews in our studios remotely for the last 120 days But going further than that, you know, the data economy is all about how a business can leverage we've compressed, you know, in the past, maybe four months. A picture of the sort of journey the maturity model that people have to take. from my perspective, is really getting to a point where you you've consolidated your that's awesome and you just described is sort of a winning data culture. Alright, So great overview of the steps in the data lifecycle and how to inject smarts into the process
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Patrick | PERSON | 0.99+ |
David | PERSON | 0.99+ |
Adam Worthington | PERSON | 0.99+ |
Adam Worthington | PERSON | 0.99+ |
Patrick Field | PERSON | 0.99+ |
Patrick Smith | PERSON | 0.99+ |
Adam | PERSON | 0.99+ |
five days | QUANTITY | 0.99+ |
June 6th | DATE | 0.99+ |
10 | QUANTITY | 0.99+ |
tomorrow | DATE | 0.99+ |
five years | QUANTITY | 0.99+ |
third year | QUANTITY | 0.99+ |
North Star | ORGANIZATION | 0.99+ |
Lester | PERSON | 0.99+ |
Siri | TITLE | 0.99+ |
10 years | QUANTITY | 0.99+ |
80% | QUANTITY | 0.99+ |
second episode | QUANTITY | 0.99+ |
Blaise Pascal | PERSON | 0.99+ |
Leicester Waters | ORGANIZATION | 0.99+ |
15 copies | QUANTITY | 0.99+ |
53% | QUANTITY | 0.99+ |
Lester | ORGANIZATION | 0.99+ |
Today | DATE | 0.99+ |
both sides | QUANTITY | 0.99+ |
four months | QUANTITY | 0.99+ |
each | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
two years | QUANTITY | 0.99+ |
20 x | QUANTITY | 0.99+ |
Iot Tahoe | ORGANIZATION | 0.99+ |
one | QUANTITY | 0.99+ |
first interview | QUANTITY | 0.99+ |
second | QUANTITY | 0.98+ |
Celester Louis | PERSON | 0.98+ |
ORGANIZATION | 0.98+ | |
Lester Fleet | ORGANIZATION | 0.98+ |
ORGANIZATION | 0.98+ | |
Both | QUANTITY | 0.98+ |
Firstly | QUANTITY | 0.98+ |
first | QUANTITY | 0.98+ |
one year | QUANTITY | 0.98+ |
10 years ago | DATE | 0.98+ |
White House | ORGANIZATION | 0.98+ |
One | QUANTITY | 0.98+ |
two things | QUANTITY | 0.97+ |
both worlds | QUANTITY | 0.97+ |
Secondly | QUANTITY | 0.97+ |
Iot | ORGANIZATION | 0.97+ |
Iot Labs | ORGANIZATION | 0.97+ |
20% | QUANTITY | 0.96+ |
Cove | ORGANIZATION | 0.96+ |
First | QUANTITY | 0.96+ |
Dave Outcomes | PERSON | 0.95+ |
firstly | QUANTITY | 0.95+ |
three big problems | QUANTITY | 0.94+ |
three core | QUANTITY | 0.94+ |
Israel | LOCATION | 0.94+ |
three | QUANTITY | 0.94+ |
KohI City | ORGANIZATION | 0.91+ |
Kohi City | LOCATION | 0.9+ |
one thing | QUANTITY | 0.89+ |
Leicester | ORGANIZATION | 0.89+ |
Io-Tahoe Smart Data Lifecycle CrowdChat | Digital
(upbeat music) >> Voiceover: From around the globe, it's theCUBE with digital coverage of Data Automated. An event series brought to you by Io-Tahoe. >> Welcome everyone to the second episode in our Data Automated series made possible with support from Io-Tahoe. Today, we're going to drill into the data lifecycle. Meaning the sequence of stages that data travels through from creation to consumption to archive. The problem as we discussed in our last episode is that data pipelines are complicated, they're cumbersome, they're disjointed and they involve highly manual processes. A smart data lifecycle uses automation and metadata to improve agility, performance, data quality and governance. And ultimately, reduce costs and time to outcomes. Now, in today's session we'll define the data lifecycle in detail and provide perspectives on what makes a data lifecycle smart? And importantly, how to build smarts into your processes. In a moment we'll be back with Adam Worthington from Ethos to kick things off. And then, we'll go into an expert power panel to dig into the tech behind smart data lifecyles. And, then we'll hop into the crowd chat and give you a chance to ask questions. So, stay right there, you're watching theCUBE. (upbeat music) >> Voiceover: Innovation. Impact. Influence. Welcome to theCUBE. Disruptors. Developers. And, practitioners. Learn from the voices of leaders, who share their personal insights from the hottest digital events around the globe. Enjoy the best this community has to offer on theCUBE. Your global leader in high tech digital coverage. >> Okay, we're back with Adam Worthington. Adam, good to see you, how are things across the pond? >> Good thank you, I'm sure our weather's a little bit worse than yours is over the other side, but good. >> Hey, so let's set it up, tell us about yourself, what your role is as CTO and--- >> Yeah, Adam Worthington as you said, CTO and co-founder of Ethos. But, we're a pretty young company ourselves, so we're in our sixth year. And, we specialize in emerging disruptive technology. So, within the infrastructure data center kind of cloud space. And, my role is a technical lead, so I, it's kind of my job to be an expert in all of the technologies that we work with. Which can be a bit of a challenge if you have a huge portfolio. One of the reasons we got to deliberately focus on. And also, kind of pieces of technical validation and evaluation of new technologies. >> So, you guys are really technology experts, data experts, and probably also expert in process and delivering customer outcomes, right? >> That's a great word there Dave, outcomes. I mean, that's a lot of what I like to speak to customers about. >> Let's talk about smart data you know, when you throw out terms like this it kind of can feel buzz wordy but what are the critical aspects of so-called smart data? >> Cool, well typically I had to step back a little bit and set the scene a little bit more in terms of kind of where I came from. So, and the types of problems I've sorted out. So, I'm really an infrastructure or solution architect by trade. And, what I kind of, relatively organically, but over time my personal framework and approach. I focused on three core design principles. So, simplicity, flexibility and efficiency. So, whatever it was I was designing and obviously they need different things depending on what the technology area is that we're working with. So, that's for me a pretty good step. So, they're the kind of areas that a smart approach in data will directly address both reducing silos. So, that comes from simplifying. So, moving away from complexity of infrastructure. Reducing the amount of copies of data that we have across the infrastructure. And, reducing the amount of application environment for the need for different areas. So, the smarter we get with data it's in my eyes anyway, the further we move away from those traditional legacy. >> But, how does it work? I mean, how, in other words, what's involved in injecting smarts into your data lifecycle? >> I think one of my, well actually I didn't have this quote ready, but genuinely one of my favorite quotes is from the French philosopher and mathematician, Blaise Pascal and he says, if I get this right, "I'd have written you a shorter letter, but I didn't have the time." So, there's real, I love that quote for lots of reasons. >> Dave: Alright. >> That's direct applications in terms of what we're talking about. In terms of, it's actually really complicated to develop a technology capability to make things simple. Be more directly meeting the needs of the business through tech. So, you provide self-service capability. And, I don't just mean self-driving, I mean making data and infrastructure make sense to the business users that are using it. >> Your job, correct me if I'm wrong, is to kind of put that all together in a solution. And then, help the customer you know, realize what we talked about earlier that business out. >> Yeah, and that's, it's sitting at both sides and understanding both sides. So, kind of key to us in our abilities to be able to deliver on exactly what you've just said, is being experts in the capabilities and new and better ways of doing things. But also, having the kind of, better business understanding to be able to ask the right questions to identify how can you better approach this 'cause it helps solve these issues. But, another area that I really like is the, with the platforms you can do more with less. And, that's not just about reducing data redundancy, that's about creating application environments that can service, an infrastructure to service different requirements that are able to do the random IO thing without getting too kind of low level tech. As well as the sequential. So, what that means is, that you don't necessarily have to move data from application environment A, do one thing with it, collate it and then move it to the application environment B, to application environment C, in terms of an analytics kind of left to right workload, you keep your data where it is, use it for different requirements within the infrastructure and again, do more with less. And, what that does, it's not just about simplicity and efficiency, it significantly reduces the times of value that that faces, as well. >> Do you have examples that you can share with us, even if they're anonymized of customers that you've worked with, that are maybe a little further down on the journey. Or, maybe not and--- >> Looking at the, you mentioned data protection earlier. So, another organization this is a project which is just coming nearing completion at the moment. Huge organization, that literally petabytes of data that was servicing their backup and archive. And, what they had is not just this reams of data. They had, I think I'm right in saying, five different backup applications that they had depending on the, what area of infrastructure they were backing up. So, whether it was virtualization, that was different to if they were backing up, different if they were backing up another data base environment they were using something else in the cloud. So, a consolidated approach that we recommended to work with them on. They were able to significantly reduce complexity and reduce the amount of time that it took them. So, what they were able to achieve and this was again, one of the key departments they had. They'd gone above the threshold of being able to backup all of them. >> Adam, give us the final thoughts, bring us home in this segment. >> Well, the final thoughts, so this is something, yeah we didn't particularly touch on. But, I think it's kind of slightly hidden, it isn't spoken about as much as I think it could be. Is the traditional approaches to infrastructure. We've already touched on that they can be complicated and there's a lack of efficiency. It impacts a user's ability to be agile. But, what you find with traditional approaches and we've already touched on some of the kind of benefits to new approaches there, is that they're often very prescriptive. They're designed for a particular firm. The infrastructure environment, the way that it's served up to the users in a kind of a packaged kind of way, means that they need to use it in that, whatever way it's been dictated. So, that kind of self-service aspect, as it comes in from a flexibility standpoint. But, these platforms and these platform approaches is the right way to address technology in my eyes. Enables the infrastructure to be used flexibly. So, the business users and the data users, what we find is that if we put in this capability into their hands. They start innovating the way that they use that data. And, the way that they bring benefits. And, if a platform is too prescriptive and they aren't able to do that, then what you're doing with these new approaches is get all of the metrics that we've touched on. It's fantastic from a cost standpoint, from an agility standpoint. But, what it means is that the innovators in the business, the ones that really understand what they're looking to achieve, they now have the tools to innovate with that. And, I think, and I've started to see that with projects that we've completed, if you do it in the right way, if you articulate the capability and you empower the business users in the right way. Then, they're in a significantly better position, these businesses to take advantages and really sort of match and significantly beat off their competition environment spaces. >> Super Adam, I mean a really exciting space. I mean we spent the last 10 years gathering all this data. You know, trying to slog through it and figure it out and now, with the tools that we have and the automation capabilities, it really is a new era of innovation and insight. So, Adam Worthington, thanks so much for coming in theCUBE and participating in this program. >> Yeah, exciting times and thank you very much Dave for inviting me, and yeah big pleasure. >> Now, we're going to go into the power panel and go deeper into the technologies that enable smart data lifecyles. And, stay right there, you're watching theCUBE. (light music) >> Voiceover: Are you interested in test-driving the Io-Tahoe platform? Kickstart the benefits of Data Automation for your business through the IoLabs program. A flexible, scalable, sandbox environment on the cloud of your choice. With setup, service and support provided by Io-Tahoe. Click on the link and connect with a data engineer to learn more and see Io-Tahoe in action. >> Welcome back everybody to the power panel, driving business performance with smart data lifecyles. Lester Waters is here, he's the Chief Technology Officer from Io-Tahoe. He's joined by Patrick Smith, who is field CTO from Pure Storage. And, Ezat Dayeh who is Assistant Engineering Manager at Cohesity. Gentlemen, good to see you, thanks so much for coming on this panel. >> Thank you, Dave. >> Yes. >> Thank you, Dave. >> Let's start with Lester, I wonder if each of you could just give us a quick overview of your role and what's the number one problem that you're focused on solving for your customers? Let's start with Lester, please. >> Ah yes, I'm Lester Waters, Chief Technology Officer for Io-Tahoe. And really, the number one problem that we are trying to solve for our customers is to help them understand what they have. 'Cause if they don't understand what they have in terms of their data, they can't manage it, they can't control it, they can't monitor it, they can't ensure compliance. So, really that's finding all that you can about your data that you have and building a catalog that can be readily consumed by the entire business is what we do. >> Patrick, field CTO in your title, that says to me you're talking to customers all the time so you've got a good perspective on it. Give us you know, your take on things here. >> Yeah absolutely, so my patch is in the air and talk to customers and prospects in lots of different verticals across the region. And, as they look at their environments and their data landscape, they're faced with massive growth in the data that they're trying to analyze. And, demands to be able to get inside are faster. And, to deliver business value faster than they've ever had to do in the past, so. >> Got it and then Ezat at Cohesity, you're like the new kid on the block. You guys are really growing rapidly. You created this whole notion of data management, backup and beyond, but from Assistant Engineering Manager what are you seeing from customers, your role and the number one problem that you're solving? >> Yeah sure, so the number one problem I see you know, time and again speaking with customers it's all around data fragmentation. So, due to things like organic growth you know, even maybe budgetary limitations, infrastructure has grown you know, over time, very piecemeal. And, it's highly distributed internally. And, just to be clear you know, when I say internally you know, that could be that it's on multiple platforms or silos within an on-prem infrastructure. But, that it also does extend to the cloud, as well. >> Right hey, cloud is cool, everybody wants to be in the cloud, right? So, you're right it creates maybe unattended consequences. So, let's start with the business outcome and kind of try to work backwards. I mean people you know, they want to get more insights from data, they want to have a more efficient data lifecyle. But, so Lester let me start with you, in thinking about like, the North Star, creating data driven cultures you know, what is the North Star for customers here? >> I think the North Star in a nutshell is driving value from your data. Without question, I mean we differentiate ourselves these days by even the nuances in our data. Now, underpinning that there's a lot of things that have to happen to make that work out well. You know for example, making sure you adequately protect your data. You know, do you have a good storage system? Do you have a good backup and recovery point objectives, recovering time objectives? Do you, are you fully compliant? Are you ensuring that you're ticking all the boxes? There's a lot of regulations these days in terms, with respect to compliance, data retention, data privacy and so fourth. Are you ticking those boxes? Are you being efficient with your data? You know, in other words I think there's a statistic that someone mentioned to me the other day that 53% of all businesses have between three and 15 copies of the same data. So you know, finding and eliminating those is part of the problems you need to chase. >> I like to think of you know, you're right. Lester, no doubt, business value and a lot of that comes from reducing the end to end cycle times. But, anything that you guys would add to that, Patrick and Ezat, maybe start with Patrick. >> Yeah, I think getting value from data really hits on, it hits on what everyone wants to achieve. But, I think there are a couple of key steps in doing that. First of all is getting access to the data. And that's, that really hits three big problems. Firstly, working out what you've got. Secondly, after working out what you've got, how to get access to it. Because, it's all very well knowing that you've got some data but if you can't get access to it. Either, because of privacy reasons, security reasons. Then, that's a big challenge. And then finally, once you've got access to the data, making sure that you can process that data in a timely manner. >> For me you know, it would be that an organization has got a really good global view of all of its data. It understands the data flow and dependencies within their infrastructure. Understands the precise legal and compliance requirements. And, has the ability to action changes or initiatives within their environment. Forgive the pun, but with a cloud like agility. You know, and that's no easy feat, right? That is hard work. >> Okay, so we've talked about the challenges and some of the objectives, but there's a lot of blockers out there and I want to understand how you guys are helping remove them? So, Lester what do you see as some of the big blockers in terms of people really leaning in to this smart data lifecycle. >> Yeah silos, is probably one of the biggest one I see in businesses. Yes, it's my data not your data. Lots of compartmentalization. And, breaking that down is one of the challenges. And, having the right tools to help you do that is only part of the solution. There's obviously a lot of cultural things that need to take place to break down those silos and work together. If you can identify where you have redundant data across your enterprise, you might be able to consolidate those. >> Yeah so, over to Patrick, so you know, one of the blockers that I see is legacy infrastructure, technical debt sucking all the budget. You got you know, too many people having to look after. >> As you look at the infrastructure that supports peoples data landscapes today. For primarily legacy reasons, the infrastructure itself is siloed. So, you have different technologies with different underlying hardware, different management methodologies that are there for good reason. Because, historically you had to have specific fitness for purpose for different data requirements. >> Dave: Ah-hm. >> And, that's one of the challenges that we tackled head on at Pure. With the flash plate technology and the concept of the data hub. A platform that can deliver in different characteristics for the different workloads. But, from a consistent data platform. >> Now, Ezat I want to go to you because you know, in the world, in your world which to me goes beyond backup and one of the challenges is you know, they say backup is one thing, recovery is everything. But as well, the CFO doesn't want to pay for just protection. Now, one of the things that I like about what you guys have done is you've broadened the perspective to get more value out of your what was once seen as an insurance policy. >> I do see one of the biggest blockers as the fact that the task at hand can you know, be overwhelming for customers. But, the key here is to remember that it's not an overnight change, it's not you know, the flick of the switch. It's something that can be tackled in a very piecemeal manner. And, absolutely like you've said you know, reduction in TCO and being able to leverage the data for other purposes is a key driver for this. So you know, this can be resolved. It can be very you know, pretty straightforward. It can be quite painless, as well. Same goes for unstructured data, which is very complex to manage. And you know, we've all heard the stats from the analysts, you know data obviously is growing at an extremely rapid rate. But, actually when you look at that you know, how is it actually growing? 80% of that growth is actually in unstructured data and only 20% of that growth is in structured data. So you know, these are quick win areas that the customers can realize immediate TCO improvement and increased agility, as well. >> Let's paint a picture of this guys, if I can bring up the lifecyle. You know what you can see here is you've got this cycle, the data lifecycle and what we're wanting to do is inject intelligence or smarts into this lifecyle. So, you can see you start with ingestion or creation of data. You're storing it, you've got to put it somewhere, right? You've got to classify it, you've got to protect it. And then, of course you want to you know, reduce the copies, make it you know, efficient. And then, you want to prepare it so that businesses can actually consume it and then you've got compliance and governance and privacy issues. And, I wonder if we could start with you Lester, this is you know, the picture of the lifecycle. What role does automation play in terms of injecting smarts into the lifecycle? >> Automation is key here, you know. Especially from the discover, catalog and classify perspective. I've seen companies where they go and we'll take and dump all of their data base schemes into a spreadsheet. So, that they can sit down and manually figure out what attribute 37 means for a column name. And, that's only the tip of the iceberg. So, being able to automatically detect what you have, automatically deduce where, what's consuming the data, you know upstream and downstream, being able to understand all of the things related to the lifecycle of your data backup, archive, deletion, it is key. And so, having good toolage areas is very important. >> So Patrick, obviously you participate in the store piece of this picture. So, I wondered if you could just talk more specifically about that, but I'm also interested in how you affect the whole system view, the end-to-end cycle time. >> Yeah, I think Lester kind of hit the nail on the head in terms of the importance of automation. Because, the data volumes are just so massive now that you can't effectively manage or understand or catalog your data without automation. Once you understand the data and the value of the data, then that's where you can work out where the data needs to be at any point in time. >> Right, so Pure and Cohesity obviously partnered to do that and of course, Ezat you guys are part of the protect, you're certainly part of the retain. But also, you provide data management capabilities and analytics, I wonder if you could add some color there? >> Yeah absolutely, so like you said you know, we focus pretty heavily on data protection as just one of our areas. And, that infrastructure it is just sitting there really can you know, the legacy infrastructure it's just sitting there you know, consuming power, space, cooling and pretty inefficient. And, automating that process is a key part of that. If I have a modern day platform such as you know, the Cohesity data platform I can actually do a lot of analytics on that through applications. So, we have a marketplace for apps. >> I wonder if we could talk about metadata. It's increasingly important you know, metadata is data about the data. But, Lester maybe explain why it's so important and what role it plays in terms of creating smart data lifecycle. >> A lot of people think it's just about the data itself. But, there's a lot of extended characteristics about your data. So, imagine if for my data lifecycle I can communicate with the backup system from Cohesity. And, find out when the last time that data was backed up or where it's backed up to. I can communicate, exchange data with Pure Storage and find out what tier it's on. Is the data at the right tier commencer with it's use level? If I could point it out. And, being able to share that metadata across systems. I think that's the direction that we're going in. Right now, we're at the stage we're just identifying the metadata and trying to bring it together and catalog it. The next stage will be okay, using the APIs and that we have between our systems. Can we communicate and share that data and build good solutions for customers to use? >> I think it's a huge point that you just made, I mean you know 10 years ago, automating classification was the big problem. And you know, with machine intelligence you know, we're obviously attacking that. But, your point about as machines start communicating to each other and you start you know, it's cloud to cloud. There's all kinds of metadata, kind of new metadata that's being created. I often joke that some day there's going to be more metadata than data. So, that brings us to cloud and Ezat, I'd like to start with you. >> You know, I do think that you know, having the cloud is a great thing. And, it has got its role to play and you can have many different you know, permutations and iterations of how you use it. And, you know, as I've may have sort of mentioned previously you know, I've seen customers go into the cloud very, very quickly and actually recently they're starting to remove workloads from the cloud. And, the reason why this happens is that you know, cloud has got its role to play but it's not right for absolutely everything. Especially in their current form, as well. A good analogy I like to use and this may sound a little bit clique but you know, when you compare clouds versus on premises data centers. You can use the analogies of houses and hotels. So, to give you an idea, so you know, when we look at hotels that's like the equivalent of a cloud, right? I can get everything I need from there. I can get my food, my water, my outdoor facilities, if I need to accommodate more people, I can rent some more rooms. I don't have to maintain the hotel, it's all done for me. When you look at houses the equivalent to you know, on premises infrastructure. I pretty much have to do everything myself, right? So, I have to purchase the house, I have to maintain it, I have buy my own food and water, eat it, I have to make improvements myself. But, then why do we all live in houses, not in hotels? And, the simple answer that I can only think of is, is that it's cheaper, right? It's cheaper to do it myself, but that's not to say that hotels haven't got their role to play. You know, so for example if I've got loads of visitors coming over for the weekend, I'm not going to go and build an extension to my house, just for them. I will burst into my hotel, into the cloud. And, you use it for you know, for things like that. So, what I'm really saying is the cloud is great for many things, but it can work out costlier for certain applications, while others are a perfect fit. >> That's an interesting analogy, I hadn't thought of that before. But, you're right, 'cause I was going to say well part of it is you want the cloud experience everywhere. But, you don't always want the cloud experience, especially you know, when you're with your family, you want certain privacy. I've not heard that before, Ezat. So, that's a new perspective, so thank you. But, Patrick I do want to come back to that cloud experience because in fact that's what's happening in a lot of cases. Organizations are extending the cloud properties of automation on-prem. >> Yeah, I thought Ezat brought up a really interesting point and a great analogy for the use of the public cloud. And, it really reinforces the importance of the Hybrid and the multicloud environment. Because, it gives you that flexibility to choose where is the optimal environment to run your business workloads. And, that's what it's all about. And, the flexibility to change which environment you're running in, either from one month to the next or from one year to the next. Because, workloads change and the characteristics that are available in the cloud change. The Hybrid cloud is something that we've lived with ourselves at Pure. So, our Pure management technology actually sits in a Hybrid cloud environment. We started off entirely cloud native but now, we use the public cloud for compute and we use our own technology at the end of a high performance network link to support our data platform. So, we're getting the best of both worlds. I think that's where a lot of our customers are trying to get to. >> All right, I want to come back in a moment there. But before we do, Lester I wonder if we could talk a little bit about compliance and governance and privacy. I think the Brits on this panel, we're still in the EU for now but the EU are looking at new rules, new regulations going beyond GDPR. Where does sort of privacy, governance, compliance fit in for the data lifecycle. And Ezat, I want your thought on this as well? >> Ah yeah, this is a very important point because the landscape for compliance around data privacy and data retention is changing very rapidly. And, being able to keep up with those changing regulations in an automated fashion is the only way you're going to be able to do it. Even, I think there's a some sort of a maybe ruling coming out today or tomorrow with a change to GDPR. So, this is, these are all very key points and being able to codify those rules into some software whether you know, Io-Tahoe or your storage system or Cohesity, it'll help you be compliant is crucial. >> Yeah, Ezat anything you can add there, I mean this really is your wheel house? >> Yeah, absolutely, so you know, I think anybody who's watching this probably has gotten the message that you know, less silos is better. And, it absolutely it also applies to data in the cloud, as well. So you know, by aiming to consolidate into you know, fewer platforms customers can realize a lot better control over their data. And, the natural affect of this is that it makes meeting compliance and governance a lot easier. So, when it's consolidated you can start to confidently understand who's accessing your data, how frequently are they accessing the data. You can also do things like you know, detecting an ominous file access activities and quickly identify potential threats. >> Okay Patrick, we were talking, you talked earlier about storage optimization. We talked to Adam Worthington about the business case, you've got the sort numerator which is the business value and then a denominator which is the cost. And, what's unique about Pure in this regard? >> Yeah, and I think there are multiple dimensions to that. Firstly, if you look at the difference between legacy storage platforms, they used to take up racks or aisles of space in a data center. With flash technology that underpins flash played we effectively switch out racks for rack units. And, it has a big play in terms of data center footprint and the environmentals associated with a data center. If you look at extending out storage efficiencies and the benefits it brings. Just the performance has a direct effect on staff. Whether that's you know, the staff and the simplicity of the platform so that it's easy and efficient to manage. Or, whether it's the efficiency you get from your data scientists who are using the outcomes from the platform and making them more efficient. If you look at some of our customers in the financial space their time to results are improved by 10 or 20 x by switching to our technology. From legacy technologies for their analytics platforms. >> So guys, we've been running you know, CUBE interviews in our studios remotely for the last 120 days. This is probably the first interview I've done where I haven't started off talking about COVID. Lester, I wondered if you could talk about smart data lifecycle and how it fits into this isolation economy and hopefully what will soon be a post-isolation economy? >> Yeah, COVID has dramatically accelerated the data economy. I think you know, first and foremost we've all learned to work at home. I you know, we've all had that experience where you know, people would hum and har about being able to work at home just a couple of days a week. And, here we are working five days a week. That's had a knock on impact to infrastructure to be able to support that. But, going further than that you know, the data economy is all about how a business can leverage their data to compete in this new world order that we are now in. COVID has really been a forcing function to you know, it's probably one of the few good things that have come out of COVID is that we've been forced to adapt. And, it's been an interesting journey and it continues to be so. >> Like Lester said you know, we're seeing huge impact here. You know, working from home has pretty much become the norm now. You know, companies have been forced into making it work. If you look at online retail, that's accelerated dramatically, as well. Unified communications and video conferencing. So, really you know, that the point here is that, yes absolutely we've compressed you know, in the past maybe four months what probably would have taken maybe even five years, maybe 10 years or so. >> We've got to wrap, but so Lester let me ask you, sort of paint a picture of the sort of journey the maturity model that people have to take. You know, if they want to get into it, where do they start and where are they going? Give us that view. >> Yeah, I think first is knowing what you have. If you don't know what you have you can't manage it, you can't control it, you can't secure it, you can't ensure it's compliant. So, that's first and foremost. The second is really you know, ensuring that you're compliant once you know what you have, are you securing it? Are you following the regulatory, the regulations? Are you able to evidence that? How are you storing your data? Are you archiving it? Are you storing it effectively and efficiently? You know, have you, nirvana from my perspective is really getting to a point where you've consolidated your data, you've broken down the silos and you have a virtually self-service environment by which the business can consume and build upon their data. And, really at the end of the day as we said at the beginning, it's all about driving value out of your data. And, automation is key to this journey. >> That's awesome and you've just described like sort of a winning data culture. Lester, Patrick, Ezat, thanks so much for participating in this power panel. >> Thank you, David. >> Thank you. >> All right, so great overview of the steps in the data lifecyle and how to inject smarts into the processes, really to drive business outcomes. Now, it's your turn, hop into the crowd chat. Please log in with Twitter or LinkedIn or Facebook, ask questions, answer questions and engage with the community. Let's crowd chat! (bright music)
SUMMARY :
to you by Io-Tahoe. and give you a chance to ask questions. Enjoy the best this community Adam, good to see you, how Good thank you, I'm sure our of the technologies that we work with. I like to speak to customers about. So, and the types of is from the French of the business through tech. And then, help the customer you know, to identify how can you that you can share with us, and reduce the amount of Adam, give us the final thoughts, the kind of benefits to and the automation capabilities, thank you very much Dave and go deeper into the technologies on the cloud of your choice. he's the Chief Technology I wonder if each of you So, really that's finding all that you can Give us you know, your in the data that they're and the number one problem And, just to be clear you know, I mean people you know, they is part of the problems you need to chase. from reducing the end to end cycle times. making sure that you can process And, has the ability to action changes So, Lester what do you see as some of And, having the right tools to help you Yeah so, over to Patrick, so you know, So, you have different technologies and the concept of the data hub. the challenges is you know, the analysts, you know to you know, reduce the copies, And, that's only the tip of the iceberg. in the store piece of this picture. the data needs to be at any point in time. and analytics, I wonder if you it's just sitting there you know, It's increasingly important you know, And, being able to share to each other and you start So, to give you an idea, so you know, especially you know, when And, the flexibility to change compliance fit in for the data lifecycle. in an automated fashion is the only way You can also do things like you know, about the business case, Whether that's you know, you know, CUBE interviews forcing function to you know, So, really you know, that of the sort of journey And, really at the end of the day for participating in this power panel. the processes, really to
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Patrick | PERSON | 0.99+ |
David | PERSON | 0.99+ |
Ezat Dayeh | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Adam Worthington | PERSON | 0.99+ |
Patrick Smith | PERSON | 0.99+ |
Adam | PERSON | 0.99+ |
Ezat | PERSON | 0.99+ |
80% | QUANTITY | 0.99+ |
10 | QUANTITY | 0.99+ |
second episode | QUANTITY | 0.99+ |
Blaise Pascal | PERSON | 0.99+ |
53% | QUANTITY | 0.99+ |
five years | QUANTITY | 0.99+ |
tomorrow | DATE | 0.99+ |
10 years | QUANTITY | 0.99+ |
EU | ORGANIZATION | 0.99+ |
sixth year | QUANTITY | 0.99+ |
Io-Tahoe | ORGANIZATION | 0.99+ |
Ethos | ORGANIZATION | 0.99+ |
North Star | ORGANIZATION | 0.99+ |
Lester | PERSON | 0.99+ |
Cohesity | ORGANIZATION | 0.99+ |
second | QUANTITY | 0.99+ |
both sides | QUANTITY | 0.99+ |
first interview | QUANTITY | 0.99+ |
each | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
one month | QUANTITY | 0.99+ |
Lester Waters | PERSON | 0.99+ |
GDPR | TITLE | 0.98+ |
today | DATE | 0.98+ |
Firstly | QUANTITY | 0.98+ |
one year | QUANTITY | 0.98+ |
15 copies | QUANTITY | 0.98+ |
ORGANIZATION | 0.98+ | |
First | QUANTITY | 0.98+ |
Today | DATE | 0.98+ |
20 x | QUANTITY | 0.98+ |
one | QUANTITY | 0.98+ |
both | QUANTITY | 0.97+ |
10 years ago | DATE | 0.97+ |
four months | QUANTITY | 0.97+ |
five days a week | QUANTITY | 0.97+ |
Secondly | QUANTITY | 0.97+ |
ORGANIZATION | 0.97+ | |
both worlds | QUANTITY | 0.97+ |
ORGANIZATION | 0.97+ | |
three | QUANTITY | 0.96+ |
One | QUANTITY | 0.96+ |
Pure Storage | ORGANIZATION | 0.95+ |
Lester | ORGANIZATION | 0.94+ |
20% | QUANTITY | 0.94+ |
Pure | ORGANIZATION | 0.93+ |
fourth | QUANTITY | 0.93+ |
Paula D'Amico, Webster Bank | Io Tahoe | Enterprise Data Automation
>>from around the globe. It's the Cube with digital coverage of enterprise data automation, an event Siri's brought to you by Iot. Tahoe, >>my buddy, We're back. And this is Dave Volante, and we're covering the whole notion of automating data in the Enterprise. And I'm really excited to have Paul Damico here. She's a senior vice president of enterprise data Architecture at Webster Bank. Good to see you. Thanks for coming on. >>Hi. Nice to see you, too. Yes. >>So let's let's start with Let's start with Webster Bank. You guys are kind of a regional. I think New York, New England, uh, leave headquartered out of Connecticut, but tell us a little bit about the bank. >>Yeah, Um, Webster Bank >>is regional Boston And that again, and New York, Um, very focused on in Westchester and Fairfield County. Um, they're a really highly rated saying regional bank for this area. They, um, hold, um, quite a few awards for the area for being supportive for the community and, um, are really moving forward. Technology lives. They really want to be a data driven bank, and they want to move into a more robust Bruce. >>Well, we got a lot to talk about. So data driven that is an interesting topic. And your role as data architect. The architecture is really senior vice president data architecture. So you got a big responsibility as it relates to It's kind of transitioning to this digital data driven bank. But tell us a little bit about your role in your organization, >>right? Um, currently, >>today we have, ah, a small group that is just working toward moving into a more futuristic, more data driven data warehouse. That's our first item. And then the other item is to drive new revenue by anticipating what customers do when they go to the bank or when they log into there to be able to give them the best offer. The only way to do that is you >>have uh huh. >>Timely, accurate, complete data on the customer and what's really a great value on off something to offer that or a new product or to help them continue to grow their savings or do and grow their investment. >>Okay. And I really want to get into that. But before we do and I know you're sort of part way through your journey, you got a lot of what they do. But I want to ask you about Cove. It how you guys you're handling that? I mean, you had the government coming down and small business loans and P p p. And huge volume of business and sort of data was at the heart of that. How did you manage through that? >>But we were extremely successful because we have a big, dedicated team that understands where their data is and was able to switch much faster than a larger bank to be able to offer. The TPP longs at to our customers within lightning speeds. And part of that was is we adapted to Salesforce very, for we've had salesforce in house for over 15 years. Um, you know, pretty much, uh, that was the driving vehicle to get our CPP is loans in on and then developing logic quickly. But it was a 24 7 development role in get the data moving, helping our customers fill out the forms. And a lot of that was manual. But it was a It was a large community effort. >>Well, think about that. Think about that too. Is the volume was probably much, much higher the volume of loans to small businesses that you're used to granting. But and then also, the initial guidelines were very opaque. You really didn't know what the rules were, but you were expected to enforce them. And then finally, you got more clarity. So you had to essentially code that logic into the system in real time, right? >>I wasn't >>directly involved, but part of my data movement Team Waas, and we had to change the logic overnight. So it was on a Friday night was released. We've pushed our first set of loans through and then the logic change, Um, from, you know, coming from the government and changed. And we had to re develop our our data movement piece is again and we design them and send them back. So it was It was definitely kind of scary, but we were completely successful. We hit a very high peak and I don't know the exact number, but it was in the thousands of loans from, you know, little loans to very large loans, and not one customer who buy it's not yet what they needed for. Um, you know, that was the right process and filled out the rate and pace. >>That's an amazing story and really great support for the region. New York, Connecticut, the Boston area. So that's that's fantastic. I want to get into the rest of your story. Now let's start with some of the business drivers in banking. I mean, obviously online. I mean, a lot of people have sort of joked that many of the older people who kind of shunned online banking would love to go into the branch and see their friendly teller had no choice, You know, during this pandemic to go to online. So that's obviously a big trend you mentioned. So you know the data driven data warehouse? I wanna understand that. But well, at the top level, what were some of what are some of the key business drivers there catalyzing your desire for change? >>Um, the ability to give the customer what they need at the time when they need it. And what I mean by that is that we have, um, customer interactions in multiple ways, right? >>And I want >>to be able for the customer, too. Walk into a bank, um, or online and see the same the same format and being able to have the same feel, the same look, and also to be able to offer them the next best offer for them. But they're you know, if they want looking for a new a mortgage or looking to refinance or look, you know, whatever it iss, um, that they have that data, we have the data and that they feel comfortable using it. And that's a untethered banker. Um, attitude is, you know, whatever my banker is holding and whatever the person is holding in their phone, that that is the same. And it's comfortable, so they don't feel that they've, you know, walked into the bank and they have to do a lot of different paperwork comparative filling out paperwork on, you know, just doing it on their phone. >>So you actually want the experience to be better. I mean, and it is in many cases now, you weren't able to do this with your existing against mainframe based Enterprise data warehouse. Is is that right? Maybe talk about that a little bit. >>Yeah, we were >>definitely able to do it with what we have today. The technology we're using, but one of the issues is that it's not timely, Um, and and you need a timely process to be able to get the customers to understand what's happening. Um, you want you need a timely process so we can enhance our risk management. We can apply for fraud issues and things like that. >>Yeah, so you're trying to get more real time in the traditional e g W. It's it's sort of a science project. There's a few experts that know how to get it. You consider line up. The demand is tremendous, and often times by the time you get the answer, you know it's outdated. So you're trying to address that problem. So So part of it is really the cycle time, the end end cycle, time that you're pressing. And then there's if I understand it, residual benefits that are pretty substantial from a revenue opportunity. Other other offers that you can you can make to the right customer, Um, that that you, you maybe know through your data. Is that right? >>Exactly. It's drive new customers, Teoh new opportunities. It's enhanced the risk, and it's to optimize the banking process and then obviously, to create new business. Um, and the only way we're going to be able to do that is that we have the ability to look at the data right when the customer walks in the door or right when they open up their app. And, um, by doing, creating more to New York time near real time data for the data warehouse team that's giving the lines of business the ability to to work on the next best offer for that customer. >>Paulo, we're inundated with data sources these days. Are there their data sources that you maybe maybe had access to before? But perhaps the backlog of ingesting and cleaning and cataloging and you know of analyzing. Maybe the backlog was so great that you couldn't perhaps tap some of those data sources. You see the potential to increase the data sources and hence the quality of the data, Or is that sort of premature? >>Oh, no. Um, >>exactly. Right. So right now we ingest a lot of flat files and from our mainframe type of Brennan system that we've had for quite a few years. But now that we're moving to the cloud and off Prem and on France, you know, moving off Prem into like an s three bucket. Where That data king, We can process that data and get that data faster by using real time tools to move that data into a place where, like, snowflake could utilize that data or we can give it out to our market. >>Okay, so we're >>about the way we do. We're in batch mode. Still, so we're doing 24 hours. >>Okay, So when I think about the data pipeline and the people involved, I mean, maybe you could talk a little bit about the organization. I mean, you've got I know you have data. Scientists or statisticians? I'm sure you do. Ah, you got data architects, data engineers, quality engineers, you know, developers, etcetera, etcetera. And oftentimes, practitioners like yourself will will stress about pay. The data's in silos of the data quality is not where we want it to be. We have to manually categorize the data. These are all sort of common data pipeline problems, if you will. Sometimes we use the term data ops, which is kind of a play on Dev Ops applied to the data pipeline. I did. You just sort of described your situation in that context. >>Yeah. Yes. So we have a very large data ops team and everyone that who is working on the data part of Webster's Bay has been there 13 14 years. So they get the data, they understand that they understand the lines of business. Um, so it's right now, um, we could we have data quality issues, just like everybody else does. We have. We have places in him where that gets clans, Um, and we're moving toward. And there was very much silo data. The data scientists are out in the lines of business right now, which is great, cause I think that's where data science belongs. We should give them on. And that's what we're working towards now is giving them more self service, giving them the ability to access the data, um, in a more robust way. And it's a single source of truth. So they're not pulling the data down into their own like tableau dashboards and then pushing the data back out. Um, so they're going to more not, I don't want to say a central repository, but a more of a robust repository that's controlled across multiple avenues where multiple lines of business can access. That said, how >>got it? Yes, and I think that one of the key things that I'm taking away from your last comment is the cultural aspects of this bite having the data. Scientists in the line of business, the line of lines of business, will feel ownership of that data as opposed to pointing fingers, criticizing the data quality they really own that that problem, as opposed to saying, Well, it's it's It's Paulus problem, >>right? Well, I have. My problem >>is, I have a date. Engineers, data architects, they database administrators, right, Um, and then data traditional data forwarding people. Um, and because some customers that I have that our business customers lines of business, they want to just subscribe to a report. They don't want to go out and do any data science work. Um, and we still have to provide that. So we still want to provide them some kind of regimen that they wake up in the morning and they open up their email. And there's the report that they just drive, um, which is great. And it works out really well. And one of the things is why we purchase I o waas. I would have the ability to give the lines of business the ability to do search within the data. And we read the data flows and data redundancy and things like that help me cleanup the data and also, um, to give it to the data. Analysts who say All right, they just asked me. They want this certain report, and it used to take Okay, well, we're gonna four weeks, we're going to go. We're gonna look at the data, and then we'll come back and tell you what we dio. But now with Iot Tahoe, they're able to look at the data and then, in one or two days of being able to go back and say, yes, we have data. This is where it is. This is where we found that this is the data flows that we've found also, which is that what I call it is the birth of a column. It's where the calm was created and where it went live as a teenager. And then it went to, you know, die very archive. Yeah, it's this, you know, cycle of life for a column. And Iot Tahoe helps us do that, and we do. Data lineage has done all the time. Um, and it's just takes a very long time. And that's why we're using something that has AI and machine learning. Um, it's it's accurate. It does it the same way over and over again. If an analyst leads, you're able to utilize talked something like, Oh, to be able to do that work for you. I get that. >>Yes. Oh, got it. So So a couple things there is in in, In researching Iot Tahoe, it seems like one of the strengths of their platform is the ability to visualize data the data structure and actually dig into it. But also see it, um, and that speeds things up and gives everybody additional confidence. And then the other pieces essentially infusing AI or machine intelligence into the data pipeline is really how you're attacking automation, right? And you're saying it's repeatable and and then that helps the data quality, and you have this virtuous cycle. Is there a firm that and add some color? Perhaps >>Exactly. Um, so you're able to let's say that I have I have seven cause lines of business that are asking me questions and one of the questions I'll ask me is. We want to know if this customer is okay to contact, right? And you know, there's different avenues, so you can go online to go. Do not contact me. You can go to the bank and you can say I don't want, um, email, but I'll take tests and I want, you know, phone calls. Um, all that information. So seven different lines of business asked me that question in different ways once said okay to contact the other one says, you know, customer one to pray All these, You know, um, and each project before I got there used to be siloed. So one customer would be 100 hours for them to do that and analytical work, and then another cut. Another analysts would do another 100 hours on the other project. Well, now I can do that all at once, and I can do those type of searches and say, Yes, we already have that documentation. Here it is. And this is where you can find where the customer has said, you know, you don't want I don't want to get access from you by email, or I've subscribed to get emails from you. >>Got it. Okay? Yeah. Okay. And then I want to come back to the cloud a little bit. So you you mentioned those three buckets? So you're moving to the Amazon cloud. At least I'm sure you're gonna get a hybrid situation there. You mentioned Snowflake. Um, you know what was sort of the decision to move to the cloud? Obviously, snowflake is cloud only. There's not an on Prem version there. So what precipitated that? >>Alright, So, from, um, I've been in >>the data I t Information field for the last 35 years. I started in the US Air Force and have moved on from since then. And, um, my experience with off brand waas with Snowflake was working with G McGee capital. And that's where I met up with the team from Iot to house as well. And so it's a proven. So there's a couple of things one is symptomatic of is worldwide. Now to move there, right, Two products, they have the on frame in the offering. Um, I've used the on Prem and off Prem. They're both great and it's very stable and I'm comfortable with other people are very comfortable with this. So we picked. That is our batch data movement. Um, we're moving to her, probably HBR. It's not a decision yet, but we're moving to HP are for real time data which has changed capture data, you know, moves it into the cloud. And then So you're envisioning this right now in Petrit, you're in the S three and you have all the data that you could possibly want. And that's Jason. All that everything is sitting in the S three to be able to move it through into snowflake and snowflake has proven cto have a stability. Um, you only need to learn in train your team with one thing. Um, aws has is completely stable at this 10.2. So all these avenues, if you think about it going through from, um, you know, this is your your data lake, which is I would consider your s three. And even though it's not a traditional data leg like you can touch it like a like a progressive or a dupe and into snowflake and then from snowflake into sandboxes. So your lines of business and your data scientists and just dive right in, Um, that makes a big, big win. and then using Iot. Ta ho! With the data automation and also their search engine, um, I have the ability to give the data scientists and eight analysts the the way of they don't need to talk to i t to get, um, accurate information or completely accurate information from the structure. And we'll be right there. >>Yes, so talking about, you know, snowflake and getting up to speed quickly. I know from talking to customers you get from zero to snowflake, you know, very fast. And then it sounds like the i o Ta ho is sort of the automation cloud for your data pipeline within the cloud. This is is that the right way to think about it? >>I think so. Um, right now I have I o ta >>ho attached to my >>on Prem. And, um, I >>want to attach it to my offering and eventually. So I'm using Iot Tahoe's data automation right now to bring in the data and to start analyzing the data close to make sure that I'm not missing anything and that I'm not bringing over redundant data. Um, the data warehouse that I'm working off is not a It's an on Prem. It's an Oracle database and its 15 years old. So it has extra data in it. It has, um, things that we don't need anymore. And Iot. Tahoe's helping me shake out that, um, extra data that does not need to be moved into my S three. So it's saving me money when I'm moving from offering on Prem. >>And so that was a challenge prior because you couldn't get the lines of business to agree what to delete or what was the issue there. >>Oh, it was more than that. Um, each line of business had their own structure within the warehouse, and then they were copying data between each other and duplicating the data and using that, uh so there might be that could be possibly three tables that have the same data in it. But it's used for different lines of business. And so I had we have identified using Iot Tahoe. I've identified over seven terabytes in the last, um, two months on data that is just been repetitive. Um, it just it's the same exact data just sitting in a different scheme. >>And and that's not >>easy to find. If you only understand one schema that's reporting for that line of business so that >>yeah, more bad news for the storage companies out there. Okay to follow. >>It's HCI. That's what that's what we were telling people you >>don't know and it's true, but you still would rather not waste it. You apply it to, you know, drive more revenue. And and so I guess Let's close on where you see this thing going again. I know you're sort of part way through the journey. May be you could sort of describe, you know, where you see the phase is going and really what you want to get out of this thing, You know, down the road Midterm. Longer term. What's your vision or your your data driven organization? >>Um, I want >>for the bankers to be able to walk around with on iPad in their hands and be able to access data for that customer really fast and be able to give them the best deal that they can get. I want Webster to be right there on top, with being able to add new customers and to be able to serve our existing customers who had bank accounts. Since you were 12 years old there and now our, you know, multi. Whatever. Um, I want them to be able to have the best experience with our our bankers, and >>that's awesome. I mean, that's really what I want is a banking customer. I want my bank to know who I am, anticipate my needs and create a great experience for me. And then let me go on with my life. And so that is a great story. Love your experience, your background and your knowledge. Can't thank you enough for coming on the Cube. >>No, thank you very much. And you guys have a great day. >>Alright, Take care. And thank you for watching everybody keep it right there. We'll take a short break and be right back. >>Yeah, yeah, yeah, yeah.
SUMMARY :
of enterprise data automation, an event Siri's brought to you by Iot. And I'm really excited to have Paul Damico here. Hi. Nice to see you, too. So let's let's start with Let's start with Webster Bank. awards for the area for being supportive for the community So you got a big responsibility as it relates to It's kind of transitioning to And then the other item is to drive new revenue Timely, accurate, complete data on the customer and what's really But I want to ask you about Cove. And part of that was is we adapted to Salesforce very, And then finally, you got more clarity. Um, from, you know, coming from the government and changed. I mean, a lot of people have sort of joked that many of the older people Um, the ability to give the customer what they a new a mortgage or looking to refinance or look, you know, whatever it iss, So you actually want the experience to be better. Um, you want you need a timely process so we can enhance Other other offers that you can you can make to the right customer, Um, and the only way we're going to be You see the potential to Prem and on France, you know, moving off Prem into like an s three bucket. about the way we do. quality engineers, you know, developers, etcetera, etcetera. Um, so they're going to more not, I don't want to say a central criticizing the data quality they really own that that problem, Well, I have. We're gonna look at the data, and then we'll come back and tell you what we dio. it seems like one of the strengths of their platform is the ability to visualize data the data structure and to contact the other one says, you know, customer one to pray All these, You know, So you you mentioned those three buckets? All that everything is sitting in the S three to be able to move it through I know from talking to customers you get from zero to snowflake, Um, right now I have I o ta Um, the data warehouse that I'm working off is And so that was a challenge prior because you couldn't get the lines Um, it just it's the same exact data just sitting If you only understand one schema that's reporting Okay to That's what that's what we were telling people you You apply it to, you know, drive more revenue. for the bankers to be able to walk around with on iPad And so that is a great story. And you guys have a great day. And thank you for watching everybody keep it right there.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Paul Damico | PERSON | 0.99+ |
Dave Volante | PERSON | 0.99+ |
Webster Bank | ORGANIZATION | 0.99+ |
Westchester | LOCATION | 0.99+ |
Paula D'Amico | PERSON | 0.99+ |
iPad | COMMERCIAL_ITEM | 0.99+ |
New York | LOCATION | 0.99+ |
one | QUANTITY | 0.99+ |
Connecticut | LOCATION | 0.99+ |
100 hours | QUANTITY | 0.99+ |
S three | COMMERCIAL_ITEM | 0.99+ |
15 years | QUANTITY | 0.99+ |
Jason | PERSON | 0.99+ |
France | LOCATION | 0.99+ |
Siri | TITLE | 0.99+ |
first item | QUANTITY | 0.99+ |
three tables | QUANTITY | 0.99+ |
24 hours | QUANTITY | 0.99+ |
thousands | QUANTITY | 0.99+ |
two months | QUANTITY | 0.99+ |
each line | QUANTITY | 0.99+ |
Fairfield County | LOCATION | 0.99+ |
HP | ORGANIZATION | 0.99+ |
Friday night | DATE | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
Two products | QUANTITY | 0.99+ |
Boston | LOCATION | 0.99+ |
four weeks | QUANTITY | 0.99+ |
US Air Force | ORGANIZATION | 0.98+ |
over 15 years | QUANTITY | 0.98+ |
two days | QUANTITY | 0.98+ |
New England | LOCATION | 0.98+ |
each project | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
Iot Tahoe | PERSON | 0.98+ |
Paulo | PERSON | 0.98+ |
Iot Tahoe | ORGANIZATION | 0.98+ |
both | QUANTITY | 0.97+ |
one thing | QUANTITY | 0.97+ |
first set | QUANTITY | 0.97+ |
TPP | TITLE | 0.97+ |
Paulus | PERSON | 0.97+ |
seven cause | QUANTITY | 0.97+ |
one schema | QUANTITY | 0.97+ |
one customer | QUANTITY | 0.96+ |
13 14 years | QUANTITY | 0.96+ |
over seven terabytes | QUANTITY | 0.96+ |
three | QUANTITY | 0.96+ |
single source | QUANTITY | 0.95+ |
Webster's Bay | ORGANIZATION | 0.95+ |
Webster | ORGANIZATION | 0.94+ |
seven different lines | QUANTITY | 0.94+ |
Cove | ORGANIZATION | 0.94+ |
Prem | ORGANIZATION | 0.93+ |
Enterprise Data Automation | ORGANIZATION | 0.92+ |
eight analysts | QUANTITY | 0.92+ |
10.2 | QUANTITY | 0.89+ |
12 years old | QUANTITY | 0.89+ |
Iot | ORGANIZATION | 0.88+ |
three buckets | QUANTITY | 0.88+ |
Snowflake | EVENT | 0.86+ |
last 35 years | DATE | 0.84+ |
Team Waas | ORGANIZATION | 0.8+ |
Io Tahoe | PERSON | 0.79+ |
24 7 development | QUANTITY | 0.72+ |
Salesforce | ORGANIZATION | 0.68+ |
each | QUANTITY | 0.68+ |
Amazon cloud | ORGANIZATION | 0.66+ |
Tahoe | PERSON | 0.66+ |
zero | QUANTITY | 0.64+ |
snowflake | EVENT | 0.61+ |
things | QUANTITY | 0.57+ |
Ajay Vohora, Io Tahoe | Enterprise Data Automation
>>from around the globe. It's the Cube with digital coverage of enterprise data automation an event Siri's brought to you by Iot. Tahoe. >>Okay, we're back. Welcome back to data Automated. A J ahora is CEO of I o Ta ho, JJ. Good to see you. How have things in London? >>Big thing. Well, thinking well, where we're making progress, I could see you hope you're doing well and pleasure being back here on the Cube. >>Yeah, it's always great to talk to. You were talking enterprise data automation. As you know, with within our community, we've been pounding the whole data ops conversation. Little different, though. We're gonna We're gonna dig into that a little bit. But let's start with a J how you've seen the response to Covert and I'm especially interested in the role that data has played in this pandemic. >>Yeah, absolutely. I think everyone's adapting both essentially, um, and and in business, the customers that I speak to on day in, day out that we partner with, um they're busy adapting their businesses to serve their customers. It's very much a game of and showing the week and serve our customers to help their customers um, you know, the adaptation that's happening here is, um, trying to be more agile, kind of the most flexible. Um, a lot of pressure on data. A lot of demand on data and to deliver more value to the business, too. Serve that customer. >>Yeah. I mean, data machine intelligence and cloud, or really three huge factors that have helped organizations in this pandemic. And, you know, the machine intelligence or AI piece? That's what automation is all about. How do you see automation helping organizations evolve maybe faster than they thought they might have to >>Sure. I think the necessity of these times, um, there's there's a says a lot of demand doing something with data data. Uh huh. A lot of a lot of businesses talk about being data driven. Um, so interesting. I sort of look behind that when we work with our customers, and it's all about the customer. You know, the mic is cios invested shareholders. The common theme here is the customer. That customer experience starts and ends with data being able to move from a point that is reacting. So what the customer is expecting and taking it to that step forward where you can be proactive to serve what that customer's expectation to and that's definitely come alive now with they, um, the current time. >>Yes. So, as I said, we've been talking about data ops a lot. The idea being Dev Ops applied to the data pipeline. But talk about enterprise data automation. What is it to you and how is it different from data off? >>Yeah, Great question. Thank you. I am. I think we're all familiar with felt more more awareness around. So as it's applied, Teoh, uh, processes methodologies that have become more mature of the past five years around devil that managing change, managing an application, life cycles, managing software development data about, you know, has been great. But breaking down those silos between different roles functions and bringing people together to collaborate. Andi, you know, we definitely see that those tools, those methodologies, those processes, that kind of thinking, um, landing itself to data with data is exciting. We're excited about that, Andi shifting the focus from being I t versus business users to you know who are the data producers. And here the data consumers in a lot of cases, it concert in many different lines of business. So in data role, those methods those tools and processes well we look to do is build on top of that with data automation. It's the is the nuts and bolts of the the algorithms, the models behind machine learning that the functions. That's where we investors our R and D and bringing that in to build on top of the the methods, the ways of thinking that break down those silos on injecting that automation into the business processes that are going to drive a business to serve its customers. It's, um, a layer beyond Dev ops data ops. They can get to that point where well, I think about it is, Is the automation behind the automation we can take? I'll give you an example. Okay, a bank where we did a lot of work to do make move them into accelerating that digital transformation. And what we're finding is that as we're able to automate the jobs related to data a managing that data and serving that data that's going into them as a business automating their processes for their customer. Um, so it's it's definitely having a compound effect. >>Yeah, I mean I think that you did. Data ops for a lot of people is somewhat new to the whole Dev Ops. The data ops thing is is good and it's a nice framework. Good methodology. There is obviously a level of automation in there and collaboration across different roles. But it sounds like you're talking about so supercharging it, if you will, the automation behind the automation. You know, I think organizations talk about being data driven. You hear that? They have thrown around a lot of times. People sit back and say, We don't make decisions without data. Okay? But really, being data driven is there's a lot of aspects there. There's cultural, but it's also putting data at the core of your organization, understanding how it effects monetization. And, as you know, well, silos have been built up, whether it's through M and a, you know, data sprawl outside data sources. So I'm interested in your thoughts on what data driven means and specifically Hi, how Iot Tahoe plays >>there. Yeah, I'm sure we'll be happy. That look that three David, we've We've come a long way in the last four years. We started out with automating some of those simple, um, to codify. Um, I have a high impact on organization across the data, a data warehouse. There's data related tasks that classify data on and a lot of our original pattern. Senai people value that were built up is is very much around. They're automating, classifying data across different sources and then going out to so that for some purpose originally, you know, some of those simpler I'm challenges that we have. Ah, custom itself, um, around data privacy. You know, I've got a huge data lake here. I'm a telecoms business. I've got millions of six subscribers. Um, quite often the chief data office challenges. How do I cover the operational risk? Where, um, I got so much data I need to simplify my approach to automating, classifying that data. Recent is you can't do that manually. We can for people at it. And the the scale of that is is prohibitive, right? Often, if you had to do it manually by the time you got a good picture of it, it's already out of date. Then, starting with those those simple challenges that we've been able to address, we're then going on and build on that to say, What else do we serve? What else do we serve? The chief data officer, Chief marketing officer on the CFO. Within these times, um, where those decision makers are looking for having a lot of choices in the platform options that they say that the tooling they're very much looking for We're that Swiss army. Not being able to do one thing really well is is great, but more more. Where that cost pressure challenge is coming in is about how do we, um, offer more across the organization, bring in those business lines of business activities that depend on data to not just with a T. Okay, >>so we like the cube. Sometimes we like to talk about Okay, what is it? And then how does it work? And what's the business impact? We kind of covered what it is but love to get into the tech a little bit in terms of how it works. And I think we have a graphic here that gets into that a little bit. So, guys, if you bring that up, I wonder if you could tell us and what is the secret sauce behind Iot Tahoe? And if you could take us through this slot. >>Sure. I mean, right there in the middle that the heart of what we do It is the intellectual property. Yeah, that was built up over time. That takes from Petra genius data sources Your Oracle relational database, your your mainframe. If they lay in increasingly AP eyes and devices that produce data and that creates the ability to automatically discover that data, classify that data after it's classified them have the ability to form relationships across those different, uh, source systems, silos, different lines of business. And once we've automated that that we can start to do some cool things that just puts a contact and meaning around that data. So it's moving it now from bringing data driven on increasingly well. We have really smile, right people in our customer organizations you want do some of those advanced knowledge tasks, data scientists and, uh, quants in some of the banks that we work with. The the onus is on, then, putting everything we've done there with automation, pacifying it, relationship, understanding that equality policies that you apply to that data. I'm putting it in context once you've got the ability to power. A a professional is using data, um, to be able to put that data and contacts and search across the entire enterprise estate. Then then they can start to do some exciting things and piece together the tapestry that fabric across that different systems could be crm air P system such as s AP on some of the newer cloud databases that we work with. Snowflake is a great Well, >>yes. So this is you're describing sort of one of the one of the reasons why there's so many stove pipes and organizations because data is gonna locked in the silos of applications. I also want to point out, you know, previously to do discovery to do that classification that you talked about form those relationship to glean context from data. A lot of that, if not most of that in some cases all that would have been manual. And of course, it's out of date so quickly. Nobody wants to do it because it's so hard. So this again is where automation comes into the the the to the idea of really becoming data driven. >>Sure. I mean the the efforts. If we if I look back, maybe five years ago, we had a prevalence of daily technologies at the cutting edge. Those have said converging me to some of these cloud platforms. So we work with Google and AWS, and I think very much is, as you said it, those manual attempts to try and grasp. But it is such a complex challenge at scale. I quickly runs out of steam because once, um, once you've got your hat, once you've got your fingers on the details Oh, um, what's what's in your data estate? It's changed, you know, you've onboard a new customer. You signed up a new partner, Um, customer has no adopted a new product that you just Lawrence and there that that slew of data it's keeps coming. So it's keeping pace with that. The only answer really is is some form of automation. And what we found is if we can tie automation with what I said before the expertise the, um, the subject matter expertise that sometimes goes back many years within an organization's people that augmentation between machine learning ai on and on that knowledge that sits within inside the organization really tends to involve a lot of value in data? >>Yes, So you know Well, a J you can't be is a smaller company, all things to all people. So your ecosystem is critical. You working with AWS? You're working with Google. You got red hat. IBM is as partners. What is attracting those folks to your ecosystem and give us your thoughts on the importance of ecosystem? >>Yeah, that's that's fundamental. So I mean, when I caimans, we tell her here is the CEO of one of the, um, trends that I wanted us to to be part of was being open, having an open architecture that allowed one thing that was nice to my heart, which is as a CEO, um, a C I O where you've got a budget vision and you've already made investments into your organization, and some of those are pretty long term bets. They should be going out 5 10 years, sometimes with CRM system training up your people, getting everybody working together around a common business platform. What I wanted to ensure is that we could openly like it using ap eyes that were available, the love that some investment on the cost that has already gone into managing in organizations I t. But business users to before So part of the reason why we've been able to be successful with, um, the partners like Google AWS and increasingly, a number of technology players. That red hat mongo DB is another one where we're doing a lot of good work with, um, and snowflake here is, um it's those investments have been made by the organizations that are our customers, and we want to make sure we're adding to that, and they're leveraging the value that they've already committed to. >>Okay, so we've talked about kind of what it is and how it works, and I want to get into the business impact. I would say what I would be looking for from from this would be Can you help me lower my operational risk? I've got I've got tasks that I do many year sequential, some who are in parallel. But can you reduce my time to task? And can you help me reduce the labor intensity and ultimately, my labor costs? And I put those resources elsewhere, and ultimately, I want to reduce the end and cycle time because that is going to drive Telephone number R. A. Y So, um, I missing anything? Can you do those things? And maybe you could give us some examples of the tiara y and the business impact. >>Yeah. I mean, the r a y David is is built upon on three things that I mentioned is a combination off leveraging the existing investment with the existing state, whether that's home, Microsoft, Azure or AWS or Google IBM. And I'm putting that to work because, yeah, the customers that we work with have had made those choices. On top of that, it's, um, is ensuring that we have you got the automation that is working right down to the level off data, a column level or the file level so we don't do with meta data. It is being very specific to be at the most granular level. So as we've grown our processes and on the automation, gasification tagging, applying policies from across different compliance and regulatory needs, that an organization has to the data, everything that then happens downstream from that is ready to serve a business outcome. It could be a customer who wants that experience on a mobile device. A tablet oh, face to face within, within the store. I mean game. Would you provision the right data and enable our customers do that? But their customers, with the right data that they can trust at the right time, just in that real time moment where decision or an action is being expected? That's, um, that's driving the r a y two b in some cases, 20 x but and that's that's really satisfying to see that that kind of impact it is taking years down to months and in many cases, months of work down to days. In some cases, our is the time to value. I'm I'm impressed with how quickly out of the box with very little training a customer and think about, too. And you speak just such a search. They discovery knowledge graph on DM. I don't find duplicates. Onda Redundant data right off the bat within hours. >>Well, it's why investors are interested in this space. I mean, they're looking for a big, total available market. They're looking for a significant return. 10 X is you gotta have 10 x 20 x is better. So so that's exciting and obviously strong management and a strong team. I want to ask you about people and culture. So you got people process technology we've seen with this pandemic that processes you know are really unpredictable. And the technology has to be able to adapt to any process, not the reverse. You can't force your process into some static software, so that's very, very important. But the end of the day you got to get people on board. So I wonder if you could talk about this notion of culture and a data driven culture. >>Yeah, that's that's so important. I mean, current times is forcing the necessity of the moment to adapt. But as we start to work their way through these changes on adapt ah, what with our customers, But that is changing economic times. What? What we're saying here is the ability >>to I >>have, um, the technology Cartman, in a really smart way, what those business uses an I T knowledge workers are looking to achieve together. So I'll give you an example. We have quite often with the data operations teams in the companies that we, um, partnering with, um, I have a lot of inbound enquiries on the day to day level. I really need this set of data they think it can help my data scientists run a particular model? Or that what would happen if we combine these two different silence of data and gets the Richmond going now, those requests you can, sometimes weeks to to realize what we've been able to do with the power is to get those answers being addressed by the business users themselves. And now, without without customers, they're coming to the data. And I t folks saying, Hey, I've now built something in the development environment. Why don't we see how that can scale up with these sets of data? I don't need terabytes of it. I know exactly the columns and the feet in the data that I'm going to use on that gets seller wasted in time, um, angle to innovate. >>Well, that's huge. I mean, the whole notion of self service and the lines of business actually feeling like they have ownership of the data as opposed to, you know, I t or some technology group owning the data because then you've got data quality issues or if it doesn't line up there their agenda, you're gonna get a lot of finger pointing. So so that is a really important. You know a piece of it. I'll give you last word A J. Your final thoughts, if you would. >>Yeah, we're excited to be the only path. And I think we've built great customer examples here where we're having a real impact in in a really fast pace, whether it helping them migrate to the cloud, helping the bean up their legacy, Data lake on and write off there. Now the conversation is around data quality as more of the applications that we enable to a more efficiently could be data are be a very robotic process automation along the AP, eyes that are now available in the cloud platforms. A lot of those they're dependent on data quality on and being able to automate. So business users, um, to take accountability off being able to so look at the trend of their data quality over time and get the signals is is really driving trust. And that trust in data is helping in time. Um, the I T teams, the data operations team, with do more and more quickly that comes back to culture being out, supply this technology in such a way that it's visual insensitive. Andi. How being? Just like Dev Ops tests with with a tty Dave drops putting intelligence in at the data level to drive that collaboration. We're excited, >>you know? You remind me of something. I lied. I don't want to go yet. It's OK, so I know we're tight on time, but you mentioned migration to the cloud. And I'm thinking about conversation with Paula from Webster Webster. Bank migrations. Migrations are, you know, they're they're a nasty word for for organizations. So our and we saw this with Webster. How are you able to help minimize the migration pain and and why is that something that you guys are good at? >>Yeah. I mean, there were many large, successful companies that we've worked with. What's There's a great example where, you know, I'd like to give you the analogy where, um, you've got a lot of people in your teams if you're running a business as a CEO on this bit like a living living grade. But imagine if those different parts of your brain we're not connected, that with, um, so diminish how you're able to perform. So what we're seeing, particularly with migration, is where banks retailers. Manufacturers have grown over the last 10 years through acquisition on through different initiatives, too. Um, drive customer value that sprawl in their data estate hasn't been fully dealt with. It sometimes been a good thing, too. Leave whatever you're fired off the agent incent you a side by side with that legacy mainframe on your oracle, happy and what we're able to do very quickly with that migration challenges shine a light on all the different parts. Oh, data application at the column level or higher level if it's a day late and show an enterprise architect a CDO how everything's connected, where they may not be any documentation. The bright people that created some of those systems long since moved on or retired or been promoted into so in the rose on within days, being out to automatically generate Anke refreshed the states of that data across that man's game on and put it into context, then allows you to look at a migration from a confidence that you did it with the back rather than what we've often seen in the past is teams of consultant and business analysts. Data around this spend months getting an approximation and and a good idea of what it could be in the current state and try their very best to map that to the future Target state. Now, without all hoping out, run those processes within hours of getting started on, um well, that picture visualize that picture and bring it to life. You know, the Yarra. Why, that's off the bat with finding data that should have been deleted data that was copies off on and being able to allow the architect whether it's we're working on gcb or migration to any other clouds such as AWS or a multi cloud landscape right now with yeah, >>that visibility is key. Teoh sort of reducing operational risks, giving people confidence that they can move forward and being able to do that and update that on an ongoing basis, that means you can scale a J. Thanks so much for coming on the Cube and sharing your insights and your experience is great to have >>you. Thank you, David. Look towards smoking in. >>Alright, keep it right there, everybody. We're here with data automated on the Cube. This is Dave Volante and we'll be right back. Short break. >>Yeah, yeah, yeah, yeah
SUMMARY :
enterprise data automation an event Siri's brought to you by Iot. Good to see you. Well, thinking well, where we're making progress, I could see you hope As you know, with within A lot of demand on data and to deliver more value And, you know, the machine intelligence I sort of look behind that What is it to you that automation into the business processes that are going to drive at the core of your organization, understanding how it effects monetization. that for some purpose originally, you know, some of those simpler I'm challenges And if you could take us through this slot. produce data and that creates the ability to that you talked about form those relationship to glean context from data. customer has no adopted a new product that you just Lawrence those folks to your ecosystem and give us your thoughts on the importance of ecosystem? that are our customers, and we want to make sure we're adding to that, that is going to drive Telephone number R. A. Y So, um, And I'm putting that to work because, yeah, the customers that we work But the end of the day you got to get people on board. necessity of the moment to adapt. I have a lot of inbound enquiries on the day to day level. of the data as opposed to, you know, I t or some technology group owning the data intelligence in at the data level to drive that collaboration. is that something that you guys are good at? I'd like to give you the analogy where, um, you've got a lot of people giving people confidence that they can move forward and being able to do that and update We're here with data automated on the Cube.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
David | PERSON | 0.99+ |
Paula | PERSON | 0.99+ |
Ajay Vohora | PERSON | 0.99+ |
ORGANIZATION | 0.99+ | |
IBM | ORGANIZATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Dave Volante | PERSON | 0.99+ |
millions | QUANTITY | 0.99+ |
Siri | TITLE | 0.99+ |
Webster | ORGANIZATION | 0.99+ |
London | LOCATION | 0.99+ |
Iot Tahoe | ORGANIZATION | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
Io Tahoe | PERSON | 0.99+ |
10 | QUANTITY | 0.99+ |
five years ago | DATE | 0.98+ |
Onda | ORGANIZATION | 0.98+ |
Webster Webster | ORGANIZATION | 0.98+ |
Covert | PERSON | 0.97+ |
two | QUANTITY | 0.97+ |
both | QUANTITY | 0.97+ |
5 10 years | QUANTITY | 0.97+ |
three | QUANTITY | 0.96+ |
20 x | QUANTITY | 0.94+ |
10 X | QUANTITY | 0.94+ |
Cube | COMMERCIAL_ITEM | 0.93+ |
Andi | PERSON | 0.93+ |
one | QUANTITY | 0.93+ |
Azure | ORGANIZATION | 0.92+ |
six subscribers | QUANTITY | 0.91+ |
three things | QUANTITY | 0.91+ |
I o Ta ho | ORGANIZATION | 0.91+ |
Google AWS | ORGANIZATION | 0.91+ |
Yarra | ORGANIZATION | 0.89+ |
J ahora | PERSON | 0.89+ |
Anke | ORGANIZATION | 0.89+ |
Dave | PERSON | 0.85+ |
Iot Tahoe | PERSON | 0.84+ |
a day | QUANTITY | 0.82+ |
Lawrence | PERSON | 0.82+ |
one thing | QUANTITY | 0.81+ |
Petra | PERSON | 0.78+ |
pandemic | EVENT | 0.78+ |
Iot. Tahoe | PERSON | 0.78+ |
last four years | DATE | 0.78+ |
past five years | DATE | 0.77+ |
Swiss | ORGANIZATION | 0.76+ |
JJ | PERSON | 0.75+ |
Enterprise Data Automation | ORGANIZATION | 0.73+ |
last 10 years | DATE | 0.62+ |
Dev Ops | ORGANIZATION | 0.59+ |
Richmond | ORGANIZATION | 0.55+ |
Cartman | ORGANIZATION | 0.55+ |
Snowflake | EVENT | 0.51+ |
terabytes | QUANTITY | 0.5+ |
factors | QUANTITY | 0.46+ |
data | TITLE | 0.45+ |
Lester Waters, Io Tahoe | Enterprise Data Automation
(upbeat music) >> Reporter: From around the globe, it's The Cube with digital coverage of enterprise data automation and event series brought to you by Io-Tahoe. >> Okay, we're back. Focusing on enterprise data automation, we're going to talk about the journey to the cloud. Remember, the hashtag is data automated. We're here with Lester Waters who's the CTO of Io-Tahoe, Lester, good to see you from across the pond on video, wish we were face to face, but it's great to have you on The Cube. >> Also I do, thank you for having me. >> Oh, you're very welcome. Hey, give us a little background on CTO, you got a deep expertise in a lot of different areas, but what do we need to know? >> Well, David, I started my career basically at Microsoft, where I started the Information Security Cryptography Group. They're the very first one that the company had and that led to a career in information security and of course, as you go along with the information security, data is the key element to be protected. So I always had my hands in data and that naturally progressed into a role with Io-Tahoe as their CTO. >> Guys, I have to invite you back, we'll talk crypto all day we'd love to do that but we're here talking about yeah, awesome, right? But we're here talking about the cloud and here we'll talk about the journey to the cloud and accelerate. Everybody's really interested obviously in cloud, even more interested now with the pandemic, but what's that all about? >> Well, moving to the cloud is quite an undertaking for most organizations. First of all, we've got as probably if you're a large enterprise, you probably have thousands of applications, you have hundreds and hundreds of database instances, and trying to shed some light on that, just to plan your move to the cloud is a real challenge. And some organizations try to tackle that manually. Really what Io-Tahoe is bringing is trying to tackle that in an automated version to help you with your journey to the cloud. >> Well, look at migrations are sometimes just an evil word to a lot of organizations, but at the same time, building up technical debt veneer after veneer and year, and year, and year is something that many companies are saying, "Okay, it's got to stop." So what's the prescription for that automation journey and simplifying that migration to the cloud? >> Well, I think the very first thing that's all about is data hygiene. You don't want to pick up your bad habits and take them to the cloud. You've got an opportunity here, so I see the journey to the cloud is an opportunity to really clean house, reorganize things, like moving out. You might move all your boxes, but you're kind of probably cherry pick what you're going to take with you and then you're going to organize it as you end up at your new destination. So from that, I get there's seven key principles that I like to operate by when I advise on the cloud migration. >> Okay. So, where do you start? >> Well, I think the first thing is understanding what you got, so discover and cataloging your data and your applications. If I don't know what I have, I can't move it, I can't improve it, I can't build up on it. And I have to understand there is dependency, so building that data catalog is the very first step. What do I got? >> Now, is that a metadata exercise? Sometimes there's more metadata than there is data. Is metadata part of that first step or? >> In deed, metadata is the first step so the metadata really describes the data you have. So, the metadata is going to tell me I have 2000 tables and maybe of those tables, there's an average of 25 columns each, and so that gives me a sketch if you will, of what I need to move. How big are the boxes I need to pack for my move to the cloud? >> Okay, and you're saying you can automate that data classification, categorization, discovery, correct using math machine intelligence, is that correct? >> Yeah, that's correct. So basically we go, and we will discover all of the schema, if you will, that's the metadata description of your tables and columns in your database in the data types. So we take, we will ingest that in, and we will build some insights around that. And we do that across a variety of platforms because everybody's organization has you've got a one yeah, an Oracle Database here, and you've got a Microsoft SQL Database here, you might have something else there that you need to bring site onto. And part of this journey is going to be about breaking down your data silos and understanding what you've got. >> Okay. So, we've done the audit, we know what we've got, what's next? Where do we go next? >> So the next thing is remediating that data. Where do I have duplicate data? Often times in an organization, data will get duplicated. So, somebody will take a snapshot of a data, and then ended up building a new application, which suddenly becomes dependent on that data. So it's not uncommon for an organization of 20 master instances of a customer. And you can see where that will go when trying to keep all that stuff in sync becomes a nightmare all by itself. So you want to understand where all your redundant data is. So when you go to the cloud, maybe you have an opportunity here to consolidate that data. >> Yeah, because you like to borrow in an Einstein or apply an Einstein Bromide right. Keep as much data as you can, but no more. >> Correct. >> Okay. So you get to the point to the second step you're kind of a one to reduce costs, then what? You figure out what to get rid of, or actually get rid of it, what's next? >> Yes, that would be the next step. So figuring out what you need and what you don't need often times I've found that there's obsolete columns of data in your databases that you just don't need, or maybe it's been superseded by another, you've got tables that have been superseded by other tables in your database. So you got to understand what's being used and what's not and then from that, you can decide, "I'm going to leave this stuff behind, "or I'm going to archive this stuff "cause I might need it for data retention "or I'm just going to delete it, "I don't need it at all." >> Well, Lester, most organizations, if they've been around a while, and the so-called incumbents, they've got data all over the place, their data marts, data warehouses, there are all kinds of different systems and the data lives in silos. So, how do you kind of deal with that problem? Is that part of the journey? >> That's a great point Dave, because you're right that the data silos happen because this business unit is chartered with this task another business unit has this task and that's how you get those instantiations of the same data occurring in multiple places. So as part of your cloud migration journey, you really want to plan where there's an opportunity to consolidate your data, because that means there'll be less to manage, there'll be less data to secure, and it'll have a smaller footprint, which means reduced costs. >> So, people always talk about a single version of the truth, data quality is a huge issue. I've talked to data practitioners and they've indicated that the quality metrics are in the single digits and they're trying to get to 90% plus, but maybe you could address data quality. Where does that fit in on the journey? >> That's, a very important point. First of all, you don't want to bring your legacy issues with you. As the point I made earlier, if you've got data quality issues, this is a good time to find those and identify and remediate them. But that can be a laborious task. We've had customers that have tried to do this by hand and it's very, very time consuming, cause you imagine if you've got 200 tables, 50,000 columns, imagine, the manual labor involved in doing that. And you could probably accomplish it, but it'll take a lot of work. So the opportunity to use tools here and automate that process is really will help you find those outliers there's that bad data and correct it before you move to the cloud. >> And you're just talking about that automation it's the same thing with data catalog and that one of the earlier steps. Organizations would do this manually or they try to do it manually and that's a lot of reason for the failure. They just, it's like cleaning out your data like you just don't want to do it (laughs). Okay, so then what's next? I think we're plowing through your steps here. What what's next on the journey? >> The next one is, in a nutshell, preserve your data format. Don't boil the ocean here to use a cliche. You want to do a certain degree of lift and shift because you've got application dependencies on that data and the data format, the tables on which they sit, the columns and the way they're named. So, some degree you are going to be doing a lift and shift, but it's an intelligent lift and shift using all the insights you've gathered by cataloging the data, looking for data quality issues, looking for duplicate columns, doing planning consolidation. You don't want to also rewrite your application. So, in that aspect, I think it's important to do a bit of lift and shift and preserve those data formats as they sit. >> Okay, so let me follow up on that. That sounds really important to me, because if you're doing a conversion and you're rewriting applications, that means that you're going to have to freeze the existing application, and then you going to be refueling the plane as you're in midair and a lot of times, especially with mission critical systems, you're never going to bring those together and that's a recipe for disaster, isn't it? >> Great analogy unless you're with the air force, you'll (mumbles) (laughs). Now, that's correct. It's you want to have bite-sized steps and that's why it's important to plan your journey, take these steps. You're using automation where you can to make that journey to the cloud much easier and more straightforward. >> All right, I like that. So we're taking a kind of a systems view and end to end view of the data pipeline, if you will. What's next? I think we're through. I think I've counted six. What's the lucky seven? >> Lucky seven, involve your business users. Really, when you think about it, your data is in silos. Part of this migration to the cloud is an opportunity to break down these silos, these silos that naturally occur as part of the business unit. You've got to break these cultural barriers that sometimes exist between business and say, so for example, I always advise, there's an opportunity here to consolidate your sensitive data, your PII, your personally identifiable information, and if three different business units have the same source of truth for that, there's was an opportunity to consolidate that into one as you migrate. That might be a little bit of tweaking to some of the apps that you have that are dependent on it, but in the long run, that's what you really want to do. You want to have a single source of truth, you want to ring fence that sensitive data, and you want all your business users talking together so that you're not reinventing the wheel. >> Well, the reason I think too that's so important is that you're now I would say you're creating a data driven culture. I know that's sort of a buzz word, but what it's true and what that means to me is that your users, your lines of business feel like they actually own the data rather than pointing fingers at the data group, the IT group, the data quality people, data engineers, saying, "Oh, I don't believe it." If the lines of business own the data, they're going to lean in, they're going to maybe bring their own data science resources to the table, and it's going to be a much more collaborative effort as opposed to a non-productive argument. >> Yeah. And that's where we want to get to. DataOps is key, and maybe that's a term that's still evolving. But really, you want the data to drive the business because that's where your insights are, that's where your value is. You want to break down the silos between not only the business units, as I mentioned, but also as you pointed out, the roles of the people that are working with it. A self service data culture is the right way to go with the right security controls, putting on my security hat of course in place so that if I'm a developer and I'm building a new application, I'd love to be able to go to the data catalog, "Oh, there's already a database that has the customer "what the customers have clicked on when shopping." I could use that. I don't have to rebuild that, I'll just use that as for my application. That's the kind of problems you want to be able to solve and that's where your cost reductions come in across the board. >> Yeah. I want to talk a little bit about the business context here. We always talk about data, it's the new source of competitive advantage, I think there's not a lot of debate about that, but it's hard. A lot of companies are struggling to get value out of their data because it's so difficult. All the things we've talked about, the silos, the data quality, et cetera. So, you mentioned the term data apps, data apps is all about streamlining, that data, pipelining, infusing automation and machine intelligence into that pipeline and then ultimately taking a systems view and compressing that time to insights so that you can drive monetization, whether it's cut costs, maybe it's new revenue, drive productivity, but it's that end to end cycle time reduction that successful practitioners talk about as having the biggest business impact. Are you seeing that? >> Absolutely, but it is a journey and it's a huge cultural change for some companies that are. I've worked in many companies that are ticket based IT-driven and just do even the marginalist of change or get insight, raise a ticket, wait a week and then out the other end will pop maybe a change that I needed and it'll take a while for us to get to a culture that truly has a self service data-driven nature where I'm the business owner, and I want to bring in a data scientist because we're losing. For example, a business might be losing to a competitor and they want to find what insights, why is the customer churn, for example, happening every Tuesday? What is it about Tuesday? This is where your data scientist comes in. The last thing you want is to raise a ticket, wait for the snapshot of the data, you want to enable that data scientist to come in, securely connect into the data, and do his analysis, and come back and give you those insights, which will give you that competitive advantage. >> Well, I love your point about churn, maybe it talks about the Andreessen quote that "Software's eating the world," and all companies are our software companies, and SaaS companies, and churn is the killer of SaaS companies. So very, very important point you're making. My last question for you before we summarize is the tech behind all of these. What makes Io-Tahoe unique in its ability to help automate that data pipeline? >> Well, we've done a lot of research, we have I think now maybe 11 pending patent applications, I think one has been approved to be issued (mumbles), but really, it's really about sitting down and doing the right kind of analysis and figuring out how we can optimize this journey. Some of these stuff isn't rocket science. You can read a schema and into an open source solution, but you can't necessarily find the hidden insights. So if I want to find my foreign key dependencies, which aren't always declared in the database, or I want to identify columns by their content, which because the columns might be labeled attribute one, attribute two, attribute three, or I want to find out how my data flows between the various tables in my database. That's the point at which you need to bring in automation, you need to bring in data science solutions, and there's even a degree of machine learning because for example, we might deduce that data is flowing from this table to this table and upon when you present that to the user with a 87% confidence, for example, and the user can go, or the administrator can go. Now, it really goes the other way, it was an invalid collusion and that's the machine learning cycle. So the next time we see that pattern again, in that environment we will be able to make a better recommendation because some things aren't black and white, they need that human intervention loop. >> All right, I just want to summarize with Lester Waters' playbook to moving to the cloud and I'll go through them. Hopefully, I took some notes, hopefully, I got them right. So step one, you want to do that data discovery audit, you want to be fact-based. Two is you want to remediate that data redundancy, and then three identify what you can get rid of. Oftentimes you don't get rid of stuff in IT, or maybe archive it to cheaper media. Four is consolidate those data silos, which is critical, breaking down those data barriers. And then, five is attack the quality issues before you do the migration. Six, which I thought was really intriguing was preserve that data format, you don't want to do the rewrite applications and do that conversion. It's okay to do a little bit of lifting and shifting >> This comes in after the task. >> Yeah, and then finally, and probably the most important is you got to have that relationship with the lines of business, your users, get them involved, begin that cultural shift. So I think great recipe Lester for safe cloud migration. I really appreciate your time. I'll give you the final word if you will bring us home. >> All right. Well, I think the journey to the cloud it's a tough one. You will save money, I have heard people say, you got to the cloud, it's too expensive, it's too this, too that, but really, there is an opportunity for savings. I'll tell you when I run data services as a PaaS service in the cloud, it's wonderful because I can scale up and scale down almost by virtually turning a knob. And so I'll have complete control and visibility of my costs. And so for me, that's very important. Io also, it gives me the opportunity to really ring fence my sensitive data, because let's face it, most organizations like being in a cheese grater when you talk about security, because there's so many ways in and out. So I find that by consolidating and bringing together the crown jewels, if you will. As a security practitioner, it's much more easy to control. But it's very important. You can't get there without some automation and automating this discovery and analysis process. >> Well, great advice. Lester, thanks so much. It's clear that the capex investments on data centers are generally not a good investment for most companies. Lester, really appreciate, Lester waters CTO of Io-Tahoe. Let's watch this short video and we'll come right back. You're watching The Cube, thank you. (upbeat music)
SUMMARY :
to you by Io-Tahoe. but it's great to have you on The Cube. you got a deep expertise in and that led to a career Guys, I have to invite you back, to help you with your and simplifying that so I see the journey to is the very first step. Now, is that a metadata exercise? and so that gives me a sketch if you will, that you need to bring site onto. we know what we've got, what's next? So you want to understand where Yeah, because you like point to the second step and then from that, you can decide, and the data lives in silos. and that's how you get Where does that fit in on the journey? So the opportunity to use tools here and that one of the earlier steps. and the data format, the and then you going to to plan your journey, and end to end view of the and you want all your business and it's going to be a much database that has the customer and compressing that time to insights and just do even the marginalist of change and churn is the killer That's the point at which you and do that conversion. after the task. and probably the most important is the journey to the cloud It's clear that the capex
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
David | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
200 tables | QUANTITY | 0.99+ |
hundreds | QUANTITY | 0.99+ |
90% | QUANTITY | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Lester Waters | PERSON | 0.99+ |
six | QUANTITY | 0.99+ |
first step | QUANTITY | 0.99+ |
87% | QUANTITY | 0.99+ |
Information Security Cryptography Group | ORGANIZATION | 0.99+ |
25 columns | QUANTITY | 0.99+ |
Io-Tahoe | ORGANIZATION | 0.99+ |
seven key principles | QUANTITY | 0.99+ |
2000 tables | QUANTITY | 0.99+ |
Andreessen | PERSON | 0.99+ |
Six | QUANTITY | 0.99+ |
second step | QUANTITY | 0.99+ |
Io Tahoe | PERSON | 0.99+ |
Tuesday | DATE | 0.99+ |
50,000 columns | QUANTITY | 0.99+ |
Lester | PERSON | 0.98+ |
11 pending patent applications | QUANTITY | 0.98+ |
five | QUANTITY | 0.97+ |
a week | QUANTITY | 0.97+ |
20 master instances | QUANTITY | 0.97+ |
Einstein | PERSON | 0.97+ |
Four | QUANTITY | 0.97+ |
first one | QUANTITY | 0.96+ |
one | QUANTITY | 0.96+ |
first thing | QUANTITY | 0.95+ |
Lester | ORGANIZATION | 0.95+ |
Two | QUANTITY | 0.93+ |
First | QUANTITY | 0.93+ |
Enterprise Data Automation | ORGANIZATION | 0.93+ |
three | QUANTITY | 0.93+ |
seven | QUANTITY | 0.92+ |
step one | QUANTITY | 0.92+ |
single version | QUANTITY | 0.92+ |
pandemic | EVENT | 0.91+ |
SQL Database | TITLE | 0.91+ |
single source | QUANTITY | 0.86+ |
three different business units | QUANTITY | 0.82+ |
The Cube | ORGANIZATION | 0.8+ |
Oracle Database | TITLE | 0.79+ |
thousands of applications | QUANTITY | 0.76+ |
single digits | QUANTITY | 0.76+ |
capex | ORGANIZATION | 0.74+ |
CTO | PERSON | 0.73+ |
Waters' | PERSON | 0.69+ |
each | QUANTITY | 0.68+ |
attribute two | OTHER | 0.65+ |
attribute three | OTHER | 0.59+ |
The Cube | TITLE | 0.57+ |
attribute one | OTHER | 0.44+ |
Paula D'Amico, Webster Bank | Io Tahoe | Enterprise Data Automation
>> Narrator: From around the Globe, it's theCube with digital coverage of Enterprise Data Automation, and event series brought to you by Io-Tahoe. >> Everybody, we're back. And this is Dave Vellante, and we're covering the whole notion of Automated Data in the Enterprise. And I'm really excited to have Paula D'Amico here. Senior Vice President of Enterprise Data Architecture at Webster Bank. Paula, good to see you. Thanks for coming on. >> Hi, nice to see you, too. >> Let's start with Webster bank. You guys are kind of a regional I think New York, New England, believe it's headquartered out of Connecticut. But tell us a little bit about the bank. >> Webster bank is regional Boston, Connecticut, and New York. Very focused on in Westchester and Fairfield County. They are a really highly rated regional bank for this area. They hold quite a few awards for the area for being supportive for the community, and are really moving forward technology wise, they really want to be a data driven bank, and they want to move into a more robust group. >> We got a lot to talk about. So data driven is an interesting topic and your role as Data Architecture, is really Senior Vice President Data Architecture. So you got a big responsibility as it relates to kind of transitioning to this digital data driven bank but tell us a little bit about your role in your Organization. >> Currently, today, we have a small group that is just working toward moving into a more futuristic, more data driven data warehousing. That's our first item. And then the other item is to drive new revenue by anticipating what customers do, when they go to the bank or when they log in to their account, to be able to give them the best offer. And the only way to do that is you have timely, accurate, complete data on the customer and what's really a great value on offer something to offer that, or a new product, or to help them continue to grow their savings, or do and grow their investments. >> Okay, and I really want to get into that. But before we do, and I know you're, sort of partway through your journey, you got a lot to do. But I want to ask you about Covid, how you guys handling that? You had the government coming down and small business loans and PPP, and huge volume of business and sort of data was at the heart of that. How did you manage through that? >> We were extremely successful, because we have a big, dedicated team that understands where their data is and was able to switch much faster than a larger bank, to be able to offer the PPP Long's out to our customers within lightning speed. And part of that was is we adapted to Salesforce very for we've had Salesforce in house for over 15 years. Pretty much that was the driving vehicle to get our PPP loans in, and then developing logic quickly, but it was a 24 seven development role and get the data moving on helping our customers fill out the forms. And a lot of that was manual, but it was a large community effort. >> Think about that too. The volume was probably much higher than the volume of loans to small businesses that you're used to granting and then also the initial guidelines were very opaque. You really didn't know what the rules were, but you were expected to enforce them. And then finally, you got more clarity. So you had to essentially code that logic into the system in real time. >> I wasn't directly involved, but part of my data movement team was, and we had to change the logic overnight. So it was on a Friday night it was released, we pushed our first set of loans through, and then the logic changed from coming from the government, it changed and we had to redevelop our data movement pieces again, and we design them and send them back through. So it was definitely kind of scary, but we were completely successful. We hit a very high peak. Again, I don't know the exact number but it was in the thousands of loans, from little loans to very large loans and not one customer who applied did not get what they needed for, that was the right process and filled out the right amount. >> Well, that is an amazing story and really great support for the region, your Connecticut, the Boston area. So that's fantastic. I want to get into the rest of your story now. Let's start with some of the business drivers in banking. I mean, obviously online. A lot of people have sort of joked that many of the older people, who kind of shunned online banking would love to go into the branch and see their friendly teller had no choice, during this pandemic, to go to online. So that's obviously a big trend you mentioned, the data driven data warehouse, I want to understand that, but what at the top level, what are some of the key business drivers that are catalyzing your desire for change? >> The ability to give a customer, what they need at the time when they need it. And what I mean by that is that we have customer interactions in multiple ways. And I want to be able for the customer to walk into a bank or online and see the same format, and being able to have the same feel the same love, and also to be able to offer them the next best offer for them. But they're if they want looking for a new mortgage or looking to refinance, or whatever it is that they have that data, we have the data and that they feel comfortable using it. And that's an untethered banker. Attitude is, whatever my banker is holding and whatever the person is holding in their phone, that is the same and it's comfortable. So they don't feel that they've walked into the bank and they have to do fill out different paperwork compared to filling out paperwork on just doing it on their phone. >> You actually do want the experience to be better. And it is in many cases. Now you weren't able to do this with your existing I guess mainframe based Enterprise Data Warehouses. Is that right? Maybe talk about that a little bit? >> Yeah, we were definitely able to do it with what we have today the technology we're using. But one of the issues is that it's not timely. And you need a timely process to be able to get the customers to understand what's happening. You need a timely process so we can enhance our risk management. We can apply for fraud issues and things like that. >> Yeah, so you're trying to get more real time. The traditional EDW. It's sort of a science project. There's a few experts that know how to get it. You can so line up, the demand is tremendous. And then oftentimes by the time you get the answer, it's outdated. So you're trying to address that problem. So part of it is really the cycle time the end to end cycle time that you're progressing. And then there's, if I understand it residual benefits that are pretty substantial from a revenue opportunity, other offers that you can make to the right customer, that you maybe know, through your data, is that right? >> Exactly. It's drive new customers to new opportunities. It's enhanced the risk, and it's to optimize the banking process, and then obviously, to create new business. And the only way we're going to be able to do that is if we have the ability to look at the data right when the customer walks in the door or right when they open up their app. And by doing creating more to New York times near real time data, or the data warehouse team that's giving the lines of business the ability to work on the next best offer for that customer as well. >> But Paula, we're inundated with data sources these days. Are there other data sources that maybe had access to before, but perhaps the backlog of ingesting and cleaning in cataloging and analyzing maybe the backlog was so great that you couldn't perhaps tap some of those data sources. Do you see the potential to increase the data sources and hence the quality of the data or is that sort of premature? >> Oh, no. Exactly. Right. So right now, we ingest a lot of flat files and from our mainframe type of front end system, that we've had for quite a few years. But now that we're moving to the cloud and off-prem and on-prem, moving off-prem, into like an S3 Bucket, where that data we can process that data and get that data faster by using real time tools to move that data into a place where, like snowflake could utilize that data, or we can give it out to our market. Right now we're about we do work in batch mode still. So we're doing 24 hours. >> Okay. So when I think about the data pipeline, and the people involved, maybe you could talk a little bit about the organization. You've got, I don't know, if you have data scientists or statisticians, I'm sure you do. You got data architects, data engineers, quality engineers, developers, etc. And oftentimes, practitioners like yourself, will stress about, hey, the data is in silos. The data quality is not where we want it to be. We have to manually categorize the data. These are all sort of common data pipeline problems, if you will. Sometimes we use the term data Ops, which is sort of a play on DevOps applied to the data pipeline. Can you just sort of describe your situation in that context? >> Yeah, so we have a very large data ops team. And everyone that who is working on the data part of Webster's Bank, has been there 13 to 14 years. So they get the data, they understand it, they understand the lines of business. So it's right now. We could the we have data quality issues, just like everybody else does. But we have places in them where that gets cleansed. And we're moving toward and there was very much siloed data. The data scientists are out in the lines of business right now, which is great, because I think that's where data science belongs, we should give them and that's what we're working towards now is giving them more self service, giving them the ability to access the data in a more robust way. And it's a single source of truth. So they're not pulling the data down into their own, like Tableau dashboards, and then pushing the data back out. So they're going to more not, I don't want to say, a central repository, but a more of a robust repository, that's controlled across multiple avenues, where multiple lines of business can access that data. Is that help? >> Got it, Yes. And I think that one of the key things that I'm taking away from your last comment, is the cultural aspects of this by having the data scientists in the line of business, the lines of business will feel ownership of that data as opposed to pointing fingers criticizing the data quality. They really own that that problem, as opposed to saying, well, it's Paula's problem. >> Well, I have my problem is I have data engineers, data architects, database administrators, traditional data reporting people. And because some customers that I have that are business customers lines of business, they want to just subscribe to a report, they don't want to go out and do any data science work. And we still have to provide that. So we still want to provide them some kind of regiment that they wake up in the morning, and they open up their email, and there's the report that they subscribe to, which is great, and it works out really well. And one of the things is why we purchased Io-Tahoe was, I would have the ability to give the lines of business, the ability to do search within the data. And we'll read the data flows and data redundancy and things like that, and help me clean up the data. And also, to give it to the data analysts who say, all right, they just asked me they want this certain report. And it used to take okay, four weeks we're going to go and we're going to look at the data and then we'll come back and tell you what we can do. But now with Io-Tahoe, they're able to look at the data, and then in one or two days, they'll be able to go back and say, Yes, we have the data, this is where it is. This is where we found it. This is the data flows that we found also, which is what I call it, is the break of a column. It's where the column was created, and where it went to live as a teenager. (laughs) And then it went to die, where we archive it. And, yeah, it's this cycle of life for a column. And Io-Tahoe helps us do that. And we do data lineage is done all the time. And it's just takes a very long time and that's why we're using something that has AI in it and machine running. It's accurate, it does it the same way over and over again. If an analyst leaves, you're able to utilize something like Io-Tahoe to be able to do that work for you. Is that help? >> Yeah, so got it. So a couple things there, in researching Io-Tahoe, it seems like one of the strengths of their platform is the ability to visualize data, the data structure and actually dig into it, but also see it. And that speeds things up and gives everybody additional confidence. And then the other piece is essentially infusing AI or machine intelligence into the data pipeline, is really how you're attacking automation. And you're saying it repeatable, and then that helps the data quality and you have this virtual cycle. Maybe you could sort of affirm that and add some color, perhaps. >> Exactly. So you're able to let's say that I have seven cars, lines of business that are asking me questions, and one of the questions they'll ask me is, we want to know, if this customer is okay to contact, and there's different avenues so you can go online, do not contact me, you can go to the bank and you can say, I don't want email, but I'll take texts. And I want no phone calls. All that information. So, seven different lines of business asked me that question in different ways. One said, "No okay to contact" the other one says, "Customer 123." All these. In each project before I got there used to be siloed. So one customer would be 100 hours for them to do that analytical work, and then another analyst would do another 100 hours on the other project. Well, now I can do that all at once. And I can do those types of searches and say, Yes, we already have that documentation. Here it is, and this is where you can find where the customer has said, "No, I don't want to get access from you by email or I've subscribed to get emails from you." >> Got it. Okay. Yeah Okay. And then I want to go back to the cloud a little bit. So you mentioned S3 Buckets. So you're moving to the Amazon cloud, at least, I'm sure you're going to get a hybrid situation there. You mentioned snowflake. What was sort of the decision to move to the cloud? Obviously, snowflake is cloud only. There's not an on-prem, version there. So what precipitated that? >> Alright, so from I've been in the data IT information field for the last 35 years. I started in the US Air Force, and have moved on from since then. And my experience with Bob Graham, was with snowflake with working with GE Capital. And that's where I met up with the team from Io-Tahoe as well. And so it's a proven so there's a couple of things one is Informatica, is worldwide known to move data. They have two products, they have the on-prem and the off-prem. I've used the on-prem and off-prem, they're both great. And it's very stable, and I'm comfortable with it. Other people are very comfortable with it. So we picked that as our batch data movement. We're moving toward probably HVR. It's not a total decision yet. But we're moving to HVR for real time data, which is changed capture data, moves it into the cloud. And then, so you're envisioning this right now. In which is you're in the S3, and you have all the data that you could possibly want. And that's JSON, all that everything is sitting in the S3 to be able to move it through into snowflake. And snowflake has proven to have a stability. You only need to learn and train your team with one thing. AWS as is completely stable at this point too. So all these avenues if you think about it, is going through from, this is your data lake, which is I would consider your S3. And even though it's not a traditional data lake like, you can touch it like a Progressive or Hadoop. And then into snowflake and then from snowflake into sandbox and so your lines of business and your data scientists just dive right in. That makes a big win. And then using Io-Tahoe with the data automation, and also their search engine. I have the ability to give the data scientists and data analysts the way of they don't need to talk to IT to get accurate information or completely accurate information from the structure. And we'll be right back. >> Yeah, so talking about snowflake and getting up to speed quickly. I know from talking to customers you can get from zero to snowflake very fast and then it sounds like the Io-Tahoe is sort of the automation cloud for your data pipeline within the cloud. Is that the right way to think about it? >> I think so. Right now I have Io-Tahoe attached to my on-prem. And I want to attach it to my off-prem eventually. So I'm using Io-Tahoe data automation right now, to bring in the data, and to start analyzing the data flows to make sure that I'm not missing anything, and that I'm not bringing over redundant data. The data warehouse that I'm working of, it's an on-prem. It's an Oracle Database, and it's 15 years old. So it has extra data in it. It has things that we don't need anymore, and Io-Tahoe's helping me shake out that extra data that does not need to be moved into my S3. So it's saving me money, when I'm moving from off-prem to on-prem. >> And so that was a challenge prior, because you couldn't get the lines of business to agree what to delete, or what was the issue there? >> Oh, it was more than that. Each line of business had their own structure within the warehouse. And then they were copying data between each other, and duplicating the data and using that. So there could be possibly three tables that have the same data in it, but it's used for different lines of business. We have identified using Io-Tahoe identified over seven terabytes in the last two months on data that has just been repetitive. It's the same exact data just sitting in a different schema. And that's not easy to find, if you only understand one schema, that's reporting for that line of business. >> More bad news for the storage companies out there. (both laughs) So far. >> It's cheap. That's what we were telling people. >> And it's true, but you still would rather not waste it, you'd like to apply it to drive more revenue. And so, I guess, let's close on where you see this thing going. Again, I know you're sort of partway through the journey, maybe you could sort of describe, where you see the phase is going and really what you want to get out of this thing, down the road, mid-term, longer term, what's your vision or your data driven organization. >> I want for the bankers to be able to walk around with an iPad in their hand, and be able to access data for that customer, really fast and be able to give them the best deal that they can get. I want Webster to be right there on top with being able to add new customers, and to be able to serve our existing customers who had bank accounts since they were 12 years old there and now our multi whatever. I want them to be able to have the best experience with our bankers. >> That's awesome. That's really what I want as a banking customer. I want my bank to know who I am, anticipate my needs, and create a great experience for me. And then let me go on with my life. And so that follow. Great story. Love your experience, your background and your knowledge. I can't thank you enough for coming on theCube. >> Now, thank you very much. And you guys have a great day. >> All right, take care. And thank you for watching everybody. Keep right there. We'll take a short break and be right back. (gentle music)
SUMMARY :
to you by Io-Tahoe. And I'm really excited to of a regional I think and they want to move it relates to kind of transitioning And the only way to do But I want to ask you about Covid, and get the data moving And then finally, you got more clarity. and filled out the right amount. and really great support for the region, and being able to have the experience to be better. to be able to get the customers that know how to get it. and it's to optimize the banking process, and analyzing maybe the backlog was and get that data faster and the people involved, And everyone that who is working is the cultural aspects of this the ability to do search within the data. and you have this virtual cycle. and one of the questions And then I want to go back in the S3 to be able to move it Is that the right way to think about it? and to start analyzing the data flows and duplicating the data and using that. More bad news for the That's what we were telling people. and really what you want and to be able to serve And so that follow. And you guys have a great day. And thank you for watching everybody.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
Paula D'Amico | PERSON | 0.99+ |
Paula | PERSON | 0.99+ |
Connecticut | LOCATION | 0.99+ |
Westchester | LOCATION | 0.99+ |
Informatica | ORGANIZATION | 0.99+ |
24 hours | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
13 | QUANTITY | 0.99+ |
thousands | QUANTITY | 0.99+ |
100 hours | QUANTITY | 0.99+ |
Bob Graham | PERSON | 0.99+ |
iPad | COMMERCIAL_ITEM | 0.99+ |
Webster Bank | ORGANIZATION | 0.99+ |
GE Capital | ORGANIZATION | 0.99+ |
first item | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
two products | QUANTITY | 0.99+ |
seven | QUANTITY | 0.99+ |
New York | LOCATION | 0.99+ |
Boston | LOCATION | 0.99+ |
three tables | QUANTITY | 0.99+ |
Each line | QUANTITY | 0.99+ |
first set | QUANTITY | 0.99+ |
two days | QUANTITY | 0.99+ |
DevOps | TITLE | 0.99+ |
Webster bank | ORGANIZATION | 0.99+ |
14 years | QUANTITY | 0.99+ |
over 15 years | QUANTITY | 0.99+ |
seven cars | QUANTITY | 0.98+ |
each project | QUANTITY | 0.98+ |
Friday night | DATE | 0.98+ |
Enterprise Data Automation | ORGANIZATION | 0.98+ |
New England | LOCATION | 0.98+ |
Io-Tahoe | ORGANIZATION | 0.98+ |
today | DATE | 0.98+ |
Webster's Bank | ORGANIZATION | 0.98+ |
one schema | QUANTITY | 0.97+ |
Fairfield County | LOCATION | 0.97+ |
One | QUANTITY | 0.97+ |
one customer | QUANTITY | 0.97+ |
over seven terabytes | QUANTITY | 0.97+ |
Salesforce | ORGANIZATION | 0.96+ |
both | QUANTITY | 0.95+ |
single source | QUANTITY | 0.93+ |
one thing | QUANTITY | 0.93+ |
US Air Force | ORGANIZATION | 0.93+ |
Webster | ORGANIZATION | 0.92+ |
S3 | COMMERCIAL_ITEM | 0.92+ |
Enterprise Data Architecture | ORGANIZATION | 0.91+ |
Io Tahoe | PERSON | 0.91+ |
Oracle | ORGANIZATION | 0.9+ |
15 years old | QUANTITY | 0.9+ |
Io-Tahoe | PERSON | 0.89+ |
12 years old | QUANTITY | 0.88+ |
Tableau | TITLE | 0.87+ |
four weeks | QUANTITY | 0.86+ |
S3 Buckets | COMMERCIAL_ITEM | 0.84+ |
Covid | PERSON | 0.81+ |
Data Architecture | ORGANIZATION | 0.79+ |
JSON | TITLE | 0.79+ |
Senior Vice President | PERSON | 0.78+ |
24 seven development role | QUANTITY | 0.77+ |
last 35 years | DATE | 0.77+ |
both laughs | QUANTITY | 0.75+ |
Io-Tahoe | TITLE | 0.73+ |
each | QUANTITY | 0.72+ |
loans | QUANTITY | 0.71+ |
zero | QUANTITY | 0.71+ |
Yusef Khan, Io Tahoe | Enterprise Data Automation
>>from around the globe. It's the Cube with digital coverage of enterprise data automation, an event Siri's brought to you by Iot. Tahoe, everybody, We're back. We're talking about enterprise data automation. The hashtag is data automated, and we're going to really dig into data migrations, data, migrations. They're risky. They're time consuming, and they're expensive. Yousef con is here. He's the head of partnerships and alliances at I o ta ho coming again from London. Hey, good to see you, Seth. Thanks very much. >>Thank you. >>So your role is is interesting. We're talking about data migrations. You're gonna head of partnerships. What is your role specifically? And how is it relevant to what we're gonna talk about today? >>Uh, I work with the various businesses such as cloud companies, systems integrators, companies that sell operating systems, middleware, all of whom are often quite well embedded within a company. I t infrastructures and have existing relationships. Because what we do fundamentally makes migrating to the cloud easier on data migration easier. A lot of businesses that are interested in partnering with us. Um, we're interested in parting with, So >>let's set up the problem a little bit. And then I want to get into some of the data. You know, I said that migration is a risky, time consuming, expensive. They're they're often times a blocker for organizations to really get value out of data. Why is that? >>Uh, I think I mean, all migrations have to start with knowing the facts about your data, and you can try and do this manually. But when that you have an organization that may have been going for decades or longer, they will probably have a pretty large legacy data estate so that I have everything from on premise mainframes. They may have stuff which is probably in the cloud, but they probably have hundreds, if not thousands of applications and potentially hundreds of different data stores. Um, now they're understanding of what they have. Ai's often quite limited because you can try and draw a manual maps, but they're outdated very quickly. Every time that data changes the manual that's out of date on people obviously leave organizations over time, so that kind of tribal knowledge gets built up is limited as well. So you can try a Mackel that manually you might need a db. Hey, thanks. Based analyst or ah, business analyst, and they won't go in and explore the data for you. But doing that manually is very, very time consuming this contract teams of people, months and months. Or you can use automation just like what's the bank with Iot? And they managed to do this with a relatively small team. Are in a timeframe of days. >>Yeah, we talked to Paul from Webster Bank. Awesome discussion. So I want to dig into this migration and let's let's pull up graphic it will talk about. We'll talk about what a typical migration project looks like. So what you see here it is. It's very detailed. I know it's a bit of an eye test, but let me call your attention to some of the key aspects of this Ah, and then use. If I want you to chime in. So at the top here, you see that area graph that's operational risk for a typical migration project, and you can see the timeline and the the milestones. That blue bar is the time to test so you can see the second step data analysis talking 24 weeks so, you know, very time consuming. And then Let's not get dig into the stuff in the middle of the fine print, but there's some real good detail there, but go down the bottom. That's labor intensity in the in the bottom and you can see high is that sort of brown and and you could see a number of data analysis, data staging data prep, the trial, the implementation post implementation fixtures, the transition toe B A B a year, which I think is business as usual. Those are all very labor intensive. So what do you take aways from this typical migration project? What do we need to know yourself? >>I mean, I think the key thing is, when you don't understand your data upfront, it's very difficult to scope to set up a project because you go to business stakeholders and decision makers and you say Okay, we want to migrate these data stores. We want to put them in the cloud most often, but actually, you probably don't know how much data is there. You don't necessarily know how many applications that relates to, you know, the relationships between the data. You don't know the flow of the data. So the direction in which the data is going between different data stores and tables, so you start from a position where you have pretty high risk and alleviate that risk. You could be stacking project team of lots and lots of people to do the next base, which is analysis. And so you set up a project which has got a pretty high cost. The big projects, more people, the heavy of governance, obviously on then there, then in the phase where they're trying to do lots and lots of manual analysis manage. That, in a sense, is, as we all know, on the idea of trying to relate data that's in different those stores relating individual tables and columns. Very, very time consuming, expensive. If you're hiring in resource from consultants or systems integrators externally, you might need to buy or to use party tools, Aziz said earlier. The people who understand some of those systems may have left a while ago. See you even high risks quite cost situation from the off on the same things that have developed through the project. Um, what are you doing with it, Ayatollah? Who is that? We're able to automate a lot of this process from the very beginning because we can do the initial data. Discovery run, for example, automatically you very quickly have an automated validator. A data map on the data flow has been generated automatically, much less time and effort and much less cars. Doctor Marley. >>Okay, so I want to bring back that that first chart, and I want to call your attention to the again that area graph the blue bars and then down below that labor intensity. And now let's bring up the the the same chart. But with a set of an automation injection in here and now. So you now see the So let's go Said Accelerated by Iot, Tom. Okay, great. And we're going to talk about this. But look, what happens to the operational risk. A dramatic reduction in that. That graph. And then look at the bars, the bars, those blue bars. You know, data analysis went from 24 weeks down to four weeks and then look at the labor intensity. The it was all these were high data analysis data staging data prep. Try a lot post implementation fixtures in transition to be a you. All of those went from high labor intensity. So we've now attack that and gone to low labor intensity. Explain how that magic happened. >>I think that the example off a data catalog. So every large enterprise wants to have some kind of repository where they put all their understanding about their data in its Price States catalog, if you like, um, imagine trying to do that manually. You need to go into every individual data store. You need a DB a business analyst, rich data store they need to do in extracted the data table was individually they need to cross reference that with other data school, it stores and schemers and tables. You probably were the mother of all lock Excel spreadsheets. It would be a very, very difficult exercise to do. I mean, in fact, one of our reflections as we automate lots of data lots of these things is, um it accelerates the ability to water may, But in some cases, it also makes it possible for enterprise customers with legacy systems um, take banks, for example. There quite often end up staying on mainframe systems that they've had in place for decades. Uh, no migrating away from them because they're not able to actually do the work of understanding the data g duplicating the data, deleting data isn't relevant and then confidently going forward to migrate. So they stay where they are with all the attendant problems assistance systems that are out of support. Go back to the data catalog example. Um, whatever you discover invades, discovery has to persist in a tool like a data catalog. And so we automate data catalog books, including Out Way Cannot be others, but we have our own. The only alternative to this kind of automation is to build out this very large project team or business analysts off db A's project managers processed analysts together with data to understand that the process of gathering data is correct. To put it in the repository to validate it except etcetera, we've got into organizations and we've seen them ramp up teams off 2030 people costs off £234 million a year on a time frame, 15 20 years just to try and get a data catalog done. And that's something that we can typically do in a timeframe of months, if not weeks. And the difference is using automation. And if you do what? I've just described it. In this manual situation, you make migrations to the cloud prohibitively expensive. Whatever saving you might make from shutting down your legacy data stores, we'll get eaten up by the cost of doing it. Unless you go with the more automated approach. >>Okay, so the automated approach reduces risk because you're not gonna, you know you're going to stay on project plan. Ideally, it's all these out of scope expectations that come up with the manual processes that kill you in the rework andan that data data catalog. People are afraid that their their family jewels data is not going to make it through to the other side. So So that's something that you're you're addressing and then you're also not boiling the ocean. You're really taking the pieces that are critical and stuff you don't need. You don't have to pay for >>process. It's a very good point. I mean, one of the other things that we do and we have specific features to do is to automatically and noise data for a duplication at a rover or record level and redundancy on a column level. So, as you say before you go into a migration process. You can then understand. Actually, this stuff it was replicated. We don't need it quite often. If you put data in the cloud you're paying, obviously, the storage based offer compute time. The more data you have in there that's duplicated, that is pure cost. You should take out before you migrate again if you're trying to do that process of understanding what's duplicated manually off tens or hundreds of bases stores. It was 20 months, if not years. Use machine learning to do that in an automatic way on it's much, much quicker. I mean, there's nothing I say. Well, then, that costs and benefits of guitar. Every organization we work with has a lot of money existing, sunk cost in their I t. So have your piece systems like Oracle or Data Lakes, which they've spent a good time and money investing in. But what we do by enabling them to transition everything to the strategic future repositories, is accelerate the value of that investment and the time to value that investment. So we're trying to help people get value out of their existing investments on data estate, close down the things that they don't need to enable them to go to a kind of brighter, more future well, >>and I think as well, you know, once you're able to and this is a journey, we know that. But once you're able to go live on, you're infusing sort of a data mindset, a data oriented culture. I know it's somewhat buzzword, but when you when you see it in organizations, you know it's really and what happens is you dramatically reduce that and cycle time of going from data to actually insights. Data's plentiful, but insights aren't, and that is what's going to drive competitive advantage over the next decade and beyond. >>Yeah, definitely. And you could only really do that if you get your data estate cleaned up in the first place. Um, I worked with the managed teams of data scientists, data engineers, business analysts, people who are pushing out dashboards and trying to build machine learning applications. You know, you know, the biggest frustration for lots of them and the thing that they spend far too much time doing is trying to work out what the right data is on cleaning data, which really you don't want a highly paid thanks to scientists doing with their time. But if you sort out your data stays in the first place, get rid of duplication. If that pans migrate to cloud store, where things are really accessible on its easy to build connections and to use native machine learning tools, you're well on the way up to date the maturity curve on you can start to use some of those more advanced applications. >>You said. What are some of the pre requisites? Maybe the top few that are two or three that I need to understand as a customer to really be successful here? Is it skill sets? Is it is it mindset leadership by in what I absolutely need to have to make this successful? >>Well, I think leadership is obviously key just to set the vision of people with spiky. One of the great things about Ayatollah, though, is you can use your existing staff to do this work. If you've used on automation, platform is no need to hire expensive people. Alright, I was a no code solution. It works out of the box. You just connect to force on your existing stuff can use. It's very intuitive that has these issues. User interface? >>Um, it >>was only to invest vast amounts with large consultants who may well charging the earth. Um, and you already had a bit of an advantage. If you've got existing staff who are close to the data subject matter experts or use it because they can very easily learn how to use a tool on, then they can go in and they can write their own data quality rules on. They can really make a contribution from day one, when we are go into organizations on way. Can I? It's one of the great things about the whole experience. Veritas is. We can get tangible results back within the day. Um, usually within an hour or two great ones to say Okay, we started to map relationships. Here's the data map of the data that we've analyzed. Harrison thoughts on where the sensitive data is because it's automated because it's running algorithms stater on. That's what they were really to expect. >>Um, >>and and you know this because you're dealing with the ecosystem. We're entering a new era of data and many organizations to your point, they just don't have the resources to do what Google and Amazon and Facebook and Microsoft did over the past decade To become data dominant trillion dollar market cap companies. Incumbents need to rely on technology companies to bring that automation that machine intelligence to them so they can apply it. They don't want to be AI inventors. They want to apply it to their businesses. So and that's what really was so difficult in the early days of so called big data. You have this just too much complexity out there, and now companies like Iot Tahoe or bringing your tooling and platforms that are allowing companies to really become data driven your your final thoughts. Please use it. >>That's a great point, Dave. In a way, it brings us back to where it began. In terms of partnerships and alliances. I completely agree with a really exciting point where we can take applications like Iot. Uh, we can go into enterprises and help them really leverage the value of these type of machine learning algorithms. And and I I we work with all the major cloud providers AWS, Microsoft Azure or Google Cloud Platform, IBM and Red Hat on others, and we we really I think for us. The key thing is that we want to be the best in the world of enterprise data automation. We don't aspire to be a cloud provider or even a workflow provider. But what we want to do is really help customers with their data without automated data functionality in partnership with some of those other businesses so we can leverage the great work they've done in the cloud. The great work they've done on work flows on virtual assistants in other areas. And we help customers leverage those investments as well. But our heart, we really targeted it just being the best, uh, enterprised data automation business in the world. >>Massive opportunities not only for technology companies, but for those organizations that can apply technology for business. Advantage yourself, count. Thanks so much for coming on the Cube. Appreciate. All right. And thank you for watching everybody. We'll be right back right after this short break. >>Yeah, yeah, yeah, yeah.
SUMMARY :
of enterprise data automation, an event Siri's brought to you by Iot. And how is it relevant to what we're gonna talk about today? fundamentally makes migrating to the cloud easier on data migration easier. a blocker for organizations to really get value out of data. And they managed to do this with a relatively small team. That blue bar is the time to test so you can see the second step data analysis talking 24 I mean, I think the key thing is, when you don't understand So you now see the So let's go Said Accelerated by Iot, You need a DB a business analyst, rich data store they need to do in extracted the data processes that kill you in the rework andan that data data catalog. close down the things that they don't need to enable them to go to a kind of brighter, and I think as well, you know, once you're able to and this is a journey, And you could only really do that if you get your data estate cleaned up in I need to understand as a customer to really be successful here? One of the great things about Ayatollah, though, is you can use Um, and you already had a bit of an advantage. and and you know this because you're dealing with the ecosystem. And and I I we work And thank you for watching everybody.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Paul | PERSON | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.99+ | |
Amazon | ORGANIZATION | 0.99+ |
London | LOCATION | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.99+ | |
Yusef Khan | PERSON | 0.99+ |
Seth | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
20 months | QUANTITY | 0.99+ |
Aziz | PERSON | 0.99+ |
hundreds | QUANTITY | 0.99+ |
tens | QUANTITY | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Webster Bank | ORGANIZATION | 0.99+ |
24 weeks | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
four weeks | QUANTITY | 0.99+ |
three | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Io Tahoe | PERSON | 0.99+ |
Marley | PERSON | 0.99+ |
Harrison | PERSON | 0.99+ |
Data Lakes | ORGANIZATION | 0.99+ |
Siri | TITLE | 0.99+ |
Excel | TITLE | 0.99+ |
Veritas | ORGANIZATION | 0.99+ |
second step | QUANTITY | 0.99+ |
15 20 years | QUANTITY | 0.98+ |
Tahoe | PERSON | 0.98+ |
One | QUANTITY | 0.98+ |
first chart | QUANTITY | 0.98+ |
an hour | QUANTITY | 0.98+ |
Red Hat | ORGANIZATION | 0.98+ |
one | QUANTITY | 0.97+ |
Tom | PERSON | 0.96+ |
hundreds of bases | QUANTITY | 0.96+ |
first | QUANTITY | 0.95+ |
next decade | DATE | 0.94+ |
first place | QUANTITY | 0.94+ |
Iot | ORGANIZATION | 0.94+ |
Iot | TITLE | 0.93+ |
earth | LOCATION | 0.93+ |
day one | QUANTITY | 0.92+ |
Mackel | ORGANIZATION | 0.91+ |
today | DATE | 0.91+ |
Ayatollah | PERSON | 0.89+ |
£234 million a year | QUANTITY | 0.88+ |
data | QUANTITY | 0.88+ |
Iot | PERSON | 0.83+ |
hundreds of | QUANTITY | 0.81+ |
thousands of applications | QUANTITY | 0.81+ |
decades | QUANTITY | 0.8+ |
I o ta ho | ORGANIZATION | 0.75+ |
past decade | DATE | 0.75+ |
Microsoft Azure | ORGANIZATION | 0.72+ |
two great ones | QUANTITY | 0.72+ |
2030 people | QUANTITY | 0.67+ |
Doctor | PERSON | 0.65+ |
States | LOCATION | 0.65+ |
Iot Tahoe | ORGANIZATION | 0.65+ |
a year | QUANTITY | 0.55+ |
Yousef | PERSON | 0.45+ |
Cloud Platform | TITLE | 0.44+ |
Cube | ORGANIZATION | 0.38+ |