Alex Schuchman , Colgate Palmolive | CUBE Conversation
(upbeat music) >> Hi everyone, and welcome back to managing risk across your extended attack service area with Armis Asset Intelligence Platform. I'm John Furrier, your host. We're here with the CISO Perspective. Alex Schuchman, who is the CISO of Colgate-Palmolive Company. Alex, thanks for coming on. >> Thanks for having me. >> You know, unified visibility across the enterprise service area is about knowing what you got to protect. You can't protect what you can't see. Tell me more about how you guys are able to centralize your view with network assets with Armis. >> Yeah, I think the most important part of any security program is really visibility. And that's one of the building blocks when you're building a security program. You need to understand what's in your environment, what you can control, what is being introduced new into the environment, and that's really what, any solution that gives you full visibility to your infrastructure, to your environment, to all the assets that are there, that's really one of your bread and butter pieces to your security program. >> What's been the impact on your business? >> You know, I think from an IT point of view, running the security program, you know, our key thing is really enabling the business to do their job better. So if we can give them visibility into all the assets that are available in their individual environments, and we're doing that in an automated fashion with no manual collection, you know, that's yet another thing that they don't have to worry about, and then we're delivering. Because really IT is an enabler for the business. And then they can focus really on what their job is, which is to deliver product. >> Yeah, and a lot of changes in their network. You got infrastructure, you got IOT devices, OT devices. So vulnerability management becomes more important. It's been around for a while, but it's not just IT devices anymore. There are gaps in vulnerability across the OT network. What can you tell us about Colgate's use of Armis' vulnerability management? What can you see now? What couldn't you see before? Can you share your thoughts on this? >> Yeah, I think what's really interesting about the kind of manufacturing environments today is, if you look back a number of years, most of the manufacturing equipment was really disconnected from the internet. It was really running in silos. So it was very easy to protect equipment that isn't internet-connected. You could put a firewall, you could segment it off. And it was really on an island on its own. Nowadays, you have a lot of IOT devices. you have a lot of internet-connected devices, sensors providing information to multiple different suppliers or vendor solutions. And you have to really then open up your ecosystem more, which, of course, means you have to change your security posture, and you really have to embrace if there's a vulnerability with one of those suppliers then how do you mitigate the risk associated to that vulnerability? Armis really helps us get a lot of information so that we can then make a decision with our business teams. >> That whole operational aspect of criticality is huge, on the assets knowing what's key. How has that changed the security workload for you guys? >> You know, for us, I mean, it's all about being efficient. If we can have the visibility across our manufacturing environments, then my team can easily consume that information. You know, if we spend a lot of time trying to digest the information, trying to process it, trying to prioritize it, that really hurts our efficiency as a team or as a function. What we really like is being able to use technology to help us do that work. We're not an IT shop. We're a manufacturing shop, but we're a very technical shop so we like to drive everything through automation and not be a bottleneck for any of the actions that take place. >> You know the old expression, is the juice worth the squeeze? It comes up a lot when people are buying tools around vulnerability management, and point for all this stuff. So SaaS solution is key with no agents to deploy. They have that. Talk about how you operationalize Armis in your environment. How quickly did it achieve time to value? Take us through that consumption of the product, and what was the experience like? >> Yeah, I'll definitely say in the security ecosystem, that's one of the biggest promises you hear across the industry. And when we started with Armis, we started with a very small deployment, and we wanted to make sure if it was really worth the lift, to your point. We implemented the first set of plants very quickly, actually even quicker than we had put in our project plan, which is not typical for implementing complex security solutions. And then we were so successful with that, we expanded to cover more of our manufacturing plants, and we were able to get really true visibility across our entire manufacturing organization in the first year, with the ability to also say that we extended that information, that visibility to our manufacturing organization, and they could also consume it just as easily as we could. >> That's awesome. How many assets did you guys discover? Just curious on the numbers? >> Oh, that's the really interesting part. You know, before we started this project we would've had to do a manual audit of our plants, which is typical in our industry. You know, when we started this project and we put in estimates, we really didn't have a great handle on what we were going to find. And what's really nice about the Armis solution is it's truly giving you full visibility. So you're actually seeing, besides the servers, and the PLCs, and all the equipment that you're familiar with, you're also connecting it to your wireless access points. You're connecting it to see any of those IOT devices as well. And then you're really getting full visibility through all the integrations that they offer. You're amazed how many devices you're actually seeing across your entire ecosystem. >> It's like Google maps for your infrastructure. You know, the street view. You want to look at it. You get the, you know, fake tree in there, whatever, but it gives you the picture. That's key. >> Correct. And with a nice visualization and an easy search engine, similar to your Google analogy, you know, everything is really at your fingertips. If you want to find something, you just go to the search bar, click a couple entries and boom, you get your list of the associated devices or the the associated locations devices. >> Well, Alex, I appreciate your time. I know you're super busy at CSIG a lot of your plate. Thanks for coming on sharing. Appreciate it. >> No problem, John. Thanks for having me. >> Okay. In a moment, Bryan Inman, a sales engineer at Armis will be joining me. You're watching theCUBE, the leader in high tech coverage. Thanks for watching. (upbeat music)
SUMMARY :
across your extended attack service area You can't protect what you can't see. And that's one of the building blocks running the security program, you know, Can you share your thoughts on this? the risk associated to that How has that changed the for any of the actions You know the old expression, the ability to also say Just curious on the numbers? and all the equipment You know, the street view. you get your list of CSIG a lot of your plate. Thanks for having me. Thanks for watching.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Bryan Inman | PERSON | 0.99+ |
Alex Schuchman | PERSON | 0.99+ |
Alex | PERSON | 0.99+ |
John | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
Armis | ORGANIZATION | 0.99+ |
Colgate | ORGANIZATION | 0.98+ |
Google maps | TITLE | 0.98+ |
first set | QUANTITY | 0.98+ |
today | DATE | 0.97+ |
ORGANIZATION | 0.96+ | |
first year | QUANTITY | 0.96+ |
Colgate-Palmolive Company | ORGANIZATION | 0.95+ |
one | QUANTITY | 0.94+ |
CSIG | ORGANIZATION | 0.94+ |
Palmolive | PERSON | 0.81+ |
Armis' | ORGANIZATION | 0.66+ |
couple entries | QUANTITY | 0.64+ |
Armis Asset Intelligence Platform | ORGANIZATION | 0.61+ |
Arik Pelkey, Pentaho - BigData SV 2017 - #BigDataSV - #theCUBE
>> Announcer: Live from Santa Fe, California, it's the Cube covering Big Data Silicon Valley 2017. >> Welcome, back, everyone. We're here live in Silicon Valley in San Jose for Big Data SV in conjunct with stratAHEAD Hadoop part two. Three days of coverage here in Silicon Valley and Big Data. It's our eighth year covering Hadoop and the Hadoop ecosystem. Now expanding beyond just Hadoop into AI, machine learning, IoT, cloud computing with all this compute is really making it happen. I'm John Furrier with my co-host George Gilbert. Our next guest is Arik Pelkey who is the senior director of product marketing at Pentaho that we've covered many times and covered their event at Pentaho world. Thanks for joining us. >> Thank you for having me. >> So, in following you guys I'll see Pentaho was once an independent company bought by Hitachi, but still an independent group within Hitachi. >> That's right, very much so. >> Okay so you guys some news. Let's just jump into the news. You guys announced some of the machine learning. >> Exactly, yeah. So, Arik Pelkey, Pentaho. We are a data integration and analytics software company. You mentioned you've been doing this for eight years. We have been at Big Data for the past eight years as well. In fact, we're one of the first vendors to support Hadoop back in the day, so we've been along for the journey ever since then. What we're announcing today is really exciting. It's a set of machine learning orchestration capabilities, which allows data scientists, data engineers, and data analysts to really streamline their data science processes. Everything from ingesting new data sources through data preparation, feature engineering which is where a lot of data scientists spend their time through tuning their models which can still be programmed in R, in Weka, in Python, and any other kind of data science tool of choice. What we do is we help them deploy those models inside of Pentaho as a step inside of Pentaho, and then we help them update those models as time goes on. So, really what this is doing is it's streamlining. It's making them more productive so that they can focus their time on things like model building rather than data preparation and feature engineering. >> You know, it's interesting. The market is really active right now around machine learning and even just last week at Google Next, which is their cloud event, they had made the acquisition of Kaggle, which is kind of an open data science. You mentioned the three categories: data engineer, data science, data analyst. Almost on a progression, super geek to business facing, and there's different approaches. One of the comments from the CEO of Kaggle on the acquisition when we wrote up at Sylvan Angle was, and I found this fascinating, I want to get your commentary and reaction to is, he says the data science tools are as early as generations ago, meaning that all the advances and open source and tooling and software development is far along, but now data science is still at that early stage and is going to get better. So, what's your reaction to that, because this is really the demand we're seeing is a lot of heavy lifing going on in the data science world, yet there's a lot of runway of more stuff to do. What is that more stuff? >> Right. Yeah, we're seeing the same thing. Last week I was at the Gardener Data and Analytics conference, and that was kind of the take there from one of their lead machine learning analysts was this is still really early days for data science software. So, there's a lot of Apache projects out there. There's a lot of other open source activity going on, but there are very few vendors that bring to the table an integrated kind of full platform approach to the data science workflow, and that's what we're bringing to market today. Let me be clear, we're not trying to replace R, or Python, or MLlib, because those are the tools of the data scientists. They're not going anywhere. They spent eight years in their phD program working with these tools. We're not trying to change that. >> They're fluent with those tools. >> Very much so. They're also spending a lot of time doing feature engineering. Some research reports, say between 70 and 80% of their time. What we bring to the table is a visual drag and drop environment to do feature engineering a much faster, more efficient way than before. So, there's a lot of different kind of desperate siloed applications out there that all do interesting things on their own, but what we're doing is we're trying to bring all of those together. >> And the trends are reduce the time it takes to do stuff and take away some of those tasks that you can use machine learning for. What unique capabilities do you guys have? Talk about that for a minute, just what Pentaho is doing that's unique and added value to those guys. >> So, the big thing is I keep going back to the data preparation part. I mean, that's 80% of time that's still a really big challenge. There's other vendors out there that focus on just the data science kind of workflow, but where we're really unqiue is around being able to accommodate very complex data environments, and being able to onboard data. >> Give me an example of those environments. >> Geospatial data combined with data from your ERP or your CRM system and all kinds of different formats. So, there might be 15 different data formats that need to be blended together and standardized before any of that can really happen. That's the complexity in the data. So, Pentaho, very consistent with everything else that we do outside of machine learning, is all about helping our customers solve those very complex data challenges before doing any kind of machine learning. One example is one customer is called Caterpillar Machine Asset Intelligence. So, their doing predictive maintenance onboard container ships and on ferry's. So, they're taking data from hundreds and hundreds of sensors onboard these ships, combining that kind of operational sensor data together with geospatial data and then they're serving up predictive maintenance alerts if you will, or giving signals when it's time to replace an engine or complace a compressor or something like that. >> Versus waiting for it to break. >> Versus waiting for it to break, exactly. That's one of the real differentiators is that very complex data environment, and then I was starting to move toward the other differentiator which is our end to end platform which allows customers to deliver these analytics in an embedded fashion. So, kind of full circle, being able to send that signal, but not to an operational system which is sometimes a challenge because you might have to rewrite the code. Deploying models is a really big challenge within Pentaho because it is this fully integrated application. You can deploy the models within Pentaho and not have to jump out into a mainframe environment or something like that. So, I'd say differentiators are very complex data environments, and then this end to end approach where deploying models is much easier than ever before. >> Perhaps, let's talk about alternatives that customers might see. You have a tool suite, and others might have to put together a suite of tools. Maybe tell us some of the geeky version would be the impendent mismatch. You know, like the chasms you'd find between each tool where you have to glue them together, so what are some of those pitfalls? >> One of the challenges is, you have these data scientists working in silos often times. You have data analysts working in silos, you might have data engineers working in silos. One of the big pitfalls is not really collaborating enough to the point where they can do all of this together. So, that's a really big area that we see pitfalls. >> Is it binary not collaborating, or is it that the round trip takes so long that the quality or number of collaborations is so drastically reduced that the output is of lower quality? >> I think it's probably a little bit of both. I think they want to collaborate but one person might sit in Dearborn, Michigan and the other person might sit in Silicon Valley, so there's just a location challenge as well. The other challenge is, some of the data analysts might sit in IT and some of the data scientists might sit in an analytics department somewhere, so it kind of cuts across both location and functional area too. >> So let me ask from the point of view of, you know we've been doing these shows for a number of years and most people have their first data links up and running and their first maybe one or two use cases in production, very sophisticated customers have done more, but what seems to be clear is the highest value coming from those projects isn't to put a BI tool in front of them so much as to do advanced analytics on that data, apply those analytics to inform a decision, whether a person or a machine. >> That's exactly right. >> So, how do you help customers over that hump and what are some other examples that you can share? >> Yeah, so speaking of transformative. I mean, that's what machine learning is all about. It helps companies transform their businesses. We like to talk about that at Pentaho. One customer kind of industry example that I'll share is a company called IMS. IMS is in the business of providing data and analytics to insurance companies so that the insurance companies can price insurance policies based on usage. So, it's a usage model. So, IMS has a technology platform where they put sensors in a car, and then using your mobile phone, can track your driving behavior. Then, your insurance premium that month reflects the driving behavior that you had during that month. In terms of transformative, this is completely upending the insurance industry which has always had a very fixed approach to pricing risk. Now, they understand everything about your behavior. You know, are you turning too fast? Are you breaking too fast, and they're taking it further than that too. They're able to now do kind of a retroactive look at an accident. So, after an accident, they can go back and kind of decompose what happened in the accident and determine whether or not it was your fault or was in fact the ice on the street. So, transformative? I mean, this is just changing things in a really big way. >> I want to get your thoughts on this. I'm just looking at some of the research. You know, we always have the good data but there's also other data out there. In your news, 92% of organizations plan to deploy more predictive analytics, however 50% of organizations have difficulty integrating predictive analytics into their information architecture, which is where the research is shown. So my question to you is, there's a huge gap between the technology landscapes of front end BI tools and then complex data integration tools. That seems to be the sweet spot where the value's created. So, you have the demand and then front end BI's kind of sexy and cool. Wow, I could power my business, but the complexity is really hard in the backend. Who's accessing it? What's the data sources? What's the governance? All these things are complicated, so how do you guys reconcile the front end BI tools and the backend complexity integrations? >> Our story from the beginning has always been this one integrated platform, both for complex data integration challenges together with visualizations, and that's very similar to what this announcement is all about for the data science market. We're very much in line with that. >> So, it's the cart before the horse? Is it like the BI tools are really driven by the data? I mean, it makes sense that the data has to be key. Front end BI could be easy if you have one data set. >> It's funny you say that. I presented at the Gardner conference last week and my topic was, this just in: it's not about analytics. Kind of in jest, but it drove a really big crowd. So, it's about the data right? It's about solving the data problem before you solve the analytics problem whether it's a simple visualization or it's a complex fraud machine learning problem. It's about solving the data problem first. To that quote, I think one of the things that they were referencing was the challenging information architectures into which companies are trying to deploy models and so part of that is when you build a machine learning model, you use R and Python and all these other ones we're familiar with. In order to deploy that into a mainframe environment, someone has to then recode it in C++ or COBOL or something else. That can take a really long time. With our integrated approach, once you've done the feature engineering and the data preparation using our drag and drop environment, what's really interesting is that you're like 90% of the way there in terms of making that model production ready. So, you don't have to go back and change all that code, it's already there because you used it in Pentaho. >> So obviously for those two technologies groups I just mentioned, I think you had a good story there, but it creates problems. You've got product gaps, you've got organizational gaps, you have process gaps between the two. Are you guys going to solve that, or are you currently solving that today? There's a lot of little questions in there, but that seems to be the disconnect. You know, I can do this, I can do that, do I do them together? >> I mean, sticking to my story of one integrated approach to being able to do the entire data science workflow, from beginning to end and that's where we've really excelled. To the extent that more and more data engineers and data analysts and data scientists can get on this one platform even if their using R and WECCA and Python. >> You guys want to close those gaps down, that's what you guys are doing, right? >> We want to make the process more collaborative and more efficient. >> So Dave Alonte has a question on CrowdChat for you. Dave Alonte was in the snowstorm in Boston. Dave, good to see you, hope you're doing well shoveling out the driveway. Thanks for coming in digitally. His question is HDS has been known for mainframes and storage, but Hitachi is an industrial giant. How is Pentaho leveraging Hitatchi's IoT chops? >> Great question, thanks for asking. Hitatchi acquired Pentaho about two years ago, this is before my time. I've been with Pentaho about ten months ago. One of the reasons that they acquired Pentaho is because a platform that they've announced which is called Lumata which is their IoT platform, so what Pentaho is, is the analytics engine that drives that IoT platform Lumata. So, Lumata is about solving more of the hardware sensor, bringing data from the edge into being able to do the analytics. So, it's an incredibly great partnership between Lumata and Pentaho. >> Makes an eternal customer too. >> It's a 90 billion dollar conglomerate so yeah, the acquisition's been great and we're still very much an independent company going to market on our own, but we now have a much larger channel through Hitatchi's reps around the world. >> You've got IoT's use case right there in front of you. >> Exactly. >> But you are leveraging it big time, that's what you're saying? >> Oh yeah, absolutely. We're a very big part of their IoT strategy. It's the analytics. Both of the examples that I shared with you are in fact IoT, not by design but it's because there's a lot of demand. >> You guys seeing a lot of IoT right now? >> Oh yeah. We're seeing a lot of companies coming to us who have just hired a director or vice president of IoT to go out and figure out the IoT strategy. A lot of these are manufacturing companies or coming from industries that are inefficient. >> Digitizing the business model. >> So to the other point about Hitachi that I'll make, is that as it relates to data science, a 90 billion dollar manufacturing and otherwise giant, we have a very deep bench of phD data scientists that we can go to when there's very complex data science problems to solve at customer sight. So, if a customer's struggling with some of the basic how do I get up and running doing machine learning, we can bring our bench of data scientist at Hitatchi to bear in those engagements, and that's a really big differentiator for us. >> Just to be clear and one last point, you've talked about you handle the entire life cycle of modeling from acquiring the data and prepping it all the way through to building a model, deploying it, and updating it which is a continuous process. I think as we've talked about before, data scientists or just the DEV ops community has had trouble operationalizing the end of the model life cycle where you deploy it and update it. Tell us how Pentaho helps with that. >> Yeah, it's a really big problem and it's a very simple solution inside of Pentaho. It's basically a step inside of Pentaho. So, in the case of fraud let's say for example, a prediction might say fraud, not fraud, fraud, not fraud, whatever it is. We can then bring that kind of full lifecycle back into the data workflow at the beginning. It's a simple drag and drop step inside of Pentaho to say which were right and which were wrong and feed that back into the next prediction. We could also take it one step further where there has to be a manual part of this too where it goes to the customer service center, they investigate and they say yes fraud, no fraud, and then that then gets funneled back into the next prediction. So yeah, it's a big challenge and it's something that's relatively easy for us to do just as part of the data science workflow inside of Pentaho. >> Well Arick, thanks for coming on The Cube. We really appreciate it, good luck with the rest of the week here. >> Yeah, very exciting. Thank you for having me. >> You're watching The Cube here live in Silicon Valley covering Strata Hadoop, and of course our Big Data SV event, we also have a companion event called Big Data NYC. We program with O'Reilley Strata Hadoop, and of course have been covering Hadoop really since it's been founded. This is The Cube, I'm John Furrier. George Gilbert. We'll be back with more live coverage today for the next three days here inside The Cube after this short break.
SUMMARY :
it's the Cube covering Big Data Silicon Valley 2017. and the Hadoop ecosystem. So, in following you guys I'll see Pentaho was once You guys announced some of the machine learning. We have been at Big Data for the past eight years as well. One of the comments from the CEO of Kaggle of the data scientists. environment to do feature engineering a much faster, and take away some of those tasks that you can use So, the big thing is I keep going back to the data That's the complexity in the data. So, kind of full circle, being able to send that signal, You know, like the chasms you'd find between each tool One of the challenges is, you have these data might sit in IT and some of the data scientists So let me ask from the point of view of, the driving behavior that you had during that month. and the backend complexity integrations? is all about for the data science market. I mean, it makes sense that the data has to be key. It's about solving the data problem before you solve but that seems to be the disconnect. To the extent that more and more data engineers and more efficient. shoveling out the driveway. One of the reasons that they acquired Pentaho the acquisition's been great and we're still very much Both of the examples that I shared with you of IoT to go out and figure out the IoT strategy. is that as it relates to data science, from acquiring the data and prepping it all the way through and feed that back into the next prediction. of the week here. Thank you for having me. for the next three days here inside The Cube
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
George Gilbert | PERSON | 0.99+ |
Hitachi | ORGANIZATION | 0.99+ |
Dave Alonte | PERSON | 0.99+ |
Pentaho | ORGANIZATION | 0.99+ |
Dave | PERSON | 0.99+ |
90% | QUANTITY | 0.99+ |
Arik Pelkey | PERSON | 0.99+ |
Boston | LOCATION | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
Hitatchi | ORGANIZATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
one | QUANTITY | 0.99+ |
50% | QUANTITY | 0.99+ |
eight years | QUANTITY | 0.99+ |
Arick | PERSON | 0.99+ |
One | QUANTITY | 0.99+ |
Lumata | ORGANIZATION | 0.99+ |
Last week | DATE | 0.99+ |
two technologies | QUANTITY | 0.99+ |
15 different data formats | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
92% | QUANTITY | 0.99+ |
One example | QUANTITY | 0.99+ |
Both | QUANTITY | 0.99+ |
Three days | QUANTITY | 0.99+ |
Python | TITLE | 0.99+ |
Kaggle | ORGANIZATION | 0.99+ |
one customer | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
eighth year | QUANTITY | 0.99+ |
last week | DATE | 0.99+ |
Santa Fe, California | LOCATION | 0.99+ |
two | QUANTITY | 0.99+ |
each tool | QUANTITY | 0.99+ |
90 billion dollar | QUANTITY | 0.99+ |
80% | QUANTITY | 0.99+ |
Caterpillar | ORGANIZATION | 0.98+ |
both | QUANTITY | 0.98+ |
NYC | LOCATION | 0.98+ |
first data | QUANTITY | 0.98+ |
Pentaho | LOCATION | 0.98+ |
San Jose | LOCATION | 0.98+ |
The Cube | TITLE | 0.98+ |
Big Data SV | EVENT | 0.97+ |
COBOL | TITLE | 0.97+ |
70 | QUANTITY | 0.97+ |
C++ | TITLE | 0.97+ |
IMS | TITLE | 0.96+ |
MLlib | TITLE | 0.96+ |
one person | QUANTITY | 0.95+ |
R | TITLE | 0.95+ |
Big Data | EVENT | 0.95+ |
Gardener Data and Analytics | EVENT | 0.94+ |
Gardner | EVENT | 0.94+ |
Strata Hadoop | TITLE | 0.93+ |