John Hennessy, Knight-Hennessy Scholars | ACG SV Grow! Awards 2019
(upbeat techno music) >> From Mountain View California, it's the Cube covering the 15th Annual Grow Awards. Brought to you by ACG SV. >> Hi, Lisa Martin with the Cube on the ground at the Computer History Museum for the 15th annual ACG SV Awards. And in Mountain View California excited to welcome to the Cube for the first time, John Hennessy, the chairman of Alphabet and the co-founder of the Knight-Hennessy Scholars Program at Stanford. JOHN, it's truly a pleasure to have you on the Cube today. >> Well delighted to be here, Lisa. >> So I was doing some research on you. And I see Marc Andreessen has called you the godfather of Silicon Valley. >> Marc very generous (loughs) >> so I thought I was pretty cool I'm going to sit down with the godfather tonight. (loughs) >> I have not done that yet. So you are keynoting the 15th Annual ACG SV Awards tonight. Talk to us a little bit about the takeaways that the audience is going to hear from you tonight. >> Well, they're going to hear some things about leadership the importance of leadership, obviously the importance of innovation. We're in the middle of Silicon Valley innovation is a big thing. And the role that technology plays in our lives and how we should be thinking about that, and how do we ensure the technology is something that serves the public good. >> Definitely. So there's about I think over 230 attendees expected tonight over 100 sea levels, the ACG SV Is has been it's it's much more than a networking organization. there's a lot of opportunities for collaboration for community. Tell me a little bit about your experience with that from a collaboration standpoint? >> Well, I think collaboration is a critical ingredient. I mean, for so many years, you look at the collaboration is gone. Just take between between the universities, my own Stanford and Silicon Valley and how that collaboration has developed over time and lead the founding of great companies, but also collaboration within the valley. This is the place to be a technology person in the whole world it's the best place partly because of this collaboration, and this innovative spirit that really is a core part of what we are as a place. >> I agree. The innovative spirit is one of the things that I enjoy, about not only being in technology, but also living in Silicon Valley. You can't go to a Starbucks without hearing a conversation or many conversations about new startups or cloud technology. So the innovative spirit is pervasive here. And it's also one that I find in an in an environment like ASG SV. You just hear a lot of inspiring stories and I was doing some research on them in the last 18 months. Five CEO positions have been seated and materialized through ACG SV. Number of venture deals initiated several board positions. So a lot of opportunity in this group here tonight. >> Right, well I think that's important because so much of the leadership has got to come by recruiting new young people. And with the increase in concerned about diversity and our leadership core and our boards, I think building that network out and trying to stretch it a little bit from the from perhaps the old boys network of an earlier time in the Valley is absolutely crucial. >> Couldn't agree more. So let's now talk a little bit about the Knight-Hennessy Scholars Program at Stanford. Tell us a little bit about it. When was it founded? >> So we are we are in our very first year, actually, this year, our first year of scholars, we founded it in 2016. The motivation was, I think, an increasing gap we perceived in terms of the need for great leadership and what was available. And it was in government. It was in the nonprofit world, it was in the for profit world. So I being a lifelong educator said, What can we do about this? Let's try to recruit and develop a core of younger people who show that they're committed to the greater good and who are excellent, who are innovative, who are creative, and prepare them for leadership roles in the future. >> So you're looking for are these undergraduate students? >> They are graduate students, so they've completed their undergraduate, it's a little hard to tell when somebody's coming out of high school, what their civic commitment is, what their ability to lead is. But coming out of coming out of undergraduate experience, and often a few years of work experience, we can tell a lot more about whether somebody has the potential to be a future leader. >> So you said, found it just in 2016. And one of the things I saw that was very interesting is projecting in the next 50 years, there's going to be 5000 Knight-Hennessy scholars at various stages of their careers and government organizations, NGOs, as you mentioned, so looking out 50 years you have a strong vision there, but really expect this organization to be able to make a lasting impact. >> That's what our goal is lasting impact over decades, because people who go into leadership positions often take a decade or two to rise to that position. But that's what our investment is our investment is in the in the future. And when I went to Phil Knight who's my co-founder and donor, might lead donor to the program, he was enthusiastic. His view was that we had a we had a major gap in leadership. And we needed to begin training, we need to do multiple things. We need to do things like we're doing tonight. But we also need to think about that next younger generation is up and coming. >> Some terms of inspiring the next generation of innovative diversity thinkers. Talk to me about some of the things that this program is aimed at, in addition to just, you know, some of the knowledge about leadership, but really helping them understand this diverse nature in which we now all find ourselves living. >> So one of the things we do is we try to bring in leaders from all different walks of life to meet and have a conversation with our scholars. This morning, we had the UN High Commissioner for Human Rights in town, Michelle Bachelet, and she sat down and talked about how she thought about her role as addressing human rights, how to move things forward in very complex situations we face around the world with collapse of many governments and many human rights violations. And how do you how do you make that forward progress with a difficult problem? So that kind of exposure to leaders who are grappling with really difficult problems is a critical part of our program. >> And they're really seeing and experiencing real world situations? >> Absolutely. They're seeing them up close as they're really occurring. They see the challenges we had, we had Governor Brown and just before he went out of office here in California, to talk about criminal justice reform a major issue in California and around the country. And how do we make progress on that on that particular challenge? >> So you mentioned a couple of other leaders who the students I've had the opportunity to learn from and engage with, but you yourself are quite the established leader. You went to Stanford as a professor in 1977. You are a President Emeritus you were president of Stanford from 2000 to 2016. So these students also get the opportunity to learn from all that you have experienced as it as a professor of Computer Science, as well as in one of your current roles as chairman of Alphabet. Talk to us a little bit about just the massive changes that you have seen, not just in Silicon Valley, but in technology and innovation over the last 40 plus years. >> Well, it is simply amazing. When I arrived at Stanford, there was no internet. The ARPANET was in its young days, email was something that a bunch of engineers and scientists use to communicate, nobody else did. I still remember going and seeing the first demonstration of what would become Yahoo. Well, while David Filo and Jerry Yang had it set up in their office. And the thing that immediately convinced me Lisa was they showed me that their favorite Pizza Parlor would now allow orders to go online. And when I saw that I said, the World Wide Web is not just about a bunch of scientists and engineers exchanging information. It's going to change our lives and it did. And we've seen wave after wave that with Google and Facebook, social media rise. And now the rise of AI I mean this this is a transformative technology as big as anything I think we've ever seen. In terms of its potential impact. >> It is AI is so transformative. I was I was in Hawaii recently on vacation and Barracuda Networks was actually advertising about AI in Hawaii and I thought that's interesting that the people that are coming to to Hawaii on vacation, presumably, people have you know, many generations who now have AI as a common household word may not understand the massive implications and opportunities that it provides. But it is becoming pervasive at every event we're at at the Cube and there's a lot of opportunity there. It's it's a very exciting subject. Last question for you. You mentioned that this that the Knight-Hennessy Scholars Program is really aimed towards graduate students. What is your advice to those BB stem kids in high school right now who are watching this saying, oh, John, what, what? How do you advise me to be able to eventually get into a program like this? >> Well, I think it begins by really finding your passion, finding something you're really dedicated to pushing yourself challenging yourself, showing that you can do great things with it. And then thinking about the bigger role you want to have with technology. In the after all, technology is not an end in itself. It's a tool to make human lives better and that's the sort of person we're looking for in the knight-Hennessy Scholars Program, >> Best advice you've ever gotten. >> Best advice ever gotten is remember that leadership is about service to the people in the institution you lead. >> It's fantastic not about about yourself but really about service to those. >> About service to others >> JOHN, it's been a pleasure having you on the Cube tonight we wish you the best of luck in your keynote at the 15th annual ACG SV Awards and we thank you for your time. >> Thank you, Lisa. I've enjoyed it. Lisa Martin, you're watching the Cube on the ground. Thanks for watching. (upbeat tech music)
SUMMARY :
Brought to you by ACG SV. and the co-founder of the So I was doing some research on you. so I thought I was pretty cool I'm going to sit down that the audience is going to hear from you tonight. And the role that technology plays in our lives the ACG SV Is has been This is the place to be a technology person is one of the things that I enjoy, because so much of the leadership the Knight-Hennessy Scholars Program at Stanford. the need for great leadership it's a little hard to tell And one of the things I saw and donor, might lead donor to the program, in addition to just, you know, So one of the things we do They see the challenges we had, we had Governor Brown just the massive changes that you have seen, And the thing that immediately convinced me Lisa was that the people that are coming and that's the sort of person we're looking for service to the people in the institution you lead. but really about service to those. and we thank you for your time. the Cube on the ground.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Marc Andreessen | PERSON | 0.99+ |
2016 | DATE | 0.99+ |
Michelle Bachelet | PERSON | 0.99+ |
John Hennessy | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Hawaii | LOCATION | 0.99+ |
California | LOCATION | 0.99+ |
2000 | DATE | 0.99+ |
1977 | DATE | 0.99+ |
John | PERSON | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
Alphabet | ORGANIZATION | 0.99+ |
Jerry Yang | PERSON | 0.99+ |
David Filo | PERSON | 0.99+ |
ORGANIZATION | 0.99+ | |
ORGANIZATION | 0.99+ | |
first year | QUANTITY | 0.99+ |
Lisa | PERSON | 0.99+ |
ACG SV | ORGANIZATION | 0.99+ |
Yahoo | ORGANIZATION | 0.99+ |
50 years | QUANTITY | 0.99+ |
Phil Knight | PERSON | 0.99+ |
Stanford | ORGANIZATION | 0.99+ |
Barracuda Networks | ORGANIZATION | 0.99+ |
Starbucks | ORGANIZATION | 0.99+ |
this year | DATE | 0.99+ |
Governor | PERSON | 0.99+ |
two | QUANTITY | 0.99+ |
JOHN | PERSON | 0.99+ |
Marc | PERSON | 0.99+ |
tonight | DATE | 0.98+ |
first time | QUANTITY | 0.98+ |
15th Annual ACG SV Awards | EVENT | 0.98+ |
Mountain View California | LOCATION | 0.98+ |
15th Annual Grow Awards | EVENT | 0.98+ |
This morning | DATE | 0.98+ |
one | QUANTITY | 0.97+ |
Five CEO | QUANTITY | 0.97+ |
today | DATE | 0.97+ |
a decade | QUANTITY | 0.96+ |
ACG SV | EVENT | 0.96+ |
over 230 attendees | QUANTITY | 0.95+ |
ACG SV Grow! Awards 2019 | EVENT | 0.95+ |
over 100 sea levels | QUANTITY | 0.95+ |
5000 | QUANTITY | 0.95+ |
ASG SV | ORGANIZATION | 0.94+ |
first demonstration | QUANTITY | 0.93+ |
Knight-Hennessy Scholars | ORGANIZATION | 0.92+ |
President | PERSON | 0.92+ |
15th annual ACG SV Awards | EVENT | 0.91+ |
UN | ORGANIZATION | 0.9+ |
last 40 plus years | DATE | 0.9+ |
last 18 months | DATE | 0.9+ |
Cube | COMMERCIAL_ITEM | 0.85+ |
ARPANET | ORGANIZATION | 0.85+ |
knight-Hennessy Scholars Program | TITLE | 0.85+ |
High Commissioner for Human Rights | PERSON | 0.84+ |
Knight-Hennessy Scholars Program | ORGANIZATION | 0.83+ |
Knight-Hennessy Scholars Program | TITLE | 0.81+ |
over decades | QUANTITY | 0.81+ |
Computer History Museum | LOCATION | 0.75+ |
Pizza Parlor | ORGANIZATION | 0.73+ |
Emeritus | PERSON | 0.7+ |
wave after | EVENT | 0.69+ |
wave | EVENT | 0.66+ |
Cube | ORGANIZATION | 0.65+ |
Stanford | LOCATION | 0.64+ |
Brown | PERSON | 0.63+ |
Zac Mutrux, Insperity | ACG SV Grow! Awards 2019
>> (Announcer) From Mountain View, California it's the Cube. Covering the 15th Annual Grow! Awards. Brought to you by ACG SV. >> I'm Lisa Martin with the Cube, on the ground at the Computer History Museum in Mountain View, California for the 15th Annual Grow! Awards for the Association of Corporate Growth Silicon Valley, ACG SV. That's a mouthful. I'm here with one of the ACG SV board members, Zac Mutrux, the District Sales Manager at Insperity. Hey, Zack, it's great to have you on the Cube. >> Thank you so much, Lisa, I'm pleased to be here. >> So before we talk about what you're doing here at the 15th Annual Grow! Awards, tell our audience about Insperity. I was reading, I love taglines. >> Yes. >> And I see on your homepage, Insperity is obsessed with delivery HR mastery. Wow. >> Oh yeah. >> Obsessed and mastery. Those two words just jumped out. Tell us a little bit about what you guys do. >> Impressive, isn't it? Well, we actually just adjusted our tagline to HR that makes a difference. And that's really what it's all about. We feel like companies that are growing, if they're going to make it from good to the best, it has everything to do with the people. Attracting the best people and keeping them, developing them over time, and that's exactly what we do with our clients. >> So Insperity has been in business since 1986, and if I think of today's modern workforce, highly mobile, distributed, there's the whole on-demand industry. You guys have seen a tremendous amount of change that now can be massively influenced, and your customers can, using technology. Give me a little bit of that historical perspective on Insperity's inception and today's workforce, and how you're helping them attract and retain the best talent. >> Oh, absolutely. Well, when the company started it was in a maybe a 200 square foot room with one telephone between the two co-founders. There's no such thing as email. So, absolutely, there's been immense technological changes and there continues to be. I think that's one of the things that has been responsible for Insperity's success is its adoption of technology. Today we are as much a technology company as we are an employee benefits company, or an HR consulting company. It's really about creating a positive experience for the employees. That's part of being a competitive employer. >> Well it has to be a positive experience, right? For your customers. Because acquiring great talent is one thing, retaining them is another. And I want to kind of pivot off the retention there for a second. As the District Sales Manager, I was asking you before we went live, tell me maybe one of your favorite stories, and you said, "Wow". One of the great things, you guys are coming off great growth and FY18 revenue growth. One of the great things that Insperity has been really successful at is customer retention. And that's hard. You're proud of this. Tell us about that statistic that you mentioned, and how it is that Insperity is evolving and innovating over the last few decades to keep that retention number as phenomenal as it is. >> Well, Insperity's been named one of the most admired corporations in the country, actually, five years in a row by Fortune magazine. And that's the kind of press that you can't buy. One of the accolades that I'm most proud of is that in the past year our own employees named us one of the top 100 companies to work for in the United States. Which is, I think, the proof that we really know what we're doing with our clients. Because there are a lot of different companies out there, various competitors, and almost none of them are on that list. So, it's living our values and expressing through our service team, our extraordinary service team, that, I think, keeps our clients coming back to us year after year. About 85% renew. That's been consistent. A high level of client retention for the past three years. Even more extraordinary is that we've been growing both top line and bottom line revenue at the same time. So there's just a testament to our leadership, to our co-founder and CEO, Paul Sarvadi, and to the best of team-- >> But it sounds like it's a lot of symbiotic relationships between the internal retention at Insperity that is maybe leading through to your customers seeing, hey, there's not a high turnover here. These people are doing, they love what they're doing. They're working for a good company. So there's probably a lot of symbiotic behaviors. >> Well, that's exactly right. I think you really hit the nail of the head. It's about culture. It's a culture that starts from the top with leadership, and it filters down throughout the organization. And we're not looking to do business with every single company. We're looking to do business with the companies that believe the things that we believe. That is, companies that have high levels of commitment, trust, communication. They do better financially then companies that don't have those things. >> And along those lines, mentioning just before we wrap here, we are at the 15th Annual ACG SV Awards tonight, where they're honoring two award winners. The Outstanding Growth Award winner is Arista Networks. And the Emerging Growth winner is Adesto Technologies. I'm excited to talk to them later. But I wanted to get a little bit of perspective on you've been involved as a board member of ACSG since last year. Tell me a little bit about what makes ACG SV worthy of your time. >> Oh, absolutely. That's a great question. It's just an extraordinary community, I think, of the top leaders in Silicon Valley come together. The monthly Key Notes add a lot of value. It's an intimate setting and there's real conversations that are taking place on topics that are relevant to today's professionals. So for me to be able to engage and hopefully add some value as a board member is privilege. >> And you can hear probably a lot of those conversations going on right behind Zac and me tonight. Zac, it's been a pleasure to have you on the Cube. Thank you so much for giving us some of your time. >> Oh, right, thank you, Lisa. >> For the Cube, I'm Lisa Martin on the ground. Thanks for watching. (pop electronic music)
SUMMARY :
Covering the 15th Annual Grow! Hey, Zack, it's great to have you on the Cube. at the 15th Annual Grow! And I see on your homepage, Tell us a little bit about what you guys do. and that's exactly what we do with our clients. Give me a little bit of that historical perspective and there continues to be. and innovating over the last few decades And that's the kind of press that you can't buy. that is maybe leading through to your customers seeing, It's a culture that starts from the top And the Emerging Growth winner is Adesto Technologies. of the top leaders in Silicon Valley come together. Zac, it's been a pleasure to have you on the Cube. For the Cube, I'm Lisa Martin on the ground.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Zac Mutrux | PERSON | 0.99+ |
Paul Sarvadi | PERSON | 0.99+ |
Zac | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Zack | PERSON | 0.99+ |
Arista Networks | ORGANIZATION | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
United States | LOCATION | 0.99+ |
ACSG | ORGANIZATION | 0.99+ |
five years | QUANTITY | 0.99+ |
Adesto Technologies | ORGANIZATION | 0.99+ |
ACG SV | ORGANIZATION | 0.99+ |
two words | QUANTITY | 0.99+ |
Mountain View, California | LOCATION | 0.99+ |
Insperity | ORGANIZATION | 0.99+ |
last year | DATE | 0.99+ |
200 square foot | QUANTITY | 0.99+ |
two co-founders | QUANTITY | 0.99+ |
one telephone | QUANTITY | 0.99+ |
1986 | DATE | 0.99+ |
Association of Corporate Growth Silicon Valley | ORGANIZATION | 0.99+ |
one | QUANTITY | 0.99+ |
15th Annual Grow! Awards | EVENT | 0.99+ |
15th Annual ACG SV Awards | EVENT | 0.98+ |
both | QUANTITY | 0.98+ |
Today | DATE | 0.98+ |
tonight | DATE | 0.97+ |
today | DATE | 0.97+ |
One | QUANTITY | 0.97+ |
About 85% | QUANTITY | 0.94+ |
two award winners | QUANTITY | 0.92+ |
past year | DATE | 0.91+ |
ACG SV Grow! Awards 2019 | EVENT | 0.89+ |
100 companies | QUANTITY | 0.84+ |
Computer History Museum | LOCATION | 0.77+ |
One of the accolades | QUANTITY | 0.74+ |
last few decades | DATE | 0.71+ |
one thing | QUANTITY | 0.71+ |
past three years | DATE | 0.68+ |
Fortune magazine | TITLE | 0.67+ |
top | QUANTITY | 0.62+ |
single company | QUANTITY | 0.61+ |
ACG | ORGANIZATION | 0.57+ |
Growth Award | TITLE | 0.55+ |
a second | QUANTITY | 0.52+ |
Insperity | PERSON | 0.52+ |
Cube | COMMERCIAL_ITEM | 0.51+ |
stories | QUANTITY | 0.48+ |
SV | EVENT | 0.46+ |
FY18 | DATE | 0.29+ |
Shampa Banerjee, PhD, Eros Digital | ACG SV Grow! Awards 2019
>> From Mountainview, California, it's The Cube, covering the 15th annual Grow! Awards. Brought to you by ACG SV. >> Hi, Lisa Martin on the ground with The Cube at the 15th annual ACG SV Grow! Awards. At the Computer History Museum in Mountainview, can you hear the buzz of 230 plus attendees behind me? I'm very pleased to welcome one of the ACG SV board members Dr. Shampa Banerjee, technology executive, and chief product officer at Eros Digital. Shampa, thank you so much for giving us some of your time this evening. >> Thank you, it's a pleasure. >> So lots of great, innovative, and inspiring conversations, no doubt, going on behind us. >> I'm trying to listen to it. >> Yeah, so talk to us a little bit about Eros Digital, who you are, what you do. >> So Eros International is the largest studio in India. It produces or distributes around 60 to 70 percent of the films made in India, Bollywood films. So I run the streaming platform, the Netflix for Bollywood, that's what I call it. >> The Netflix for Bollywood, I love it. Now, tell us more about that. >> So, you know, it's a streaming platform, a lot of the titles are from what we produce. A lot of the titles we lease from other production houses, and that is the entire technology platform, and then how do you get to the, we connect the consumers, rather, to the entertainment, right? So we like to help them discover, we help them indulge in the whole experience, and then as they keep coming to us more and more, we personalize the experience for them, so that's really what we give them. >> You know, personalization is so key. We expect it right in our lives, and whatever it is that we're doing, we're engaging with an Amazon or a Netflix or at Eros for example, we kind of now expect that. We're sort of demanding consumers, right? We expect them to know what I want, just what I want, don't give me any things that I don't want, so is that one of the things that you've seen, maybe surprising in your career, is this increasing demand for personalization? >> Absolutely, because, you know, there's so much content out there, so much information, and unless there's a filtering mechanism that makes sense for you, people don't want to, you know, it's very hard for them, so they want you to do the work for them. It's entertainment, right? So absolutely. Everyone kind of expects it. It's not said. It's not explicit, but that's the expectation. >> And obviously, with the goal of delighting and retaining those customers, you as the chief product officer have to listen and react to that. >> I spent, I'll tell you a short story. I spent once a month going through all the customers' comments in different platforms, right? And one of the stories I read was this 17-year-old French gal in Paris, she loves watching Bollywood because she was suffering from leukemia and after she gets a treatment, she comes home, she wants to watch something that makes her happy, and we had some issues with that subtitles, and she was having a problem watching our movies and she begged "Please bring them back". And I ran out of my office, went to my team, and I said, "Guys, this is who we wake up for every day. We give her joy, we give her pleasure." So to me, that's how listening to the customers to me is primary, to me they are my biggest stakeholder, and I've told the CEO and founder that, look at the end of the day, I leave and argue with you if it doesn't serve my customers. That's what I believe, listening to the customers, listening to them, understanding, of course, we do a lot of data collection and we look at what we are doing and the patterns, and based on that we make modifications, we test different things to see what makes sense, what's working, and what's not working, because people don't always tell you, and even if you ask them, they're shy to tell you. But then you can see what they're doing, and that's an indicator. >> Well that makes you feel really good, seeing and hearing and feeling the impact that you're making, and speaking of impact, you have been, in the last minute or so that we have, you've been on the board of ACG SV for about the last five years. We're here tonight to honor Arista as the Outstanding Growth Award winner and (mumbles) Technology as the Emerging Growth winner, but really quickly, what makes ACG SV worth your time? >> So ACG honestly is a fantastic organization and you know, living in the Bay Area, there are many organizations, there are many events that are always going on, you know. ACG has been a place where I've seen it's a very, very, very, very diverse organization, of course I still wish there were more females, you know, but it's a very diverse organization, people of all ages, people from different walks of life, from different kinds of companies, you know, and people are very, very collaborative and help each other to do business. I've become personal friends with many of them, but the main thing is, you know, you come here, if you're new to the Valley especially, whether as a company or as an individual, this is one of the best places to come to because it's not too large, it's not too small, it has the right number of people, and it helps you quickly on board. They'll introduce you to people, introduce you to events, they give you what you need to kind of get started. So to me it's like, when I joined, I joined before I was on the board, almost, I don't know, seven or eight years ago, and I've seen this whole thing transform and it's just an excellent, supportive, the people are very open-minded, great ideas, and it's just an excellent organization, love it. So it's worth my time, you know, to take the extra hours, and I would love to see it get even bigger and more diverse and more interesting. >> Well it sounds like, I love how you kind of described ACG SV as being that Goldilocks type of organization, not too big, not too small, just right, but we thank you so much. I wish we had more time to talk, as a female in technology, but we'll have to have you back at the studio on The Cube! >> Thank you so much. >> Thank you so much for your time. For The Cube, I am Lisa Martin. Thanks for watching. (music)
SUMMARY :
Brought to you by ACG SV. Shampa, thank you so much for giving us So lots of great, innovative, and inspiring who you are, what you do. So Eros International is the largest studio in India. Now, tell us more about that. and then how do you get to the, so is that one of the things that you've seen, so they want you to do the work for them. and retaining those customers, you as the chief and even if you ask them, they're shy to tell you. and (mumbles) Technology as the Emerging Growth winner, but the main thing is, you know, you come here, just right, but we thank you so much. Thank you so much for your time.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Lisa Martin | PERSON | 0.99+ |
India | LOCATION | 0.99+ |
Eros Digital | ORGANIZATION | 0.99+ |
Paris | LOCATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Netflix | ORGANIZATION | 0.99+ |
Bollywood | ORGANIZATION | 0.99+ |
ACG | ORGANIZATION | 0.99+ |
Eros International | ORGANIZATION | 0.99+ |
Shampa Banerjee | PERSON | 0.99+ |
Eros | ORGANIZATION | 0.99+ |
seven | DATE | 0.99+ |
Mountainview, California | LOCATION | 0.99+ |
tonight | DATE | 0.99+ |
Bay Area | LOCATION | 0.99+ |
Mountainview | LOCATION | 0.98+ |
ACG SV | ORGANIZATION | 0.98+ |
one | QUANTITY | 0.98+ |
Shampa | PERSON | 0.98+ |
230 plus attendees | QUANTITY | 0.98+ |
around 60 | QUANTITY | 0.98+ |
eight years ago | DATE | 0.97+ |
17-year-old | QUANTITY | 0.97+ |
Arista | PERSON | 0.96+ |
once a month | QUANTITY | 0.96+ |
The Cube | ORGANIZATION | 0.93+ |
ACG SV Grow! Awards 2019 | EVENT | 0.92+ |
French | OTHER | 0.91+ |
Dr. | PERSON | 0.85+ |
annual Grow! Awards | EVENT | 0.83+ |
15th annual ACG SV Grow! Awards | EVENT | 0.82+ |
Computer History Museum | LOCATION | 0.81+ |
70 percent | QUANTITY | 0.81+ |
last five years | DATE | 0.77+ |
leukemia | OTHER | 0.76+ |
Goldilocks | ORGANIZATION | 0.71+ |
this evening | DATE | 0.67+ |
Outstanding Growth Award | TITLE | 0.64+ |
15th | QUANTITY | 0.56+ |
Cube | PERSON | 0.53+ |
Narbeh Derhacobian, Adesto Technologies | ACG SV Grow! Awards 2019
>> from Mountain View, California It's the Cube covering the fifteenth annual Grow Awards. Brought to you by A C. G S V. >> Hi, Lisa Martin on the ground with the Cube at the Computer History Museum for the fifteenth annual TGS Grow Awards. Can you hear the collaboration of the innovation going on behind me? Very excited to welcome to the Cube, one of tonight's award winners from a Jaso Technologies, Norby, Jericho B and the President and CEO of Modesto. Congratulations on the Emerging Growth Award that adjuster has been honored with tonight. >> Thank you very much. We're very honored to be here. So you've been at >> the helm of a desert for a long time. I'd like our audience to hear a little bit from you about whom destiny is what you do. What makes you different. >> Perfect. So we are at a technology company on our products are used primarily in Internet of things, applications across many, many segments. Most off our businesses within the industrial segment on our customers use our products to actually build a Iot solutions for their end markets. Our products include semiconductor chips that are used at the edge of Coyote EJ gateway devices that connects the local networks to the more broad networks on. Basically, we enable our customers to take data from the physical world and send it up into the clouds >> to you guys. Our have had a great great trajectory, obviously being recognized by the emerging growth winner from a C. G S B. Tell me a little bit about it was looking at some information from you guys and on twenty eighteen, You guys did a great job of executing on your strategic initiatives to really make twenty eighteen a transformative year couple of acquisitions to us about the last year, in particular in the group that you have seen the momento and you're bringing into twenty nineteen. >> Correct? Correct. So we started. We enter twenty eighteen as a provider up application specific memory devices for I ot however, we realize that for our customers to take true benefit off the technologies we provide, we need to be a more holistic supplier of solutions. So as a result, we went through a whole process off looking at other technologies that can complement what we have in a very similar way, with strategic focus in the markets that we were focused, and as a result, we made two acquisitions in past summer that ended up its expanding our market opportunity, broadening our reach within existing customer and significantly expanding our offering portfolio to foreign markets. >> Negroes have a really strong position with tear one customers in the industrial sector. You mentioned that expecting Don't be a little bit more than about your leadership here in what makes these large industrial cheer. One players say Augusto is for us, >> right? So before I asked her that let me talk a little bit about the difference between industrial I ot and Consumer >> Riley's Yes, >> So if you think about consumer, I ot, it's what grabs headlines. It's the fitness trackers, the latest home smart thermostats, and the smartwatch is on so forth. The's are new markets. Volumes are girl very fast, but if next year and new shiny object is created, it's easy for the consumers to replace. They basically buy the new one. Repent replaced the old. One interesting thing about industrial I ot is that industrial I ot has this fragmented legacy systems that today run in their businesses. So if you look at the building we're in Today there is a fired and safety system that runs there's H Vac system that runs the business. There's a security systems, and this could have been installed here decades ago. There are billions of connected things in that industrial network today, but the data is unable to go up into the cloud. Where come cloud providers? Aye, aye. Providers can actually take the data on provide benefits to the business owners. We understand the language of industrial I ot very well because off our roots in that space. And we also understand this universe very well because of our roots being in Silicon Valley. So for industrial customers to benefit from this transformation, it's very important to be able to understand the OT world operational technology world of old days on the IT world that we're very familiar with. So with addition off these acquisitions that we've done this summer very well, positions with the building blocks that way can put together on offer differentiated solutions to our customers? >> Well, no, but it's been a pleasure having you on the queue. But the fifteenth annual acey GSP grow words. Congratulations to adjust of your whole team for the emerging growth award. And we look forward to seeing what happens this year in the space with you. Thank >> you. Thank you very much. Thank you. >> Lisa. Martin, you're watching the Cube. Thanks for watching.
SUMMARY :
Brought to you by A C. Hi, Lisa Martin on the ground with the Cube at the Computer History Museum for the fifteenth Thank you very much. I'd like our audience to hear a little bit from you about whom destiny is into the clouds to you guys. in the markets that we were focused, and as a result, we made two acquisitions in past Negroes have a really strong position with tear one customers in the industrial the consumers to replace. But the fifteenth annual acey GSP grow words. Thank you very much. Thanks for watching.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Silicon Valley | LOCATION | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Jaso Technologies | ORGANIZATION | 0.99+ |
Mountain View, California | LOCATION | 0.99+ |
Adesto Technologies | ORGANIZATION | 0.99+ |
last year | DATE | 0.99+ |
Today | DATE | 0.99+ |
Narbeh Derhacobian | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
decades ago | DATE | 0.99+ |
next year | DATE | 0.99+ |
Coyote | ORGANIZATION | 0.98+ |
billions | QUANTITY | 0.98+ |
Modesto | ORGANIZATION | 0.98+ |
twenty eighteen | QUANTITY | 0.97+ |
today | DATE | 0.97+ |
one | QUANTITY | 0.97+ |
this year | DATE | 0.96+ |
tonight | DATE | 0.95+ |
A C. G S V. | PERSON | 0.94+ |
one customers | QUANTITY | 0.94+ |
two acquisitions | QUANTITY | 0.93+ |
Martin | PERSON | 0.92+ |
past summer | DATE | 0.91+ |
TGS Grow Awards | EVENT | 0.91+ |
fifteenth | QUANTITY | 0.9+ |
Riley | PERSON | 0.9+ |
One players | QUANTITY | 0.9+ |
twenty nineteen | QUANTITY | 0.9+ |
ACG SV Grow! Awards 2019 | EVENT | 0.9+ |
Norby | ORGANIZATION | 0.89+ |
Augusto | ORGANIZATION | 0.88+ |
Grow Awards | EVENT | 0.88+ |
this summer | DATE | 0.83+ |
One interesting thing | QUANTITY | 0.74+ |
Emerging Growth Award | TITLE | 0.74+ |
Jericho B | PERSON | 0.73+ |
fifteenth annual | QUANTITY | 0.65+ |
Cube | TITLE | 0.59+ |
Vac | OTHER | 0.59+ |
Cube | ORGANIZATION | 0.59+ |
History Museum | ORGANIZATION | 0.53+ |
ot | ORGANIZATION | 0.5+ |
CEO | PERSON | 0.49+ |
Computer | LOCATION | 0.48+ |
annual | EVENT | 0.32+ |
Basil Faruqui, BMC | theCUBE NYC 2018
(upbeat music) >> Live from New York, it's theCUBE. Covering theCUBE New York City 2018. Brought to you by SiliconANGLE Media and its ecosystem partners. >> Okay, welcome back everyone to theCUBE NYC. This is theCUBE's live coverage covering CubeNYC Strata Hadoop Strata Data Conference. All things data happen here in New York this week. I'm John Furrier with Peter Burris. Our next guest is Basil Faruqui lead solutions marketing manager digital business automation within BMC returns, he was here last year with us and also Big Data SV, which has been renamed CubeNYC, Cube SV because it's not just big data anymore. We're hearing words like multi cloud, Istio, all those Kubernetes. Data now is so important, it's now up and down the stack, impacting everyone, we talked about this last year with Control M, how you guys are automating in a hurry. The four pillars of pipelining data. The setup days are over; welcome to theCUBE. >> Well thank you and it's great to be back on theCUBE. And yeah, what you said is exactly right, so you know, big data has really, I think now been distilled down to data. Everybody understands data is big, and it's important, and it is really you know, it's quite a cliche, but to a larger degree, data is the new oil, as some people say. And I think what you said earlier is important in that we've been very fortunate to be able to not only follow the journey of our customers but be a part of it. So about six years ago, some of the early adopters of Hadoop came to us and said that look, we use your products for traditional data warehousing on the ERP side for orchestration workloads. We're about to take some of these projects on Hadoop into production and really feel that the Hadoop ecosystem is lacking enterprise-grade workflow orchestration tools. So we partnered with them and some of the earliest goals they wanted to achieve was build a data lake, provide richer and wider data sets to the end users to be able to do some dashboarding, customer 360, and things of that nature. Very quickly, in about five years time, we have seen a lot of these projects mature from how do I build a data lake to now applying cutting-edge ML and AI and cloud is a major enabler of that. You know, it's really, as we were talking about earlier, it's really taking away excuses for not being able to scale quickly from an infrastructure perspective. Now you're talking about is it Hadoop or is it S3 or is it Azure Blob Storage, is it Snowflake? And from a control-end perspective, we're very platform and technology agnostic, so some of our customers who had started with Hadoop as a platform, they are now looking at other technologies like Snowflake, so one of our customers describes it as kind of the spine or a power strip of orchestration where regardless of what technology you have, you can just plug and play in and not worry about how do I rewire the orchestration workflows because control end is taking care of it. >> Well you probably always will have to worry about that to some degree. But I think where you're going, and this is where I'm going to test with you, is that as analytics, as data is increasingly recognized as a strategic asset, as analytics increasingly recognizes the way that you create value out of those data assets, and as a business becomes increasingly dependent upon the output of analytics to make decisions and ultimately through AI to act differently in markets, you are embedding these capabilities or these technologies deeper into business. They have to become capabilities. They have to become dependable. They have to become reliable, predictable, cost, performance, all these other things. That suggests that ultimately, the historical approach of focusing on the technology and trying to apply it to a periodic or series of data science problems has to become a little bit more mature so it actually becomes a strategic capability. So the business can say we're operating on this, but the technologies to take that underlying data science technology to turn into business operations that's where a lot of the net work has to happen. Is that what you guys are focused on? >> Yeah, absolutely, and I think one of the big differences that we're seeing in general in the industry is that this time around, the pull of how do you enable technology to drive the business is really coming from the line of business, versus starting on the technology side of the house and then coming to the business and saying hey we've got some cool technologies that can probably help you, it's really line of business now saying no, I need better analytics so I can drive new business models for my company, right? So the need for speed is greater than ever because the pull is from the line of business side. And this is another area where we are unique is that, you know, Control M has been designed in a way where it's not just a set of solutions or tools for the technical guys. Now, the line of business is getting closer and closer, you know, it's blending into the technical side as well. They have a very, very keen interest in understanding are the dashboards going to be refreshed on time? Are we going to be able to get all the right promotional offers at the right time? I mean, we're here at NYC Strata, there's a lot of real-time promotion happening here. The line of business has direct interest in the delivery and the timing of all of this, so we have always had multiple interfaces to Control M where a business user who has an interest in understanding are the promotional offers going to happen at the right time and is that on schedule? They have a mobile app for them to do that. A developer who's building up complex, multi-application platform, they have an API and a programmatic interface to do that. Operations that has to monitor all of this has rich dashboards to be able to do that. That's one of the areas that has been key for our success over the last couple decades, and we're seeing that translate very well into the big data place. >> So I just want to go under the hood for a minute because I love that answer. And I'd like to pivot off what Peter said, tying it back to the business, okay, that's awesome. And I want to learn a little bit more about this because we talked about this last year and I kind of am seeing it now. Kubernetes and all this orchestration is about workloads. You guys nailed the workflow issue, complex workflows. Because if you look at it, if you're adding line of business into the equation, that's just complexity in and of itself. As more workflows exist within its own line of business, whether it's recommendations and offers and workflow issues, more lines of business in there is complex for even IT to deal with, so you guys have nailed that. How does that work? Do you plug it in and the lines of businesses have their own developers, so the people who work with the workflows engage how? >> So that's a good question, with sort of orchestration and automation now becoming very, very generic, it's kind of important to classify where we play. So there's a lot of tools that do release and build automation. There's a lot of tools that'll do infrastructure automation and orchestration. All of this infrastructure and release management process is done ultimately to run applications on top of it, and the workflows of the application need orchestration and that's the layer that we play in. And if you think about how does the end user, the business and consumer interact with all of this technology is through applications, k? So the orchestration of the workflow's inside the applications, whether you start all the way from an ERP or a CRM and then you land into a data lake and then do an ML model, and then out come the recommendations analytics, that's the layer we are automating today. Obviously, all of this-- >> By the way, the technical complexity for the user's in the app. >> Correct, so the line of business obviously has a lot more control, you're seeing roles like chief digital officers emerge, you're seeing CTOs that have mandates like okay you're going to be responsible for all applications that are facing customer facing where the CIO is going to take care of everything that's inward facing. It's not a settled structure or science involved. >> It's evolving fast. >> It's evolving fast. But what's clear is that line of business has a lot more interest and influence in driving these technology projects and it's important that technologies evolve in a way where line of business can not only understand but take advantage of that. >> So I think it's a great question, John, and I want to build on that and then ask you something. So the way we look at the world is we say the first fifty years of computing were known process, unknown technology. The next fifty years are going to be unknown process, known technology. It's all going to look like a cloud. But think about what that means. Known process, unknown technology, Control M and related types of technologies tended to focus on how you put in place predictable workflows in the technology layer. And now, unknown process, known technology, driven by the line of business, now we're talking about controlling process flows that are being created, bespoke, strategic, differentiating doing business. >> Well, dynamic, too, I mean, dynamic. >> Highly dynamic, and those workflows in many respects, those technologies, piecing applications and services together, become the process that differentiates the business. Again, you're still focused on the infrastructure a bit, but you've moved it up. Is that right? >> Yeah, that's exactly right. We see our goal as abstracting the complexity of the underlying application data and infrastructure. So, I mean, it's quite amazing-- >> So it could be easily reconfigured to a business's needs. >> Exactly, so whether you're on Hadoop and now you're thinking about moving to Snowflake or tomorrow something else that comes up, the orchestration or the workflow, you know, that's as a business as a product that's our goal is to continue to evolve quickly and in a manner that we continue to abstract the complexity so from-- >> So I've got to ask you, we've been having a lot of conversations around Hadoop versus Kubernetes on multi cloud, so as cloud has certainly come in and changed the game, there's no debate on that. How it changes is debatable, but we know that multiple clouds is going to be the modus operandus for customers. >> Correct. >> So I got a lot of data and now I've got pipelining complexities and workflows are going to get even more complex, potentially. How do you see the impact of the cloud, how are you guys looking at that, and what are some customer use cases that you see for you guys? >> So the, what I mentioned earlier, that being platform and technology agnostic is actually one of the unique differentiating factors for us, so whether you are an AWS or an Azure or a Google or On-Prem or still on a mainframe, a lot of, we're in New York, a lot of the banks, insurance companies here still do some of the most critical processing on the mainframe. The ability to abstract all of that whether it's cloud or legacy solutions is one of our key enablers for our customers, and I'll give you an example. So Malwarebytes is one of our customers and they've been using Control M for several years. Primarily the entire structure is built on AWS, but they are now utilizing Google cloud for some of their recommendation analysis on sentiment analysis because their goal is to pick the best of breed technology for the problem they're looking to solve. >> Service, the best breed service is in the cloud. >> The best breed service is in the cloud to solve the business problem. So from Control M's perspective, transcending from AWS to Google cloud is completely abstracted for them, so runs Google tomorrow it's Azure, they decide to build a private cloud, they will be able to extend the same workflow orchestration. >> But you can build these workflows across whatever set of services are available. >> Correct, and you bring up an important point. It's not only being able to build the workflows across platforms but being able to define dependencies and track the dependencies across all of this, because none of this is happening in silos. If you want to use Google's API to do the recommendations, well, you've got to feed it the data, and the data's pipeline, like we talked about last time, data ingestion, data storage, data processing, and analytics have very, very intricate dependencies, and these solutions should be able to manage not only the building of the workflow but the dependencies as well. >> But you're defining those elements as fundamental building blocks through a control model >> Correct. >> That allows you to treat the higher level services as reliable, consistent, capabilities. >> Correct, and the other thing I would like to add here is not only just build complex multiplatform, multiapplication workflows, but never lose focus of the business service of the business process there, so you can tie all of this to a business service and then, these things are complex, there are problems, let's say there's an ETL job that fails somewhere upstream, Control M will immediately be able to predict the impact and be able to tell you this means the recommendation engine will not be able to make the recommendations. Now, the staff that's going to work under mediation understands the business impact versus looking at a screen where there's 500 jobs and one of them has failed. What does that really mean? >> Set priorities and focal points and everything else. >> Right. >> So I just want to wrap up by asking you how your talk went at Strata Hadoop Data Conference. What were you talking about, what was the core message? Was it Control M, was it customer presentations? What was the focus? >> So the focus of yesterday's talk was actually, you know, one of the things is academic talk is great, but it's important to, you know, show how things work in real life. The session was focused on a real-use case from a customer. Navistar, they have IOT data-driven pipelines where they are predicting failures of parts inside trucks and buses that they manufacture, you know, reducing vehicle downtime. So we wanted to simulate a demo like that, so that's exactly what we did. It was very well received. In real-time, we spun up EMR environment in AWS, automatically provision control of infrastructure there, we applied spark and machine learning algorithms to the data and out came the recommendation at the end was that, you know, here are the vehicles that are-- >> Fix their brakes. (laughing) >> Exactly, so it was very, very well received. >> I mean, there's a real-world example, there's real money to be saved, maintenance, scheduling, potential liability, accidents. >> Liability is a huge issue for a lot of manufacturers. >> And Navistar has been at the leading edge of how to apply technologies in that business. >> They really have been a poster child for visual transformation. >> They sure have. >> Here's a company that's been around for 100 plus years and when we talk to them they tell us that we have every technology under the sun that has come since the mainframe, and for them to be transforming and leading in this way, we're very fortunate to be part of their journey. >> Well we'd love to talk more about some of these customer use cases. Other people love about theCUBE, we want to do more of them, share those examples, people love to see proof in real-world examples, not just talk so appreciate it sharing. >> Absolutely. >> Thanks for sharing, thanks for the insights. We're here Cube live in New York City, part of CubeNYC, we're getting all the data, sharing that with you. I'm John Furrier with Peter Burris. Stay with us for more day two coverage after this short break. (upbeat music)
SUMMARY :
Brought to you by SiliconANGLE Media with Control M, how you guys are automating in a hurry. describes it as kind of the spine or a power strip but the technologies to take that underlying of the house and then coming to the business You guys nailed the workflow issue, and that's the layer that we play in. for the user's in the app. Correct, so the line of business and it's important that technologies evolve in a way So the way we look at the world is we say that differentiates the business. of the underlying application data and infrastructure. so as cloud has certainly come in and changed the game, and what are some customer use cases that you see for the problem they're looking to solve. is in the cloud. The best breed service is in the cloud But you can build these workflows across and the data's pipeline, like we talked about last time, That allows you to treat the higher level services and be able to tell you this means the recommendation engine So I just want to wrap up by asking you at the end was that, you know, Fix their brakes. there's real money to be saved, And Navistar has been at the leading edge of how They really have been a poster child for and for them to be transforming and leading in this way, people love to see proof in real-world examples, Thanks for sharing, thanks for the insights.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
John | PERSON | 0.99+ |
Basil Faruqui | PERSON | 0.99+ |
Peter Burris | PERSON | 0.99+ |
BMC | ORGANIZATION | 0.99+ |
Peter | PERSON | 0.99+ |
500 jobs | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
New York | LOCATION | 0.99+ |
last year | DATE | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
New York City | LOCATION | 0.99+ |
SiliconANGLE Media | ORGANIZATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
Hadoop | TITLE | 0.99+ |
first fifty years | QUANTITY | 0.99+ |
theCUBE | ORGANIZATION | 0.99+ |
Navistar | ORGANIZATION | 0.99+ |
tomorrow | DATE | 0.98+ |
yesterday | DATE | 0.98+ |
one | QUANTITY | 0.98+ |
this week | DATE | 0.97+ |
Malwarebytes | ORGANIZATION | 0.97+ |
Cube | ORGANIZATION | 0.95+ |
Control M | ORGANIZATION | 0.95+ |
NYC | LOCATION | 0.95+ |
Snowflake | TITLE | 0.95+ |
Strata Hadoop Data Conference | EVENT | 0.94+ |
100 plus years | QUANTITY | 0.93+ |
CubeNYC Strata Hadoop Strata Data Conference | EVENT | 0.92+ |
last couple decades | DATE | 0.91+ |
Azure | TITLE | 0.91+ |
about five years | QUANTITY | 0.91+ |
Istio | ORGANIZATION | 0.9+ |
CubeNYC | ORGANIZATION | 0.89+ |
day | QUANTITY | 0.87+ |
about six years ago | DATE | 0.85+ |
Kubernetes | TITLE | 0.85+ |
today | DATE | 0.84+ |
NYC Strata | ORGANIZATION | 0.83+ |
Hadoop | ORGANIZATION | 0.78+ |
one of them | QUANTITY | 0.77+ |
Big Data SV | ORGANIZATION | 0.75+ |
2018 | EVENT | 0.7+ |
Kubernetes | ORGANIZATION | 0.66+ |
fifty years | DATE | 0.62+ |
Control M | TITLE | 0.61+ |
four pillars | QUANTITY | 0.61+ |
two | QUANTITY | 0.6+ |
-Prem | ORGANIZATION | 0.6+ |
Cube SV | COMMERCIAL_ITEM | 0.58+ |
a minute | QUANTITY | 0.58+ |
S3 | TITLE | 0.55+ |
Azure | ORGANIZATION | 0.49+ |
cloud | TITLE | 0.49+ |
2018 | DATE | 0.43+ |
Paul Appleby, Kinetica | theCUBE NYC 2018
>> Live from New York, it's the Cube (funky music) covering the Cube New York City 2018 brought to you by SiliconANGLE Media and its ecosystem partners. (funky music) >> Everyone welcome back to theCUBE live in New York City for Cube NYC. This is our live broadcast - two days of coverage around the big data world, AI, the future of Cloud analytics. I'm John Furrier, my cohost Peter Burris. Our next guest is Paul Appleby, CEO Kinetica. Thanks for coming back to theCUBE - good to see you. >> Great to be back again and great to visit in New York City - it's incredible to be here on this really important week. >> Last time we chatted was in our big data Silicon Valley event, which is going to be renamed Cube SV, because it's not just data anymore; there's a lot of Cloud involved, a lot of new infrastructure. But analytics has certainly changed. What's your perspective now in New York as you're in here hearing all the stories around the show and you talk to customers - what's the update from your perspective? Because certainly we're hearing a lot of Cloud this year - Cloud, multi Cloud, analytics, and eyeing infrastructure, proof in the pudding, that kind of thing. >> I'm going to come back to the Cloud thing because I think that's really important. We have shifted to this sort of hybrid multi Cloud world, and that's our future - there is no doubt about it, and that's right across all spectre of computing, not just as it relates to data. But I think this evolution of data has continued this journey that we've all been on from whatever you want to call it - systems or record - to the world of big data where we're trying to gain insights out of this massive oceans of data. But we're in a world today where we're leveraging the power of analytics and intelligence, AI machine learning, to make fundamental decisions that drive some action. Now that action may be to a human to make a decision to interact more effectively with a customer, or it could be to a machine to automate some process. And we're seeing this fundamental shift towards a focus on that problem, and associated with that, we're leveraging the power of Cloud, AI, ML, and all the rest of it. >> And the human role in all this has been talked about. I've seen in the US in the political landscape, data for good, we see Facebook up there being basically litigated publicly in front of the Senate around the role of data and the elections. People are talking in the industry about the role of humans with machines is super important. This is now coming back as a front and center issue of hey, machines do great intelligence, but what about the human piece? What's your view on the human interaction component, whether it's the curation piece, the role of the citizen analyst, or whatever we're calling it these days, and what machines do to supplement that? >> Really good question - I've spent a lot of time thinking about this. I've had the incredible privilege of being able to attend the World Economic Forum for the last five years, and this particular topic of how Robotics Automation Artificial Intelligence machine learning is impacting economies, societies, and ultimately the nature of work has been a really big thread there for a number of years. I've formed a fundamental view: first of all, any technology can be used for good purposes and bad purposes, and it's - >> It always is. >> And it always is, and it's incumbent upon society and government to apply the appropriate levels of regulation, and for corporations to obviously behave the right way, but setting aside those topics - because we could spend hours talking about those alone - there is a fundamental issue, and this is this kind of conversation about what a lot of people like to describe as the fourth industrial revolution. I've spent a lot of time, because you hear people bandy that around - what do they really mean, and what are we really talking about? I've looked at every point in time where there's been an industrial revolution - there's been a fundamental shift of work that was done by humans that's now done by machines. There's been a societal uproar, and there're being new forms of work created, and society's evolved. What I look at today is yes, there's a responsibility and a regular treaside to this, but there's also a responsibility in business and society to prepare our workers and our kids for new forms of work, cause that's what I really think we should be thinking about - what are the new forms of work that are actually unlocked by these technologies, rather than what are the roles that are displaced by this steam powered engine. (laughs softly) >> Well, Paul, we totally agree with you. There's one other step in this process. It kind of anticipates each of these revolutions, and that is there is a process of new classes of asset formation. Mhm. So if you go back to when we put new power trains inside row houses to facilitate the industrial revolution in the early 1800s, and you could say the same thing about transportation, and what the trains did and whatnot. There's always this process of new asset formation that presaged some of these changes. Today it's data - data's an asset cause businesses ultimately institutionalize, or re institutionalize, their work around what they regard as valuable. Now, when we start talking about machines telling other machines what to do, or providing options or paring off options for humans so they have clear sets of things that they can take on, speed becomes a crucial issue, right? At the end of the day, all of this is going to come back to how fast can you process data? Talk to us a little bit about how that dynamic and what you guys are doing to make it possible is impacting business choices. >> Two really important things to unpack there, and one I think I'd love to touch on later, which is data as an asset class and how corporations should treat data. You talk about speed, and I want to talk about speed in the context of perishability, because the truth is if you're going to drive these incredible insights, whether it's related to a cyber threat, or a terrorist threat, or an opportunity to expand your relationship with a customer, or to make a critical decision in a motor vehicle in an autonomous operating mode, these things are about taking massive volumes of streaming data, running analytics in real time, and making decisions in real time. These are not about gleaning insights from historic pools or oceans of data; this is about making decisions that are fundamental to - >> Right now. >> The environment that you're in right now. You think about the autonomous car - great example of the industrial Internet, one we all love to talk about. The mechanical problems associated with autonomy have been solved, fundamentally sensors in cars, and the automated processes related to that. The decisioning engines - they need to be applied at scale in millions of vehicles in real time. That's an extreme data problem. The biggest problem solved there is data, and then over time, societal and regulatory change means that this is going to take some time before it comes to fruition. >> We were just saying - I think it was 100 Teslas generating 100 terabytes of data a day based on streams from its fleet of cars its customers have. >> We firmly believe that longer term, when you get to true autonomy, each car will probably generate around ten terabytes of data a day. That is an extremely complex problem to solve, because at the end of the day, this thinking that you're able to drive that data back to some centralized brain to be making those decisions for and on behalf of the cars is just fundamentally flawed. It has to happen in the car itself. >> Totally agree. >> This is putting super computers inside cars. >> Which is kind of happening - in fact, that 100 terabytes a day is in fact the data that does get back to Tesla. >> Yeah. >> As you said, there's probably 90% of the data is staying inside the car, which is unbelievable scale. >> So the question I wanted to ask you - you mentioned the industrial revolution, so every time there's a new revolution, there's an uproar, you mentioned. But there's also a step up of new capabilities, so if there's new work being developed, usually entrepreneur activity - weird entrepreneurs figured out that everyone says they're not weird anymore; it's great. But there's a step up of new capability that's built. Someone else says hey, the way we used to do databases and networks was great for moving one gig Ethernet on top of the rack; now you got 10 terabytes coming off a car or wireless spectrum. We got to rethink spectrum, or we got to rethink database. Let's use some of these GPUs - so a new step up of suppliers have to come in to support the new work. What's your vision on some of those things that are happening now - that you think people aren't yet seeing? What are some of those new step up functions? Is it on the database side, is it on the network, is it on the 5G - where's the action? >> Wow. Because who's going to support the Teslas? (Paul laughs) Who's going to support the new mobile revolution, the new iPhones the size of my two hands put together? What's your thoughts on that? >> The answer is all of the above. Let me talk about that and what I mean by that. Because you're looking at it from the technology perspective, I'd love to come back and talk about the human perspective as well, but from the technology perspective, of course leveraging power is going to be fundamental to this, because if you think about the types of use cases where you're going to have to be gigathreading queries against massive volumes of data, both static and streaming, you can't do that with historic technology, so that's going to be a critical part of it. The other part of it that we haven't mentioned a lot here but I think we should bring into it is if you think about these types of industrial Internet use cases, or IOT - even consumer Internet IOT related use cases - a lot of the decisioning has to occur out of the H. It cannot occur in a central facility, so it means actually putting the AI or ML engine inside the vehicle, or inside the cell phone tower, or inside the oil rig, and that is going to be a really big part of you know, shifting back to this very distributive model of machine lining in AI, which brings very complex questions in of how you drive governance - (John chuckles) >> And orchestration around employing Ai and ML models at massive scale, out to edge devices. >> Inferencing at the edge, certainly. It's going to be interesting to see what happens with training - we know that some of the original training will happen at the center, but some of that maintenance training? It's going to be interesting to see where that actually - it's probably going to be a split function, but you're going to need really high performing databases across the board, and I think that's one of the big answers, John, is that everybody says oh, it's all going to be in software. It's going to be a lot of hard word answers. >> Yep. >> Well the whole idea is just it's provocative to think about it and also intoxicating if you also want to go down that rabbit hole... If you think about that car, okay, if they're going to be doing century machine learning at the edge - okay, what data are you working off of? There's got to be some storage, and then what about real time data coming from other either horizontally scalable data sets. (laughs) So the question is, what do they have access to? Are they optimized for the decision making at that time? >> Mhm. >> Again, talk about the future of work - this is a big piece, but this is the human piece as well. >> Yeah. >> Are our kids going to be in a multi massive, multi player online game called Life? >> They are. >> They are now. They're on Fortnite, they're on Call of Duty, and all this gaming culture. >> But I think this is one of the interesting things, because there's a very strong correlation between information theory and thermodynamics. >> Mhm. >> They're the same exact - in physics, they are the identical algorithms and the identical equations. There's not a lot of difference, and you go back to the original revolution, you have a series of row houses, you put a power supply all the way down, you can run a bunch of looms. The big issue is entropy - how much heat are you generating? How do you get greater efficiency out of that single power supply? Same thing today: we're worried about the amount of cost, the amount of energy, the amount of administrative overhead associated with using data as an asset, and the faster the database, the more natural it is, the more easy it is to administer, the more easy it is to apply to a lot of different cases, the better. And it's going to be very, very interesting over the next few year to see how - Does database come in memory? Does database stay out over there? A lot of questions are going to be answered in the next couple years as we try to think about where these information transducers actually reside, and how they do their job. >> Yeah, and that's going to be driven yes, partially by the technology, but more importantly by the problems that we're solving. Here we are in New York City - you look at financial services. There are two massive factors in financial services going on what is the digital bank of the future look like, and how the banks interact with their customers, and how you get that true one-to-one engagement, which historically has been virtually impossible for companies that have millions or tens of millions of customers, so fundamental transformation of customer engagement driven by these advanced or excelerated analytics engines, and the pair of AI and ML, but then on the other side if you start looking at really incredibly important things for the banks like risk and spread, historically because of the volumes of data, it's been virtually impossible for them to present their employees with a true picture of those things. Now, with these accelerated technologies, you can take all the historic trading data, and all of the real time trading data, smash that together, and run real time analytics to make the right decisions in the moment of interaction with a customer, and that is incredibly powerful for both the customer, but also for the bank in mitigating risk, and they're the sorts of things we're doing with banks up and down the city here in New York, and of course, right around the world. >> So here's a question for you, so with that in mind - this is kind of more of a thought exercise - will banks even be around in 20 years? >> Wow. (laughs) >> I mean, you've got block chains saying we're going to have new crypto models here, if you take this Tesla with ten terabytes going out every second or whatever that number is. If that's the complex problem, banking should be really easy to solve. >> I think it's incumbent on boards in every industry, not just banking, to think about what existential threats exist, because there are incredibly powerful, successful companies that have gone out of existence because of fundamental shifts and buying behaviors or technologies - I think banks need to be concerned. >> Every industry needs to be concerned. >> Every industry needs to be concerned. >> At the end of the day, every board needs to better understand how they can reduce their assets specificities, right? How they can have their assets be more fungible and more applicable or appropriable to multiple different activities? Think about a future where data and digital assets are a dominant feature of business. Asset specificities go down; today their very definition of vertical industry is defined by the assets associated with bottling, the assets associated with flying, the assets associated with any number of other things. As aspect specialist needs to go down because of data, it changes even the definition of industry, let alone banking. >> Yeah, and auto industry's a great example. Will we own cars in the future? Will we confirm them as a service? >> Exactly. >> Car order manufacturers need to come to terms with that. The banks need to come to terms with the fact that the fundamental infrastructure for payments, whether it's domestic or global, will change. I mean, it is going to change. >> It's changing. It's changing. >> It has to change, and it's in the process of changing, and I'm not talking about crypto, you know, what form of digital currency exists in the future, we can argue about forever, but a fundamental underlying platform for real time exchange - that's just the future. Now, what does that mean for banks that rely heavily on payments as part of their core driver of profitability? Now that's a really important thing to come to terms with. >> Or going back to the point you made earlier. We may not have banks, but we have bankers. There's still going to be people who're providing advice in council, helping the folks understand what businesses to buy, what businesses to sell. So whatever industry they're in, we will still have the people that bring the extra taste to the data. >> Okay, we got to break it there, we've run out of time. Paul, love to chat further about future banking, all this other stuff, and also, as we live in a connected world, what does that mean? We're obviously connected to data; we certainly know there's gonnna be a ton of data. We're bringing that to you here, New York City, with Cube NYC. Stay with us for more coverage after the short break. (funky music)
SUMMARY :
brought to you by SiliconANGLE Media Thanks for coming back to theCUBE - good to see you. in New York City - it's incredible to be here around the show and you talk to customers - Now that action may be to a human to make a decision about the role of humans with machines is super important. to attend the World Economic Forum for the last and government to apply the appropriate levels At the end of the day, all of this is going to come back to and one I think I'd love to touch on later, and the automated processes related to that. based on streams from its fleet of cars because at the end of the day, a day is in fact the data that does get back to Tesla. is staying inside the car, which is unbelievable scale. So the question I wanted to ask you - Who's going to support the new mobile revolution, a lot of the decisioning has to occur out of the H. at massive scale, out to edge devices. It's going to be interesting to see what happens There's got to be some storage, and then what about Again, talk about the future of work - this is and all this gaming culture. But I think this is one of the interesting things, the more easy it is to administer, the more easy it is and all of the real time trading data, Wow. If that's the complex problem, or technologies - I think banks need to be concerned. the assets associated with bottling, Yeah, and auto industry's a great example. The banks need to come to terms with the fact It's changing. Now that's a really important thing to come to terms with. Or going back to the point you made earlier. We're bringing that to you here,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Paul | PERSON | 0.99+ |
Paul Appleby | PERSON | 0.99+ |
John | PERSON | 0.99+ |
Peter Burris | PERSON | 0.99+ |
New York | LOCATION | 0.99+ |
millions | QUANTITY | 0.99+ |
John Furrier | PERSON | 0.99+ |
Call of Duty | TITLE | 0.99+ |
Two | QUANTITY | 0.99+ |
New York City | LOCATION | 0.99+ |
10 terabytes | QUANTITY | 0.99+ |
Tesla | ORGANIZATION | 0.99+ |
Senate | ORGANIZATION | 0.99+ |
two hands | QUANTITY | 0.99+ |
two days | QUANTITY | 0.99+ |
Kinetica | ORGANIZATION | 0.99+ |
US | LOCATION | 0.99+ |
SiliconANGLE Media | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.99+ | |
ten terabytes | QUANTITY | 0.99+ |
iPhones | COMMERCIAL_ITEM | 0.99+ |
each car | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
Today | DATE | 0.98+ |
early 1800s | DATE | 0.98+ |
both | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
millions of vehicles | QUANTITY | 0.98+ |
Fortnite | TITLE | 0.97+ |
100 terabytes a day | QUANTITY | 0.97+ |
Cube | ORGANIZATION | 0.96+ |
each | QUANTITY | 0.96+ |
Silicon Valley | LOCATION | 0.96+ |
two massive factors | QUANTITY | 0.96+ |
NYC | LOCATION | 0.95+ |
World Economic Forum | EVENT | 0.95+ |
this year | DATE | 0.94+ |
100 terabytes of data | QUANTITY | 0.94+ |
around ten terabytes of data a day | QUANTITY | 0.93+ |
20 years | QUANTITY | 0.92+ |
next couple years | DATE | 0.91+ |
Teslas | ORGANIZATION | 0.91+ |
fourth industrial revolution | EVENT | 0.91+ |
100 | QUANTITY | 0.91+ |
one gig | QUANTITY | 0.88+ |
single power | QUANTITY | 0.88+ |
last five years | DATE | 0.86+ |
theCUBE | ORGANIZATION | 0.84+ |
Cube SV | COMMERCIAL_ITEM | 0.83+ |
CEO | PERSON | 0.83+ |
tens of millions of customers | QUANTITY | 0.81+ |
every second | QUANTITY | 0.8+ |
next few year | DATE | 0.79+ |
90% of the data | QUANTITY | 0.76+ |
a day | QUANTITY | 0.75+ |
Cube | COMMERCIAL_ITEM | 0.75+ |
Cube New York City 2018 | EVENT | 0.7+ |
2018 | DATE | 0.7+ |
one other step | QUANTITY | 0.67+ |
important things | QUANTITY | 0.59+ |
Life | TITLE | 0.55+ |
data | QUANTITY | 0.51+ |
5G | TITLE | 0.28+ |
Video Report Exclusive: @theCUBE report from ACG SV's GROW! Awards
Jeffrey Kier with the Qbert Computer History Museum in Mountain View California for the 14th annual association of corporate growth Silicon Valley grow Awards we've been here for a couple years now and it's a big event 300 people coming in to talk about really an ecosystem that helping other companies grow always great to be on the cube [Music] essentially what we are is an organization that's dedicated towards providing networking opportunities educational opportunities programming for c-level executives and other senior level executives at companies to help them develop their career and also grow their businesses tonight it's about tech as a force for good and I'm gonna talk about what I call the four superpowers today mobile unlimited reach cloud unlimited scale ai unlimited intelligence and IOT bridging from the digital to the physical world and how those four superpowers are reinforcing each other today very sophisticated population I mean it's just wonderful living in this seventy some people our biggest thing that we see is just the whole better together message that all of the resources from the strategically line businesses all working together to support the customers technology is evolving at a remarkable speed you know that's being driven largely by the availability of increased processing power less and less expensive faster and faster digital transformation IT transformation security transformation and work force transformation those are the big things for us this year it's great to be able to have a computer that really understands how to generate meaningful realistic text it's our opportunity to improve the quality of lives for every human on the planet as a result of those superpowers and really how it's our responsibility as a tech community to shape those superpowers for good there are issues created operationally day to day that we have to sort of always be on the watch for like you know readiness distance or these technologies it's the two sides of the same point always you can use it for good or you can use it for bad and unfortunately the bads within the news more than the good but there's so many exciting things going on in medicine health care oh yeah agriculture energy that the opportunities are almost endless not just the first world problems those of us here in the Silicon Valley see every day but really open our eyes to what's happening in other parts of the globe the need for water clean water water filtration clean air having access to information education so these are some things that are you know really personally dear to me in the last 50 years we've taken the extreme poverty rate from over 40 percent to less than 10 percent on the planet we've increased the length of life by almost 20 years these are stunning things and largely the result of the technological breakthroughs that we're doing that's the beauty of this right that's all of these things actually create opportunities you just have to stick with it and look at solutions and there's no shortage of really talented creative people to go address these opportunities and it's so fun to be involved in it right now the scale that we're able to now conduct business to be able to develop software to reach customers and truly write to change people's lives there are in many ways the technology halves and the technology have not absolutely and a lot of it is not just about making the product but then taking the product you've made and then implementing it in various use cases that really make a change from about in the world as I say today is the fastest day of tech evolution of your life it's also the slowest day of tech devolution of the rest of your life the rest of your life I'm Jeff Rick you're watching the cube from the a cts-v Awards thanks for watching [Music]
**Summary and Sentiment Analysis are not been shown because of improper transcript**
ENTITIES
Entity | Category | Confidence |
---|---|---|
Jeffrey Kier | PERSON | 0.99+ |
Jeff Rick | PERSON | 0.99+ |
Qbert Computer History Museum | ORGANIZATION | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
300 people | QUANTITY | 0.99+ |
less than 10 percent | QUANTITY | 0.99+ |
two sides | QUANTITY | 0.99+ |
over 40 percent | QUANTITY | 0.98+ |
this year | DATE | 0.97+ |
almost 20 years | QUANTITY | 0.96+ |
today | DATE | 0.96+ |
tonight | DATE | 0.96+ |
Mountain View California | LOCATION | 0.94+ |
Silicon Valley grow Awards | EVENT | 0.93+ |
last 50 years | DATE | 0.84+ |
@theCUBE | ORGANIZATION | 0.82+ |
first | QUANTITY | 0.81+ |
four superpowers | QUANTITY | 0.8+ |
14th annual | QUANTITY | 0.73+ |
a couple years | QUANTITY | 0.68+ |
seventy some people | QUANTITY | 0.67+ |
every human | QUANTITY | 0.59+ |
GROW! Awards | TITLE | 0.54+ |
cube | TITLE | 0.48+ |
ACG SV | TITLE | 0.42+ |
cts | TITLE | 0.27+ |
Jacques Nadeau, Dremio | Big Data SV 2018
>> Announcer: Live from San Jose, it's theCUBE, presenting Big Data Silicon Valley. Brought to you by SiliconANGLE Media and it's ecosystem partners. >> Welcome back to Big Data SV in San Jose. This theCUBE, the leader in live tech coverage. My name is Dave Vellante and this is day two of our wall-to-wall coverage. We've been here most of the week, had a great event last night, about 50 or 60 of our CUBE community members were here. We had a breakfast this morning where the Wikibon research team laid out it's big data forecast, the eighth big data forecast and report that we've put out, so check out that online. Jacques Nadeau is here. He is the CTO and co-founder of Dremio. Jacque, welcome to theCUBE, thanks for coming on. >> Thanks for having me here. >> So we were talking a little bit about what you guys do. Three year old company. Well, let me start. Why did you co-found Dremio? >> So, it was a very simple thing I saw, so, over the last ten years or so, we saw a regression in the ability for people to get at data, so you see all these really cool technologies that came out to store data. Data lakes, you know, SQL systems, all these different things that make developers very agile with data. But what we were also seeing was a regression in the ability for analysts and data consumers to get at that data because the systems weren't designed for analysts, they were designed for data producers and developers. And we said, you know what, there needs to be a way to solve this. We need to be able to empower people to be self-sufficient again at the data consumption layer. >> Okay, so you solved that problem how, you said, called it a self-service of a data platform. >> Yeah, yeah, so self-service data platform and the idea is pretty simple. It's that, no matter where the data is physically, people should be able to interact with a logical view of it. And so, we talk a little bit like it's Google Docs for your data. So people can go into the system, they can see the different data sets that are available to them, collaborate around those, create changes to those that they can then share with other people in the organization, always dealing with the logical layer and then, behind the scenes, we have physical capabilities to interact with all the different system we interact with. But that's something that business users shouldn't have to think as much about and so, if you think about how people interact with data today, it's very much about copies. So every time you want to do something, typically you're going to make a copy. I want to reshape the data, I make a copy. I want to make it go faster, I make a copy. And those copies are very, very difficult for people to manage and they could have mixed the business meaning of data with the physical, I'm making copies to make them faster or whatever. And so our perspective is that, if you can separate away the physical concerns from the logical, then business users have a much more, much more likelihood to be able to do something self-service. >> So you're essentially virtualizing my corpus of data, independent of location, is that right, I mean-- >> It's part of what we do, yeah. No, it's part of what we do. So, the way we look at it is, is kind of several different components to try to make something self-service. It starts with, yeah, virtualize or abstract away the details of the physical, right? But then, on top of that, expose a very, sort of a very user-friendly interface that allows people to sort of catalog and understand the different things, you know, search for things that they want to interact with, and then curate things, even if they're non-technical users, right? So the goal is that, if you talk to sort of even large internet companies in the Valley, it's very hard to even hire the amount of data engineering that you need to satisfy all the requests of your end-users of data. And so the, and so the goal of Dremio is basically to figure out different tools that can provide a non-technical experience for getting at the data. So that's sort of the start of it but then the second step is, once you've got access to this thing and people can collaborate and sort of deal with the data, then you've got these huge volumes of data, right? It's big data and so how do you make that go faster? And then we have some components that we deal with, sort of, speed and acceleration. >> So maybe talk about how people are leveraging this capability, this platform, what the business impact is, what have you seen there? >> So a lot of people have this problem, which is, they have data all over the place and they're trying to figure out "How do I expose this "to my end-users?" And those end-users might be analysts, they might be data scientists, they might be product managers that are trying to figure out how their product is working. And so, what they're doing today is they're typically trying to build systems internally that, to provide these capabilities. And so, for example, working with a large auto manufacturer. And they've got a big initiative where they're trying to make the data that they have, they have huge amounts of data across all sort of different parts of the organization and they're trying to make that available to different data consumers. Now, of course, there's a bunch of security concerns that you need to have around that, but they just want to make the data more accessible. And so, what they're doing is they're using Dremio to figure out ways to, basically, catalog all the data below, expose that to the different users, applying lots of different security rules around that, and then create a bunch of reflections, which make the things go faster as people are interacting with the things. >> Well, what about the governance factor? I mean, you heard this in the hadoop world years ago. "Ah, we're going to make, we're going to harden hadoop, "we're going to" and really, there was no governance and it became more and more important. How do you guys handle that? Do you partner with people? Is it up to the customer to figure that out? Do you provide that? >> It's several different things, right? It's a complex ecosystem, right? So it's a combination of things. You start with partnering with different systems to make sure that you integrate well with those things. So the different things that control some parts of credentials inside the systems all the way down to "What's the file system permissions?", right? "What are the permissions inside of something like Hive and the metastore there?" And then other systems on top of that, like Sentry or Ranger are also exposing different credentialing, right? And so we work hard to sort of integrate with those things. On top of that, Dremio also provides a full security model inside of the sort of virtual space that we work. And so people can control the permissions, the ability to access or edit any object inside of Dremio based on user roles and LDAP and those kinds of things. So it's, it's kind of multiple layers that have to be working together. >> And tell me more about the company. So founded three years ago, I think a couple of raises, >> Yep >> who's backing you? >> Yeah, yeah, yeah, so we founded just under three years ago. We had great initial investors, in Red Point and Lightspeed, so two great initial investors and we raised about 15 million on that round. And then we actually just closed a B round in January of this year and we added Norwest to the portfolio there. >> Awesome, so you're now in the mode of, I mean, they always say, you know, software is such a capital-efficient business but you see software companies raising, you know, 900 million dollars and so, presumably, that's to compete, to go to market and, you know, differentiate with your messaging and branding. Is that sort of what the, the phase that you're in now? You kind of developed a product, it's technically sound, it's proven in the marketspace and now you're scaling the, the go-to-market, is that right? >> That's exactly right. So, so we've had a lot of early successes, a lot of Fortune 100 companies using Dremio today. For example, we're working with TransUnion. We're working with Intel. We actually have a great relationship with OVH, which is the third-largest hosting company in the world, so a lot of great, Daimler is another one. So working with a lot of great companies, seeing sort of great early success with the product with those companies, and really looking to say "Hey, we're out here." We've got a booth for the first time at Strata here and we're sort of letting people know about, sort of, a better way, or easier way, for people to deal with data >> Yeah. >> A happier way. >> I mean, it's a crowded space, right? There's a lot of tools out there, a lot of companies. I'm interested in how you sort of differentiate. Obviously simplification is a part of that, the breadth of your capabilities. But maybe, in your words, you could share with me how you differentiate from the competition and how you break out from the noise. >> Yeah, yeah, yeah, so it's, you're absolutely right, it's a very crowded space. Everybody's using the same words and that makes it very hard for people to understand what's going on. And so, what we've found is very simple is that typically we will actually, the first meeting we deal with a customer, within the first 10 minutes we'll demo the product. Because so many technologies are technologies, not, they're not products and so you have to figure out how to use the product. You've got to figure out how you would customize it for your certain use-case. And what we've found with our product is, by making it very, very simple, people start, the light goes on in a very short amount of time and so, we also do things on our website so that you can see, in a couple of minutes, or even less than that, little animations that sort of give you a sense of what it's about. But really, it's just "Hey, this is a product "which is about", there's this light bulb that goes on, it's great. And you figure this out over the course of working with different customers, right? But there's this light bulb that goes on for people that are so confused by all the things that are going on and if we can just sit down with them, show them the product for a few minutes, all of a sudden they're like "Wait a minute, "I can use this", right? So you're frequently talking to buyers that are not the most technical parts of the organization initially, and so most of the technologies they look at are technologies that are very difficult to understand and they have to look to others to try to even understand how it would fit into their architecture. With Dremio, we have customers that can, that have installed it and gotten up, and within an hour or two, started to see real value. And that sort of excitement happens even in the demo, with most people. >> So you kind of have this bifurcated market. Since the big data meme, everybody says they're data-driven and you've got a bifurcated market in that, you've got the companies that are data-driven and you've got companies who say they're data-driven but really aren't. Who are your customers? Are they in both? Are they predominantly in the data-driven side? Are they predominantly in the trying to be data-driven? >> Well, I would say that they all would say that they're data-driven. >> Yeah, everyone, who's going to say "Well, we're not data-driven." >> Yeah, yeah, yeah. So I would say >> We're dead. >> I would say that everybody has data and they've got some ways that they're using it well and other places where they feel like they're not using it as well as they should. And so, I mean, the reason that we exist is to make it so it's easier for people to get value out of data, and so, if they were getting all the value they think they could get out of data, then we probably wouldn't exist and they would be fully data-driven. So I think that everybody, it's a journey and people are responding well to us, in part, because we're helping them down that journey. >> Well, the reason I asked that question is that we go to a lot of shows and everybody likes to throw out the digital transformation buzzword and then use Uber and Airbnb as an example, but if you dig deeper, you see that data is at the core of those companies and they're now beginning to apply machine intelligence and they're leveraging all this data that they've built up, this data architecture that they built up over the last five or 10 years. And then you've got this set of companies where all the data lives in silos and I can see you guys being able to help them. At the same time, I can see you helping the disruptors, so how do you see that? I mean, in terms of your role, in terms of affecting either digital transformations or digital disruptions. >> Well, I'd say that in either case, so we believe in a very sort of simple thing, which is that, so going back to what I said at the beginning, which is just that I see this regression in terms of data access, right? And so what happens is that, if you have a tightly-coupled system between two layers, then it becomes very difficult for people to sort of accommodate two different sets of needs. And so, the change over the last 10 years was the rise of the developer as the primary person for controlling data and that brought a huge amount of great things to it but analysis was not one of them. And there's tools that try to make that better but that's really the problem. And so our belief is very simple, which is that a new tier needs to be introduced between the consumers and the, and the producers of data. And that, and so that tier may interact with different systems, it may be more complex or whatever, for certain organizations, but the tier is necessary in all organizations because the analysts shouldn't be shaken around every time the developers change how they're doing data. >> Great. John Furrier has a saying that "Data is the new development kit", you know. He said that, I don't know, eight years ago and it's really kind of turned out to be the case. Jacques Nadeau, thanks very much for coming on theCUBE. Really appreciate your time. >> Yeah. >> Great to meet you. Good luck and keep us informed, please. >> Yes, thanks so much for your time, I've enjoyed it. >> You're welcome. Alright, thanks for watching everybody. This is theCUBE. We're live from Big Data SV. We'll be right back. (bright music)
SUMMARY :
Brought to you by SiliconANGLE Media We've been here most of the week, So we were talking a little bit about what you guys do. And we said, you know what, there needs to be a way Okay, so you solved that problem how, and the idea is pretty simple. So the goal is that, if you talk to sort of expose that to the different users, I mean, you heard this in the hadoop world years ago. And so people can control the permissions, And tell me more about the company. And then we actually just closed a B round that's to compete, to go to market and, you know, for people to deal with data and how you break out from the noise. and so most of the technologies they look at So you kind of have this bifurcated market. that they're data-driven. Yeah, everyone, who's going to say So I would say And so, I mean, the reason that we exist is At the same time, I can see you helping the disruptors, And so, the change over the last 10 years "Data is the new development kit", you know. Great to meet you. This is theCUBE.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
Jacques Nadeau | PERSON | 0.99+ |
Daimler | ORGANIZATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
Norwest | ORGANIZATION | 0.99+ |
Intel | ORGANIZATION | 0.99+ |
Wikibon | ORGANIZATION | 0.99+ |
TransUnion | ORGANIZATION | 0.99+ |
Jacque | PERSON | 0.99+ |
San Jose | LOCATION | 0.99+ |
OVH | ORGANIZATION | 0.99+ |
Lightspeed | ORGANIZATION | 0.99+ |
second step | QUANTITY | 0.99+ |
Uber | ORGANIZATION | 0.99+ |
two layers | QUANTITY | 0.99+ |
Airbnb | ORGANIZATION | 0.99+ |
both | QUANTITY | 0.99+ |
SiliconANGLE Media | ORGANIZATION | 0.99+ |
Google Docs | TITLE | 0.99+ |
Red Point | ORGANIZATION | 0.99+ |
Strata | ORGANIZATION | 0.99+ |
60 | QUANTITY | 0.98+ |
900 million dollars | QUANTITY | 0.98+ |
three years ago | DATE | 0.98+ |
eight years ago | DATE | 0.98+ |
two | QUANTITY | 0.98+ |
Dremio | PERSON | 0.98+ |
first 10 minutes | QUANTITY | 0.98+ |
last night | DATE | 0.98+ |
about 15 million | QUANTITY | 0.97+ |
theCUBE | ORGANIZATION | 0.97+ |
first time | QUANTITY | 0.97+ |
Dremio | ORGANIZATION | 0.97+ |
Big Data SV | ORGANIZATION | 0.96+ |
an hour | QUANTITY | 0.96+ |
two great initial investors | QUANTITY | 0.95+ |
today | DATE | 0.93+ |
first meeting | QUANTITY | 0.93+ |
this morning | DATE | 0.92+ |
two different sets | QUANTITY | 0.9+ |
third | QUANTITY | 0.88+ |
Big Data | ORGANIZATION | 0.87+ |
SQL | TITLE | 0.87+ |
10 years | QUANTITY | 0.87+ |
CUBE | ORGANIZATION | 0.87+ |
years ago | DATE | 0.86+ |
Silicon Valley | LOCATION | 0.86+ |
January of this year | DATE | 0.84+ |
Dremio | TITLE | 0.84+ |
Three year old | QUANTITY | 0.81+ |
last 10 years | DATE | 0.8+ |
Sentry | ORGANIZATION | 0.77+ |
one of them | QUANTITY | 0.75+ |
about 50 | QUANTITY | 0.75+ |
day two | QUANTITY | 0.74+ |
Ranger | ORGANIZATION | 0.74+ |
SV | EVENT | 0.7+ |
last ten years | DATE | 0.68+ |
eighth big | QUANTITY | 0.68+ |
Data | ORGANIZATION | 0.66+ |
Big | EVENT | 0.65+ |
couple of minutes | QUANTITY | 0.61+ |
CTO | PERSON | 0.56+ |
one | QUANTITY | 0.55+ |
last | DATE | 0.52+ |
100 companies | QUANTITY | 0.52+ |
under | DATE | 0.51+ |
five | QUANTITY | 0.5+ |
2018 | DATE | 0.5+ |
Hive | TITLE | 0.42+ |
Ian Swanson, DataScience.com | Big Data SV 2018
(royal music) >> Announcer: John Cleese. >> There's a lot of people out there who have no idea what they're doing, but they have absolutely no idea that they have no idea what they're doing. Those are the ones with the confidence and stupidity who finish up in power. That's why the planet doesn't work. >> Announcer: Knowledgeable, insightful, and a true gentleman. >> The guy at the counter recognized me and said... Are you listening? >> John Furrier: Yes, I'm tweeting away. >> No, you're not. >> I tweet, I'm tweeting away. >> He is kind of rude that way. >> You're on your (bleep) keyboard. >> Announcer: John Cleese joins the Cube alumni. Welcome, John. >> John Cleese: Have you got any phone calls you need to answer? >> John Furrier: Hold on, let me check. >> Announcer: Live from San Jose, it's the Cube, presenting Big Data Silicon Valley, brought to you by Silicon Angle Media and its ecosystem partners. (busy music) >> Hey, welcome back to the Cube's continuing coverage of our event, Big Data SV. I'm Lisa Martin with my co-host, George Gilbert. We are down the street from the Strata Data Conference. This is our second day, and we've been talking all things big data, cloud data science. We're now excited to be joined by the CEO of a company called Data Science, Ian Swanson. Ian, welcome to the Cube. >> Thanks so much for having me. I mean, it's been a awesome two days so far, and it's great to wrap up my trip here on the show. >> Yeah, so, tell us a little bit about your company, Data Science, what do you guys do? What are some of the key opportunities for you guys in the enterprise market? >> Yeah, absolutely. My company's called datascience.com, and what we do is we offer an enterprise data science platform where data scientists get to use all they tools they love in all the languages, all the libraries, leveraging everything that is open source to build models and put models in production. Then we also provide IT the ability to be able to manage this massive stack of tools that data scientists require, and it all boils down to one thing, and that is, companies need to use the data that they've been storing for years. It's about, how do you put that data into action. We give the tools to data scientists to get that data into action. >> Let's drill down on that a bit. For a while, we thought if we just put all our data in this schema-on-read repository, that would be nirvana. But it wasn't all that transparent, and we recognized we have to sort of go in and structure it somewhat, help us take the next couple steps. >> Ian: Yeah, the journey. >> From this partially curated data sets to something that turns into a model that is actionable. >> That's actually been the theme in the show here at the Strata Data Conference. If we went back years ago, it was, how do we store data. Then it was, how do we not just store and manage, but how do we transform it and get it into a shape that we can actually use it. The theme of this year is how do we get it to that next step, the next step of putting it into action. To layer onto that, data scientists need to access data, yes, but then they need to be able to collaborate, work together, apply many different techniques, machine learning, AI, deep learning, these are all techniques of a data scientist to be able to build a model. But then there's that next step, and the next is, hey, I built this model, how do I actually get it in production? How does it actually get used? Here's the shocking thing. I was at an event where there's 500 data scientists in the audience, and I said, "Stand up if you worked on a model for more than nine months "and it never went into production." 90% of the audience stood up. That's the last mile that we're all still working on, and what's exciting is, we can make it possible today. >> Wanting to drill down into the sort of, it sounds like there's a lot of choice in the tools. But typically, to do a pipeline, you either need well established APIs that everyone understands and plugs together with, or you need an end to end sort of single vendor solution that becomes the sort of collaboration backbone. How are you organized, how are you built? >> This might be self-serving, but datascience.com, we have enterprise data science platform, we recommend a unified platform for data science. Now, that unified platform needs to be highly configurable. You need to make it so that that workbench, you can use any tool that you want. Some data scientists might want to use a hammer, others want to be able to use a screwdriver over here. The power is how configurable, how extensible it is, how open source you can adopt everything. The amazing trends that we've seen have been proprietary solutions going back decades, to now, the rise of open source. Every day, dozens if not hundreds of new machine learning libraries are being released every single day. We've got to give those capabilities to data scientists and make them scale. >> OK, so the, and I think it's pretty easy to see how you would have incorporate new machine learning libraries into a pipeline. But then there's also the tools for data preparation, and for like feature extraction and feature engineering, you might even have some tools that help you with figuring out which algorithm to select. What holds all that together? >> Yeah, so orchestrating the enterprise data science stack is the hardest challenge right now. There has to be a company like us that is the glue, that is not just, do these solutions work together, but also, how do they collaborate, what is that workflow? What are those steps in that process? There's one thing that you might have left out, and that is, model deployment, model interpretation, model management. >> George: That's the black art, yeah. >> That's where this whole thing is going next. That was the exciting thing that I heard in terms of all these discussion with business leaders throughout the last two days is model deployment, model management. >> If I can kind of take this to maybe shift the conversation a little bit to the target audience. Talked a lot about data scientists and needing to enable them. I'm curious about, we just talked with, a couple of guests ago, about the chief data officer. How, you work with enterprises, how common is the chief data officer role today? What are some of the challenges they've got that datascience.com can help them to eliminate? >> Yeah, the CIO and the chief data officer, we have CIOs that have been selecting tools for companies to use, and now the chief data officer is sitting down with the CEO and saying, "How do we actually drive business results?" We work very closely with both of those personas. But on the CDO side, it's really helping them educate their teams on the possibilities of what could be realized with the data at hand, and making sure that IT is enabling the data scientists with the right tools. We supply the tools, but we also like to go in there with our customers and help coach, help educate what is possible, and that helps with the CDO's mission. >> A question along that front. We've been talking about sort of empowering the data scientist, and really, from one end of the modeling life cycle all the way to the end or the deployment, which is currently the hardest part and least well supported. But we also have tons of companies that don't have data science trained people, or who are only modestly familiar. Where do, what do we do with them? How do we get those companies into the mainstream in terms of deploying this? >> I think whether you're a small company or a big company, digital transformation is the mandate. Digital transformation is not just, how do I make a taxi company become Uber, or how do I make a speaker company become Sonos, the smart speaker, it's how do I exploit all the sources of my data to get better and improved operational processes, new business models, increased revenue, reduced operation costs. You could start small, and so we work with plenty of smaller companies. They'll hire a couple data scientists, and they're able to do small quick wins. You don't have to go sit in the basement for a year having something that is the thing, the unicorn in the business, it's small quick wins. Now we, my company, we believe in writing code, trained, educated, data scientists. There are solutions out there that you throw data at, you push a button, it gets an output. It's this magic black box. There's risk in that. Model interpretation, what are the features it's scoring on, there's risk, but those companies are seeing some level of success. We firmly believe, though, in hiring a data science team that is trained, you can start small, two or three, and get some very quick wins. >> I was going to say, those quick wins are essential for survivability, like digital transformation is essential, but it's also, I mean, to survival at a minimum, right? >> Ian: Yes. >> Those quick wins are presumably transformative to an enterprise being able to sustain, and then eventually, or ideally, be able to take market share from their competition. >> That is key for the CDO. The CDO is there pitching what is possible, he's pitching, she's pitching the dream. In order to be able to help visualize what that dream and the outcome could be, we always say, start small, quick wins, then from there, you can build. What you don't want to do is go nine months working on something and you don't know if there's going to be outcome. A lot of data science is trial and error. This is science, we're testing hypotheses. There's not always an outcome that's to be there, so small quick wins is something we highly recommend. >> A question, one of the things that we see more and more is the idea that actionable insights are perishable, and that latency matters. In fact, you have a budget for latency, almost, like in that short amount of time, the more sort of features that you can dynamically feed into a model to get a score, are you seeing more of that? How are the use cases that you're seeing, how's that pattern unfolding? >> Yeah, so we're seeing more streaming data use cases. We work with some of the biggest technology companies in the world, so IoT, connected services, streaming real time decisions that are happening. But then, also, there are so many use cases around org that could be marketing, finance, HR related, not just tech related. On the marketing side, imagine if you're customer service, and somebody calls you, and you know instantly the lifetime value of that customer, and it kicks off a totally new talk track, maybe get escalated immediately to a new supervisor, because that supervisor can handle this top tier customer. These are decisions that can happen real time leveraging machine learning models, and these are things that, again, are small quick wins, but massive, massive impact. It's about decision process now. That's digital transformation. >> OK. Are you seeing patterns in terms of how much horsepower customers are budgeting for the training process, creating the model? Because we know it's very compute intensive, like, even Intel, some people call it, like, high performance compute, like a supercomputer type workload. How much should people be budgeting? Because we don't see any guidelines or rules of thumb for this. >> I still think the boundaries are being worked out. There's a lot of great work that Nvidia's doing with GPU, we're able to do things faster on compute power. But even if we just start from the basics, if you go and talk to a data scientist at a massive company where they have a team of over 1,000 data scientists, and you say to do this analysis, how do you spin up your compute power? Well, I go walk over to IT and I knock on the door, and I say, "Set up this machine, set up this cluster." That's ridiculous. A product like ours is able to instantly give them the compute power, scale it elastically with our cloud service partners or work with on-prem solutions to be able to say, get the power that you need to get the results in the time that's needed, quick, fast. In terms of the boundaries of the budget, that's still being defined. But at the end of the day, we are seeing return on investment, and that's what's key. >> Are you seeing a movement towards a greater scope of integration for the data science tool chain? Or is it that at the high end, where you have companies with 1,000 data scientists, they know how to deal with specialized components, whereas, when there's perhaps less of, a smaller pool of expertise, the desire for end to end integration is greater. >> I think there's this kind of thought that is not necessarily right, and that is, if you have a bigger data science team, you're more sophisticated. We actually see the same sophistication level of 1,000 person data science team, in many cases, to a 20 person data science team, and sometimes inverse, I mean, it's kind of crazy. But it's, how do we make sure that we give them the tools so they can drive value. Tools need to include collaboration and workflow, not just hammers and nails, but how do we work together, how do we scale knowledge, how do we get it in the hands of the line of business so they can use the results. It's that that is key. >> That's great, Ian. I also like that you really kind of articulated start small, quick ins can make massive impact. We want to thank you so much for stopping by the Cube and sharing that, and what you guys are doing at Data Science to help enterprises really take advantage of the value that data can really deliver. >> Thanks so much for having datascience.com on, really appreciate it. >> Lisa: Absolutely. George, thank you for being my co-host. >> You're always welcome. >> We want to thank you for watching the Cube. I'm Lisa Martin with George Gilbert, and we are at our event Big Data SV on day two. Stick around, we'll be right back with our next guest after a short break. (busy music)
SUMMARY :
Those are the ones with the confidence and stupidity and a true gentleman. The guy at the counter recognized me and said... Announcer: John Cleese joins the Cube alumni. brought to you by Silicon Angle Media We are down the street from the Strata Data Conference. and it's great to wrap up my trip here on the show. and it all boils down to one thing, and that is, the next couple steps. to something that turns into a model that is actionable. and the next is, hey, I built this model, that becomes the sort of collaboration backbone. how open source you can adopt everything. OK, so the, and I think it's pretty easy to see Yeah, so orchestrating the enterprise data science stack in terms of all these discussion with business leaders a couple of guests ago, about the chief data officer. and making sure that IT is enabling the data scientists empowering the data scientist, and really, having something that is the thing, or ideally, be able to take market share and the outcome could be, we always say, start small, the more sort of features that you can dynamically in the world, so IoT, connected services, customers are budgeting for the training process, get the power that you need to get the results Or is it that at the high end, We actually see the same sophistication level and sharing that, and what you guys are doing Thanks so much for having datascience.com on, George, thank you for being my co-host. and we are at our event Big Data SV on day two.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
George Gilbert | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Ian Swanson | PERSON | 0.99+ |
George | PERSON | 0.99+ |
Ian | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
Uber | ORGANIZATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
Silicon Angle Media | ORGANIZATION | 0.99+ |
John | PERSON | 0.99+ |
John Cleese | PERSON | 0.99+ |
500 data scientists | QUANTITY | 0.99+ |
90% | QUANTITY | 0.99+ |
dozens | QUANTITY | 0.99+ |
Nvidia | ORGANIZATION | 0.99+ |
San Jose | LOCATION | 0.99+ |
20 person | QUANTITY | 0.99+ |
Data Science | ORGANIZATION | 0.99+ |
nine months | QUANTITY | 0.99+ |
1,000 person | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
two days | QUANTITY | 0.99+ |
more than nine months | QUANTITY | 0.99+ |
second day | QUANTITY | 0.99+ |
1,000 data scientists | QUANTITY | 0.99+ |
three | QUANTITY | 0.99+ |
Big Data SV | EVENT | 0.99+ |
over 1,000 data scientists | QUANTITY | 0.99+ |
Cube | ORGANIZATION | 0.99+ |
both | QUANTITY | 0.99+ |
Strata Data Conference | EVENT | 0.98+ |
one | QUANTITY | 0.98+ |
Intel | ORGANIZATION | 0.98+ |
Sonos | ORGANIZATION | 0.98+ |
one thing | QUANTITY | 0.97+ |
a year | QUANTITY | 0.96+ |
today | DATE | 0.95+ |
day two | QUANTITY | 0.95+ |
this year | DATE | 0.94+ |
single | QUANTITY | 0.92+ |
Big Data SV 2018 | EVENT | 0.88+ |
DataScience.com | ORGANIZATION | 0.87+ |
hundreds of new machine learning libraries | QUANTITY | 0.86+ |
lot of people | QUANTITY | 0.83+ |
decades | QUANTITY | 0.82+ |
every single day | QUANTITY | 0.81+ |
years ago | DATE | 0.77+ |
last two days | DATE | 0.76+ |
datascience.com | ORGANIZATION | 0.75+ |
one end | QUANTITY | 0.7+ |
years | QUANTITY | 0.67+ |
datascience.com | OTHER | 0.65+ |
couple steps | QUANTITY | 0.64+ |
Big Data | EVENT | 0.64+ |
couple of guests | DATE | 0.57+ |
couple | QUANTITY | 0.52+ |
Silicon Valley | LOCATION | 0.52+ |
things | QUANTITY | 0.5+ |
Cube | TITLE | 0.47+ |
Blaine Mathieu, VANTIQ | Big Data SV 2018
>> Announcer: Live from San Jose, it's The Cube, presenting Big Data, Silicon Valley. Brought to you by Silicon Angle Media and its ecosystem partners. >> Welcome back to The Cube. Our continuing coverage of our event, Big Data SV continues. I am Lisa Martin joined by Peter Burris. We're in downtown San Jose at a really cool place called Forager Tasting and Eatery. Come down, hang out with us today as we have continued conversations around all things big data, everything in between. This is our second day here and we're excited to welcome to The Cube the CMO of VANTIQ, Blaine Mathieu. Blaine, great to meet you, great to have you on the program. >> Great to be here, thanks for inviting me. >> So, VANTIQ, you guys are up the street in Walnut Creek. What do you guys do, what are you about, what makes VANTIQ different? >> Well, in a nutshell, VANTIQ is a so called high productivity application development platform to allow developers to build, deploy, and manage so called event driven real time applications, the kind of applications that are critical for driving many of the digital transformation initiatives that enterprises are trying to get on top of these days. >> Digital trasformation, it's a term that can mean so many different things, but today, it's essential for companies to be able to compete, especially enterprise companies with newer companies that are more agile, more modern. But if we peel apart digital transformation, there's so many elements that are essential. How do you guys help companies, enterprises, say, evolve their application architectures that might currently not be able to support an actual transformation to a digital business? >> Well, I think that's a great question, thank you. I think the key to digital trasformation is really a lot around the concept of real time, okay. The reason Uber is disrupting or has disrupted the taxi industry is the old way of doing it was somebody called a taxi and then they waited 30 minutes for a taxi to show up and then they told the taxi where to go and hopefully they got there. Whereas, Uber, turned that into a real time business, right? You called, you pinged something on your phone. They knew your location. They knew the location of the driver. They matched those up, brought 'em together in real time. Already knew where to bring you to and ensured you had the right route and that location. All of this data flowing, all of these actions have been taken in real time. The same thing applies to a disruptor like Netflix, okay? In the old days, Blockbuster used to send you, you know, a leaflet in the mail telling you what the new movies are. Maybe it was personalized for you. Probably not. No, Netflix knows who you are instantly, gives you that information, again, in real time based on what you've done in the past and is able to give you, deliver the movie also, in real time pretty well. Every disruptor you look at around digital transformation is bringing a business or a process that was done slowly and impersonally to make it happen in real time. Unfortunately, enterprise applications and the architectures, as you said a second ago, that are being used in most applications today weren't designed to enable these real time use cases. A great example is sales force. So, a sales force is a pretty standard, what you'd call a request application. So, you make a request, a person, generally, makes a request of the system, system goes into a database, queries that database, find information and then returns it back to the user. And that whole process could take, you know, significant amounts of time, especially if the right data isn't in the database at the time and you have to go request it or find it or create it. A new type of application needs to be created that's not fundamentally database centric, but it's able to take these real time data streams coming in from devices, from people, from enterprise systems, process them in real time and then take an action. >> So, let's pretend I'm a CEO. >> Yeah. >> One of the key things you said, and I want you to explain it better, is event. What is event? What is an event and how does that translate into a digital business decision? >> This notion of complex event processing CEP has been around in technology for a long time and yet, it surprises me still a lot of folks we talk to, CEOs, have never heard of the concept. And, it's very simple really. An event is just something that happens in the context of business. That's as complex and as simple as it is. An event could be a machine increases in temperature by one degree, a car moves from one location to another location. It could be an enterprise system, like an ERP system, you know, approves a PO. It could be a person pressing a button on a mobile device. All of those, or it could be an IOT device putting off a signal about the state of a machine. Increasingly, we're getting a lot of events coming from IOT devices. So, really, any particular interesting business situation or a change in a situation that happens is an event And increasingly driven, as you know, by IOT, by augmented reality, by AI and machine learning, by autonomous vehicles, by all these new real time technologies are spinning off more and more events, streams of these events coming off in rapid fashion and we have to be able to do something about them. >> Let me take a crack at it and you tell me if I've got this right. That, historically, applications have been defined in terms of processes and so, in many respects, there was a very concrete, discreet, well established program, set of steps that were performed and then the transaction took place. And event, it seems to me is, yeah, we generally described it, but it changes in response to the data. >> Right, right. >> So, an event is kind of like an outside in driven by data. >> Right, right. >> System response, whereas, your traditional transaction processing is an inside out driven by a sequence of programmed steps, and that decision might have been made six years ago. So, the event is what's happening right now informed by data versus a transaction, traditional transaction is much more, what did we decide to do six years ago and it just gets sustained. Have I got that right? >> That's right. Absolutely right or six hours ago or even six minutes ago, which might seem wow, six minutes, that's pretty good, but take a use case for a field service agent trying to fix a machine or an air conditioner on top of a building. In today's world now, that air conditioner has hundreds of sensors that are putting off data about the state of that air conditioner in real time. A service tech has the ability to, while the machine is still putting off that data, be able to make repairs and changes and fixes, again, in the moment, see how that is changing the data coming off the machine, and then, continue to make the appropriate repairs in collaboration with a smart system or an application that's helping them. >> That's how identifying patterns about what the problem is, versus some of the old ways was where we had recipe of, you know, steps that you went through in the call center. >> Right, right. And the customer is getting more and more frustrated. >> They got their clipboard out and had the 52 steps they followed to see oh that didn't work, now the next step. No, data can help us do that much more efficiently and effectively if we're able to process it in real time. >> So, in many respects, what we're really talking about is an application world or a world looking forward where the applications, which historically have been very siloed, process driven, to a world where the application function is much more networked together and the application, the output of one application is having a significant impact through data on the performance of an application somewhere else. That seems like it's got the potential to be an extremely complex fabric. (laughing) So, do I wait until I figure all that out (laughing) and then I start building it? Or do I, I mean, how do I do it? Do I start small and create and grow into it? What's the best way for people to start working on this? >> Well, you're absolutely right. Building these complex, geeking out a little bit, you know, asynchronous, non-blocking, so called reactive applications, that's the concept that we've been using in computer science for some time, is very hard, frankly. Okay, it's much easier to build computing systems that process things step one, step, two, step three, in order, but if you have to build a system that is able to take real time inputs or changes at any point in the process at any time and go in a different direction, it's very complex. And, computer scientists have been writing applications like this for decades. It's possible to do, but that isn't possible to do at the speed that companies now want to transform themselves, right? By the time you spec out an application and spend two years writing it, your business competitors have already disrupted you. The requirements have already changed. You need to be much more rapid and agile. And so, the secret sauce to this whole thing is to be able to write these transformative applications or create them, not even write is actually the wrong word to use, to be able to create them. >> Generate them. >> Yeah, generate them in a way which is very fast, does not require a guru level developer and reactive Java or some super low level code that you'd have to use to otherwise do it, so that you can literally have business people help design the applications, conceptually build them almost in real time, get them out into the market, and then be able to modify them as you need to, you know, on the fly. >> If I can build on that for just one second. So, it used to be we had this thing called computer assisted software engineer. >> (laughs) Right, right. >> We were going to operate this very very high level language. It's kind of-- But then, we would use code and build a code and the two of them were separated and so the minute that we deployed, somebody would go off and maintain and the whole thing would break. >> Right, right. >> Do you have that problem? >> No, well, that's exactly right. So, the old, you know, the old, the previous way of doing it was about really modeling an application, maybe visually, drag and drop, but then fundamentally, you created a bunch of code and then your job, as you said after, was to maintain and deploy and manage. >> Try to sustain some connection back up to that beautiful visual model. >> And you probably didn't because that was too much. That was too much work, so forget about the model after that. Instead, what we're able to do these days is to build the applications visually, you know, really for the most part with either super low code or, in many cases, no code because we have the ability to abstract away a lot of the complexity, a lot of the complex code that you'd have to write, we can represent that, okay, with these logical abstractions, create the applications themselves, and then continue to maintain, add to, modify the application using the exact same structure. You're not now stuck on, now you're stuck with 20,000 lines of code that you have to, that you have to edit. You're continuing to run and maintain the application just the way you built it, okay. We've now got to the place in computer science where we can actually do these things. We couldn't do them, you know, 20 years ago with case, but we can absolutely do them now. >> So, I'm hearing from a customer internal perspective a lot of operational efficiencies that VANTIQ can drive. Let's look now from a customer's perspective. What are the business impacts you're able to make? You mentioned the word reactive a minute ago when you were talking about applications, but do you have an example where you've, VANTIQ, has enabled a customer, a business, to be more, to be proactive and be able to identify through, you know, complex event processing, what their customers are doing to be able to deliver relevant messages and really drive revenue, drive profit? >> Right, right. So many, you know, so many great examples. And, I mentioned field service a few minutes ago. I've got a lot of clients in that doing this real time field service using these event processing applications. One that I want to bring up right now is one of the largest global shoe manufacturers, actually, that's a client of VANTIQ. I, unfortunately, can't say the name right now 'cause they want to keep what they're doing under wraps, but we all definitely know the company. And they're using this to manage the security, primarily, around their real time global supply chain. So, they've got a big challenge with companies in different countries redirecting shipments of their shoes, selling them on the gray market, at different prices than what are allowed in different regions of the world. And so, through both sensorizing the packages, the barcode scanning, the enterprise systems bringing all that data together in real time, they can literally tell in the moment is something is be-- If a package is redirected to the wrong region or if literally a shoe or a box of shoes is being sold where it shouldn't be sold at the wrong price. They used to get a monthly report on the activities and then they would go and investigate what happened last month. Now, their fraud detection manager is literally sitting there getting this in real time, saying, oh, Singapore sold a pallet of shoes that they should not have been able to sell five minute ago. Call up the guy in Singapore and have him go down and see what's going on and fix that issue. That's pretty powerful when you think about it. >> Definitely, so like reduction in fraud or increase in fraud detection. Sounds like, too, there's a potential for a significant amount of cost savings to the business, not just meeting the external customer needs, but from a, from a cost perspective reduction. Not just some probably TCO, but in operational expenses. >> For sure, although, I would say most of the digital transformation initiatives, when we talk to CEOs and CIOs, they're not focused as much on cost savings, as they're focused on A, avoiding being disrupted by the next interesting startup, B, creating new lines of business, new revenue streams, finding out a way to do something differently dramatically better than they're currently doing it. It's not only about optimizing or squeezing some cost out of their current application. This thing that we are talking about, I guess you could say it's an improvement on their current process, but really, it's actually something they just weren't even really doing before. Just a total different way of doing fraud detection and managing their global supply chain that they just fundamentally weren't even doing. And now, of course, they're looking at many other use cases across the company, not just in supply chain, but, you know, smart manufacturing, so many use cases. Your point about savings, though, there's, you know, what value does the application itself bring? Then, there's the question of what does it cost to build and maintain and deploy the application itself, right? And, again, with these new visual development tools, they're not modeling tools, you're literally developing the application visually. You know, I've been in so many scenarios where we talked to large enterprises. You know, we talk about what we're doing, like we talk about right now, and they say, okay, we'd love to do a POC, proof of concept. We want to allocate six months for this POC, like normally you would probably do for building most enterprise applications. And, we inevitably say, well, how about Friday? How about we have the POC done by Friday? And, you know, we get the Germans laugh, you know, laugh uncomfortably and we go away and deliver the POC by Friday because of how much different it is to build applications this way versus writing low level Java or C-sharp code and sticking together a bunch of technologies and tools 'cause we abstract all that away. And, you know, the eyes drop open and the mouth drops open and it's incredible what modern technology can do to radically change how software is being developed. >> Wow, big impact in a short period of time. That's always a nice thing to be able to deliver. >> It is, it is to-- It's great to be able to surprise people like that. >> Exactly, exactly. Well, Blaine, thank you so much for stopping by, sharing what VANTIQ is doing to help companies be disruptive and for sharing those great customer examples. We appreciate your time. >> You're welcome. Appreciate the time. >> And for my co-host, Peter Burris, I'm Lisa Martin. You're watching The Cube's continuing coverage of our event, Big Data SV Live from San Jose, down the street from the Strata Data Conference. Stick around, we'll be right back with our next guest after a short breal. (techy music)
SUMMARY :
Brought to you by Silicon Angle Media the CMO of VANTIQ, Blaine Mathieu. So, VANTIQ, you guys are up the street in Walnut Creek. for driving many of the digital transformation that might currently not be able to support and the architectures, as you said a second ago, One of the key things you said, in the context of business. in response to the data. So, an event is kind of like an outside in So, the event is what's happening right now and changes and fixes, again, in the moment, of the old ways was where we had recipe of, you know, And the customer is getting more and more frustrated. they followed to see oh that didn't work, and the application, the output of one application And so, the secret sauce to this whole thing to modify them as you need to, you know, on the fly. So, it used to be we had this thing and so the minute that we deployed, So, the old, you know, the old, Try to sustain just the way you built it, okay. but do you have an example where you've, that they should not have been able to sell to the business, not just meeting and deliver the POC by Friday because to be able to deliver. It's great to be able to surprise people Well, Blaine, thank you so much for stopping by, Appreciate the time. down the street from the Strata Data Conference.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Blaine | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Peter Burris | PERSON | 0.99+ |
Singapore | LOCATION | 0.99+ |
Uber | ORGANIZATION | 0.99+ |
two years | QUANTITY | 0.99+ |
Netflix | ORGANIZATION | 0.99+ |
San Jose | LOCATION | 0.99+ |
VANTIQ | ORGANIZATION | 0.99+ |
Blaine Mathieu | PERSON | 0.99+ |
20,000 lines | QUANTITY | 0.99+ |
30 minutes | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
Silicon Angle Media | ORGANIZATION | 0.99+ |
52 steps | QUANTITY | 0.99+ |
Walnut Creek | LOCATION | 0.99+ |
six months | QUANTITY | 0.99+ |
Java | TITLE | 0.99+ |
one degree | QUANTITY | 0.99+ |
Friday | DATE | 0.99+ |
second day | QUANTITY | 0.99+ |
last month | DATE | 0.99+ |
one second | QUANTITY | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
six years ago | DATE | 0.98+ |
both | QUANTITY | 0.98+ |
Strata Data Conference | EVENT | 0.98+ |
Big Data SV Live | EVENT | 0.98+ |
One | QUANTITY | 0.98+ |
The Cube | ORGANIZATION | 0.98+ |
today | DATE | 0.98+ |
one | QUANTITY | 0.98+ |
20 years ago | DATE | 0.98+ |
Big Data SV 2018 | EVENT | 0.97+ |
six hours ago | DATE | 0.97+ |
six minutes ago | DATE | 0.97+ |
five minute ago | DATE | 0.97+ |
a minute ago | DATE | 0.96+ |
hundreds of sensors | QUANTITY | 0.95+ |
The Cube | TITLE | 0.94+ |
Blockbuster | ORGANIZATION | 0.91+ |
few minutes ago | DATE | 0.89+ |
step one | QUANTITY | 0.89+ |
step three | QUANTITY | 0.85+ |
Forager Tasting and Eatery | ORGANIZATION | 0.85+ |
decades | QUANTITY | 0.84+ |
six minutes | QUANTITY | 0.84+ |
C | TITLE | 0.83+ |
Big Data | ORGANIZATION | 0.81+ |
one location | QUANTITY | 0.78+ |
one application | QUANTITY | 0.77+ |
second ago | DATE | 0.71+ |
CEP | ORGANIZATION | 0.53+ |
big | ORGANIZATION | 0.52+ |
Germans | PERSON | 0.51+ |
techy | ORGANIZATION | 0.41+ |
Data | EVENT | 0.31+ |
Octavian Tanase, NetApp | Big Data SV 2018
>> Announcer: Live from San Jose it's The Cube presenting Big Data, Silicon Valley brought to you by SiliconANGLE Media and its ecosystem partners. >> Good morning. Welcome to The Cube. We are on day two of our coverage our event Big Data SV. I'm Lisa Martin with my cohost Dave Vellante. We're down the street from the Strata Data Conference. This is The Cube's tenth big data event and we had a great day yesterday learning a lot from myriad guests on very different nuances of big data journey where things are going. We're excited to welcome back to The Cube an alumni, Octavian Tanase, the Senior Vice President of Data ONTAP fron Net App. Octavian, welcome back to The Cube. >> Glad to be here. >> So you've been at the Strata Data Conference for the last couple of days. From a big data perspective, what are some of the things that you're hearing, in terms of from a customer's perspective on what's working, what challenges, opportunities? I'm very excited to be here and learn about the innovation of our partners in the industry and share with our partners and our customers what we're doing to enable them to drive more value out of that data. The reality is that data has become the 21st Century gold or oil that powers the business and everybody's looking to apply new techniques, a lot of times machine learning, deep learning, to draw more value of the data, make better decisions and compete in the marketplace. Octavian, you've been at NetApp now eight years and I've been watching NetApp, as we were talking about offline, for decades and I've seen the ebb and flow and this company has transformed many, many times. The latest, obviously cloud came in, flash came into play and then you're also going through a major transition in the customer based to clustered ONTAP. You seemed to negotiate that. NetApp is back, thriving, stock's up. What's happening at NetApp? What's the culture like these days? Give us the update. >> I think we've been very fortunate to have a CEO like George Kurian, who has been really focused on helping us do basically fewer things better, really focus on our core business, simplify our operations and continue to innovate and this is probably the area that I'm most excited about. It's always good to make sure that you accelerate the business, make it simpler for your customers and your partners to do business with you, but what you have to do is innovate. We are a product company. We are passionate about innovation. I believe that we are innovating with more pace than many of the startups in the space so that's probably the most exciting thing that has been part of our transformation. >> So let's talk about big data. Back in the day if you had a big data problem you would buy a big Unix box, maybe buy some Oracle licenses, try to put all your data into that box and that became your data warehouse. The brilliance of Hadoop was hey we can leave the data where it is. There's too much data to put into the box so we're going to bring five megabytes to code to a petabyte of data. And the other piece of it is CFOs loved it, because we're going to reduce the cost of our expensive data warehouse and we're going to buy off the shelf components: white box, servers and off the shelf disk drives. We're going to put that together and life will be good. Well as things matured, the old client-server days, it got very expensive, you needed enterprise grade. So where does NetApp fit into that equation, because originally big storage companies like NetApp, they weren't part of the equation? Has that changed? >> Absolutely. One of the things that has enabled that transformation, that change is we made a deliberate decision to focus on software defined and making sure that the ONTAP operating system is available wherever data is being created: on the edge in an IoT device, in the traditional data center or in the cloud. So we are in the unique position to enable analytics, big data, wherever those applications reside. One of the things that we've recently done is we've partnered with IDC and what the study, what the analysis has shown is that deploying in analytics, a Hadoop or NoSQL type of solution on top of NetApp is half the cost of DAS. So when you consider the cost of servers, the licenses that you're going to have to pay for, these commercial implementations of Hadoop as well as the storage and the data infrastructure, you are much better off choosing NetApp than a white box type of solution. >> Let's unpack that a little bit, because if I infer correctly from what you said normally you would say the operational costs are going to be dramatically lower, it's easier to manage a professional system like a NetApp ONTAP, it's integrated, great software, but am I hearing you correctly, you're saying the acquisition costs are actually less than if I'm buying white box? A lot of people are going to be skeptical about that, say Octavian no way, it's cheaper to buy white box stuff. Defend that statement. >> Absolutely. If you're looking at the whole solution that includes the server and the storage, what NetApp enables you to do if you're running the solution on top of ONTAP you reduce the need for so many servers. If you reduce that number you also reduce the licensing cost. Moreover, if you actually look at the core value proposition of the storage layer there, DAS typically makes three copies of the data. We don't. We are very greedy and we're making sure that you're using shared storage and we are applying a bunch of storage efficiency techniques to further compress, compact that data for world class storage efficiency. >> So cost efficiency is obviously a great benefit for any company when they're especially evolving, from a digital perspective. What are some of the business level benefits? You mentioned speed a minute ago. What is Data ONTAP and even ONTAP in the cloud enabling your enterprise customers to achieve at the business level, maybe from faster time to market, identifying with machine learning and AI new products? Give me an example of maybe a customer that you think really articulates the value that ONTAP in the cloud can deliver. >> One of the things that's really important is to have your data management capability, whatever the data is being produced so ONTAP being consumed either as a VM or a service ... I don't know if you've seen some of the partnerships that we have with AWS and Azure. We're able to offer the same rich data management capabilities, not only the traditional data center, but in the cloud. What that really enables customers to do is to simplify and have the same operating system, the same data management platform for the both the second platform traditional applications as well as for the third platform applications. I've seen a company like Adobe be very successful in deploying their infrastructure, their services not only on prem in their traditional data center, but using ONTAP Cloud. So we have more than about 1,500 customers right now that have adopted ONTAP in the AWS cloud. >> What are you seeing in terms of the adoption of flash and I'm particularly interested in the intersection of flash adoption and the developer angle, because we've seen, in certain instances, certain organizations are able to share data off of flash much more efficiently that you would be, for instance, of a spinning disk? Have you seen a developer impact in your customer base? >> Absolutely I think most of customers initially have adopted flash, because of high throughput and low latency. I think over time customers really understood and identified with the overall value proposition in cost of ownership in flash that it enables them to consolidate multiple workloads in a smaller footprint. So that enables you to then reduce the cost to operate that infrastructure and it really gives you a range of applications that you can deploy that you were never able to do that. Everybody's looking to do in place, in line analytics that now are possible, because of this fast media. Folks are looking to accelerate old applications in which they cannot invest anymore, but they just want to run faster. Flash also tends to be more reliable than traditional storage, so customers definitely appreciate that fewer things could go wrong so overall the value proposition of flash, it's all encompassing and we believe that in the near future flash will be the defacto standard in everybody's data center, whether it's on prem or in the cloud. >> How about backup and recovery in big data? We obviously, in the enterprise, very concerned about data protection. What's similar in big data? What's different and what's NetApp's angle on that? >> I think data protection and data security will never stop being important to our customers. Security's top of mind for everybody in the industry and it's a source of resume changing events, if you would, and they're typically not promotions. So we have invested a tremendous deal in certifications for HIPAA, for FIPS, we are enabling encryption, both at rest and in flight. We've done a lot of work to make sure that the encryption can happen in software layer, to make sure that we give the customers best storage class efficiency and what we're also leveraging is the innovation that ONTAP has done over many years to protect the data, replicate its snapshots, peering the data to the cloud. These are techniques that we're commonly using to reduce the cost of ownership, also protect the data the customers deploy. >> So security's still a hot topic and, like you said, it probably always will be, but it's a shared responsibility, right? So customers leveraging NetApps safe or on prem hybrid also using Azure or AWS, who's your target audience? If you're talking to the guys and gals that are still managing storage are you also having the CSO or the security guys come in, the gals, to understand we've got this appointment in Azure or AWS so we're going to bring in ONTAP to facilitate this? There's a shared responsibility of security. Who's at the table, from your perspective, in your customers that you need to help understand how they facilitate true security? >> It's definitely been a transformative event where more and more people in IQ organizations are involved in the decisions that are required to deploy the applications. There was a time when we would talk only to the storage admin. After a while we started talking to the application admin, the virtualization admin and now you're talking to the line of business who has that vested interest to make sure that they can harness the power of the data in their environment. So you have the CSO, you have the traditional infrastructure people, you have the app administration and you have the app owner, the business owner that are all at the table that are coming and looking to choose the best of breed solution for their data management. >> What are the conversations like with your CXO, executives? Everybody talks about digital transformation. It's kind of an overused term, but there's real substance when you actually peel the onion. What are you seeing as NetApp's role in effecting digital transformations within your customer base? >> I think we have a vision of how we can help enterprises take advantage of the digital transformation and adopt it. I think we have three tenants of that vision. Number one is we're helping customers harness the power of the cloud. Number two, we're looking to enable them to future proof their investments and build the next generation data center. And number three, nobody starts with a fresh slate so we're looking to help customers modernize their current infrastructure through storage. We have a lot of expertise in storage. We've helped, over time, customers time and again adopt disruptive technologies in nondisruptive ways. We're looking to adopt these technologies and trends on behalf of our customers and then help them use them in a seamless safe way. >> And continue their evolution to identify new revenue streams, new products, new opportunities and even probably give other lines of business access to this data that they need to understand is there value here, how can we harness it faster than our competitors, right? >> Absolutely. It's all about deriving value out of the data. I think earlier I called it the gold of the 21st Century. This is a trend that will continue. I believe there will be no enterprise or center that won't focus on using machine learning, deep learning, analytics to derive more value out of the data to find more customer touch points, to optimize their business to really compete in the marketplace. >> Data plus AI plus cloud economics are the new innovation drivers of the next 10, 20 years. >> Completely agree. >> Well Octavian thanks so much for spending time with us this morning sharing what's new at NetApp, some of the visions that you guys have and also some of the impact that you're making with customers. We look forward to having you back on the program in the near future. >> Thank you. Appreciate having the time. >> And for my cohost Dave Vellante I'm Lisa Martin. You're watching The Cube live on day two of coverage of our event, Big Data SV. We're at this really cool venue, Forager Tasting Room. Come down here, join us, get to hear all these great conversations. Stick around and we'll be right back with our next guest after a short break. (electronic music)
SUMMARY :
brought to you by SiliconANGLE Media We're down the street from the Strata Data Conference. in the customer based to clustered ONTAP. that you accelerate the business, Back in the day if you had a big data problem and making sure that the ONTAP operating system A lot of people are going to be skeptical about that, that includes the server and the storage, that ONTAP in the cloud can deliver. that have adopted ONTAP in the AWS cloud. to operate that infrastructure and it really gives you We obviously, in the enterprise, peering the data to the cloud. that you need to help understand that are required to deploy the applications. What are the conversations like with your CXO, executives? and build the next generation data center. out of the data to find more customer touch points, are the new innovation drivers of the next 10, 20 years. We look forward to having you back on the program Appreciate having the time. get to hear all these great conversations.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
George Kurian | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Octavian Tanase | PERSON | 0.99+ |
Adobe | ORGANIZATION | 0.99+ |
Octavian | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
eight years | QUANTITY | 0.99+ |
San Jose | LOCATION | 0.99+ |
SiliconANGLE Media | ORGANIZATION | 0.99+ |
NetApp | TITLE | 0.99+ |
Hadoop | TITLE | 0.99+ |
five megabytes | QUANTITY | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
second platform | QUANTITY | 0.99+ |
21st Century | DATE | 0.99+ |
HIPAA | TITLE | 0.99+ |
Strata Data Conference | EVENT | 0.99+ |
yesterday | DATE | 0.99+ |
ONTAP | TITLE | 0.99+ |
The Cube | TITLE | 0.99+ |
IDC | ORGANIZATION | 0.98+ |
both | QUANTITY | 0.98+ |
One | QUANTITY | 0.98+ |
Unix | COMMERCIAL_ITEM | 0.98+ |
NetApp | ORGANIZATION | 0.97+ |
The Cube | ORGANIZATION | 0.97+ |
Silicon Valley | LOCATION | 0.96+ |
ONTAP Cloud | TITLE | 0.95+ |
more than about 1,500 customers | QUANTITY | 0.95+ |
NetApps | TITLE | 0.93+ |
Big Data SV | EVENT | 0.93+ |
Big Data SV 2018 | EVENT | 0.93+ |
day two | QUANTITY | 0.93+ |
Forager Tasting Room | LOCATION | 0.88+ |
NoSQL | TITLE | 0.87+ |
Azure | ORGANIZATION | 0.86+ |
third platform applications | QUANTITY | 0.81+ |
a minute ago | DATE | 0.81+ |
Number two | QUANTITY | 0.8+ |
Senior Vice President | PERSON | 0.79+ |
three tenants | QUANTITY | 0.78+ |
decades | QUANTITY | 0.74+ |
a petabyte of data | QUANTITY | 0.73+ |
tenth big | QUANTITY | 0.71+ |
Number one | QUANTITY | 0.71+ |
three copies | QUANTITY | 0.7+ |
this morning | DATE | 0.69+ |
number three | QUANTITY | 0.68+ |
ONTAP | ORGANIZATION | 0.67+ |
Data ONTAP | ORGANIZATION | 0.64+ |
event | QUANTITY | 0.64+ |
Net App | TITLE | 0.64+ |
10 | QUANTITY | 0.64+ |
half | QUANTITY | 0.6+ |
flash | TITLE | 0.58+ |
much | QUANTITY | 0.58+ |
Big Data | EVENT | 0.57+ |
years | QUANTITY | 0.55+ |
Sastry Malladi, FogHorn | Big Data SV 2018
>> Announcer: Live from San Jose, it's theCUBE, presenting Big Data Silicon Valley, brought to you by SiliconANGLE Media and its ecosystem partner. (upbeat electronic music) >> Welcome back to The Cube. I'm Lisa Martin with George Gilbert. We are live at our event, Big Data SV, in downtown San Jose down the street from the Strata Data Conference. We're joined by a new guest to theCUBE, Sastry Malladi, the CTO Of FogHorn. Sastry, welcome to theCUBE. >> Thank you, thank you, Lisa. >> So FogHorn, cool name, what do you guys do, who are you? Tell us all that good stuff. >> Sure. We are a startup based in Silicon Valley right here in Mountain View. We started about three years ago, three plus years ago. We provide edge computing intelligence software for edge computing or fog computing. That's how our company name got started is FogHorn. For our particularly, for our IoT industrial sector. All of the industrial guys, whether it's transportation, manufacturing, oil and gas, smart cities, smart buildings, any of those different sectors, they use our software to predict failure conditions in real time, or do condition monitoring, or predictive maintenance, any of those use cases and successfully save a lot of money. Obviously in the process, you know, we get paid for what we do. >> So Sastry... GE populized this concept of IIoT and the analytics and, sort of the new business outcomes you could build on it, like Power by the Hour instead of selling a jet engine. >> Sastry: That's right. But there's... Actually we keep on, and David Floor did some pioneering research on how we're going to have to do a lot of analytics on the edge for latency and bandwidth. What's the FogHorn secret sauce that others would have difficulty with on the edge analytics? >> Okay, that's a great question. Before I directly answer the question, if you don't mind, I'll actually even describe why that's even important to do that, right? So a lot of these industrial customers, if you look at, because we work with a lot of them, the amount of data that's produced from all of these different machines is terabytes to petabytes of data, it's real. And it's not just the traditional digital sensors but there are video, audio, acoustic sensors out there. The amount of data is humongous, right? It's not even practical to send all of that to a Cloud environment and do data processing, for many reasons. One is obviously the connectivity, bandwidth issues, and all of that. But the two most important things are cyber security. None of these customers actually want to connect these highly expensive machines to the internet. That's one. The second is the lack of real-time decision making. What they want to know, when there is a problem, they want to know before it's too late. We want to notify them it is a problem that is occurring so that have a chance to go fix it and optimize their asset that is in question. Now, existing solutions do not work in this constrained environment. That's why FogHorn had to invent that solution. >> And tell us, actually, just to be specific, how constrained an environment you can operate in. >> We can run in about less than 100 to 150 megabytes of memory, single-core to dual-core of CPU, whether it's an ARM processor, an x86 Intel-based processor, almost literally no storage because we're a real-time processing engine. Optionally, you could have some storage if you wanted to store some of the results locally there but that's the kind of environment we're talking about. Now, when I say 100 megabytes of memory, it's like a quarter of Raspberry Pi, right? And even in that environment we have customers that run dozens of machinery models, right? And we're not talking -- >> George: Like an ensemble. >> Like an anomaly detection, a regression, a random forest, or a clustering, or a gamut, some of those. Now, if we get into more deep learning models, like image processing and neural net and all of that, you obviously need a little bit more memory. But what we have shown, we could still run, one of our largest smart city buildings customer, elevator company, runs in a raspberry Pi on millions of elevators, right? Dozens of machinery algorithms on top of that, right? So that's the kind of size we're talking about. >> Let me just follow up with one question on the other thing you said, with, besides we have to do the low-latency locally. You said a lot of customers don't want to connect these brown field, I guess, operations technology machines to the internet, and physically, I mean there was physical separation for security. So it's like security, Bill Joy used to say "Security by obscurity." Here it's security by -- >> Physical separation, absolutely. Tell me about it. I was actually coming from, if you don't mind, last week I was in Saudi Arabia. One of the oil and gas plants where we deployed our software, you have to go to five levels of security even to get to there, It's a multibillion dollar plant and refining the gas and all of that. Completely offline, no connectivity to the internet, and we installed, in their existing small box, our software, connected to their live video cameras that are actually measuring the stuff, doing the processing and detecting the specific conditions that we're looking for. >> That's my question, which was if they want to be monitoring. So there's like one low level, really low hardware low level, the sensor feeds. But you could actually have a richer feed, which is video and audio, but how much of that, then, are you doing the, sort of, inferencing locally? Or even retraining, and I assume that since it's not the OT device, and it's something that's looking at it, you might be more able to send it back up the Cloud if you needed to do retraining? >> That's exactly right. So the way the model works is particularly for image processing because you need, it's a more complex process to train than create a model. You could create a model offline, like in a GPU box, an FPGA box and whatnot. Import and bring the model back into this small little device that's running in the plant, and now the live video data is coming in, the model is inferencing the specific thing. Now there are two ways to update and revise the model: incremental revision of the model, you could do that if you want, or you can send the results to a central location. Not internet, they do have local, in this example for example a PIDB, an OSS PIDB, or some other local service out there, where you have an opportunity to gather the results from each of these different locations and then consolidate and retrain the model, put the model back again. >> Okay, the one part that I didn't follow completely is... If the model is running ultimately on the device, again and perhaps not even on a CPU, but a programmable logic controller. >> It could, even though a programmable controller also typically have some shape of CPU there as well. These days, most of the PLCs, programmable controllers, have either an RM-based processor or an x86-based processor. We can run either one of those too. >> So, okay, assume you've got the model deployed down there, for the, you know, local inferencing. Now, some retraining is going to go on in the Cloud, where you have, you're pulling in the richer perspective from many different devices. How does that model get back out to the device if it doesn't have the connectivity between the device and the Cloud? >> Right, so if there's strictly no connectivity, so what happens is once the model is regenerated or retrained, they put a model in a USB stick, it's a low attack. USB stick, bring it to the PLC device and upload the model. >> George: Oh, so this is sort of how we destroyed the Iranian centrifuges. >> That's exactly right, exactly right. But you know, some other environments, even though it's not connectivity to the Cloud environment, per se, but the devices have the ability to connect to the Cloud. Optionally, they say, "Look, I'm the device "that's coming up, do you have an upgraded model for me?" Then it can pull the model. So in some of the environments it's super strict where there are absolutely no way to connect this device, you put it in a USB stick and bring the model back here. Other environments, device can query the Cloud but Cloud cannot connect to the device. This is a very popular model these days because, in other words imagine this, an elevator sitting in a building, somebody from the Cloud cannot reach the elevator, but an elevator can reach the Cloud when it wants to. >> George: Sort of like a jet engine, you don't want the Cloud to reach the jet engine. >> That's exactly right. The jet engine can reach the Cloud it if wants to, when it wants to, but the Cloud cannot reach the jet engine. That's how we can pull the model. >> So Sastry, as a CTO you meet with customers often. You mentioned you were in Saudi Arabia last week. I'd love to understand how you're leveraging and gaging with customers to really help drive the development of FogHorn, in terms of being differentiated in the market. What are those, kind of bi-directional, symbiotic customer relationships like? And how are they helping FogHorn? >> Right, that's actually a great question. We learn a lot from customers because we started a long time ago. We did an initial version of the product. As we begin to talk to the customers, particularly that's part of my job, where I go talk to many of these customers, they give us feedback. Well, my problem is really that I can't even do, I can't even give you connectivity to the Cloud, to upgrade the model. I can't even give you sample data. How do you do that modeling, right? And sometimes they say, "You know what, "We are not technical people, help us express the problem, "the outcome, give me tools "that help me express that outcome." So we created a bunch of what we call OT tools, operational technology tools. How we distinguish ourselves in this process, from the traditional Cloud-based vendor, the traditional data science and data analytics companies, is that they think in terms of computer scientists, computer programmers, and expressions. We think in terms of industrial operators, what can they express, what do they know? They don't really necessarily care about, when you tell them, "I've got an anomaly detection "data science machine algorithm", they're going to look at you like, "What are you talking about? "I don't understand what you're talking about", right? You need to tell them, "Look, this machine is failing." What are the conditions in which the machine is failing? How do you express that? And then we translate that requirement, or that into the underlying models, underlying Vel expressions, Vel or CPU expression language. So we learned a ton from user interface, capabilities, latency issues, connectivity issues, different protocols, a number of things that we learn from customers. >> So I'm curious with... More of the big data vendors are recognizing data in motion and data coming from devices. And some, like Hortonworks DataFlow NiFi has a MiNiFi component written in C plus plus, really low resource footprint. But I assume that that's really just a transport. It's almost like a collector and that it doesn't have the analytics built in -- >> That's exactly right, NiFi has the transport, it has the real-time transport capability for sure. What it does not have is this notion of that CEP concept. How do you combine all of the streams, everything is a time series data for us, right, from the devices. Whether it's coming from a device or whether it's coming from another static source out there. How do you express a pattern, a recognition pattern definition, across these streams? That's where our CPU comes in the picture. A lot of these seemingly similar software capabilities that people talk about, don't quite exactly have, either the streaming capability, or the CPU capability, or the real-time, or the low footprint. What we have is a combination of all of that. >> And you talked about how everything's time series to you. Is there a need to have, sort of an equivalent time series database up in some central location? So that when you subset, when you determine what relevant subset of data to move up to the Cloud, or you know, on-prem central location, does it need to be the same database? >> No, it doesn't need to be the same database. It's optional. In fact, we do ship a local time series database at the edge itself. If you have a little bit of a local storage, you can down sample, take the results, and store it locally, and many customers actually do that. Some others, because they have their existing environment, they have some Cloud storage, whether it's Microsoft, it doesn't matter what they use, we have connectors from our software to send these results into their existing environments. >> So, you had also said something interesting about your, sort of, tool set, as being optimized for operations technology. So this is really important because back when we had the Net-Heads and the Bell-Heads, you know it was a cultural clash and they had different technologies. >> Sastry: They sure did, yeah. >> Tell us more about how selling to operations, not just selling, but supporting operations technology is different from IT technology and where does that boundary live? >> Right, so typical IT environment, right, you start with the boss who is the decision maker, you work with them and they approve the project and you go and execute that. In an industrial, in an OT environment, it doesn't quite work like that. Even if the boss says, "Go ahead and go do this project", if the operator on the floor doesn't understand what you're talking about, because that person is in charge of operating that machine, it doesn't quite work like that. So you need to work bottom up as well, to convincing them that you are indeed actually solving their pain point. So the way we start, where rather than trying to tell them what capabilities we have as a product, or what we're trying to do, the first thing we ask is what is their pain point? "What's your problem? What is the problem "you're trying to solve?" Some customers say, "Well I've got yield, a lot of scrap. "Help me reduce my scrap. "Help me to operate my equipment better. "Help me predict these failure conditions "before it's too late." That's how the problem starts. Then we start inquiring them, "Okay, what kind of data "do you have, what kind of sensors do you have? "Typically, do you have information about under what circumstances you have seen failures "versus not seeing failures out there?" So in the process of inauguration we begin to understand how they might actually use our software and then we tell them, "Well, here, use your software, "our software, to predict that." And, sorry, I want 30 more seconds on that. The other thing is that, typically in an IT environment, because I came from that too, I've been in this position for 30 plus years, IT, UT and all of that, where we don't right away talk about CEP, or expressions, or analytics, and we don't talk about that. We talk about, look, you have these bunch of sensors, we have OT tools here, drag and drop your sensors, express the outcome that you're trying to look for, what is the outcome you're trying to look for, and then we drive behind the scenes what it means. Is it analytics, is it machine learning, is it something else, and what is it? So that's kind of how we approach the problem. Of course, if, sometimes you do surprisingly occasionally run into very technical people. From those people we can right away talk about, "Hey, you need these analytics, you need to use machinery, "you need to use expressions" and all of that. That's kind of how we operate. >> One thing, you know, that's becoming clearer is I think this widespread recognition that's data intensive and low latency work to be done near the edge. But what goes on in the Cloud is actually closer to simulation and high-performance compute, if you want to optimize a model. So not just train it, but maybe have something that's prescriptive that says, you know, here's the actionable information. As more of your data is video and audio, how do you turn that into something where you can simulate a model, that tells you the optimal answer? >> Right, so this is actually a good question. From our experience, there are models that require a lot of data, for example, video and audio. There are some other models that do not require a lot of data for training. I'll give you an example of what customer use cases that we have. There's one customer in a manufacturing domain, where they've been seeing a lot of finished goods failures, there's a lot of scrap and the problem then was, "Hey, predict the failures, "reduce my scrap, save the money", right? Because they've been seeing a lot of failures every single day, we did not need a lot of data to train and create a model to that. So, in fact, we just needed one hour's worth of data. We created a model, put the thing, we have reduced, completely eliminated their scrap. There are other kinds of models, other kinds of models of video, where we can't do that in the edge, so we're required for example, some video files or simulated audio files, take it to an offline model, create the model, and see whether it's accurately predicting based on the real-time video coming in or not. So it's a mix of what we're seeing between those two. >> Well Sastry, thank you so much for stopping by theCUBE and sharing what it is that you guys at FogHorn are doing, what you're hearing from customers, how you're working together with them to solve some of these pretty significant challenges. >> Absolutely, it's been a pleasure. Hopefully this was helpful, and yeah. >> Definitely, very educational. We want to thank you for watching theCUBE, I'm Lisa Martin with George Gilbert. We are live at our event, Big Data SV in downtown San Jose. Come stop by Forager Tasting Room, hang out with us, learn as much as we are about all the layers of big data digital transformation and the opportunities. Stick around, we will be back after a short break. (upbeat electronic music)
SUMMARY :
brought to you by SiliconANGLE Media down the street from the Strata Data Conference. what do you guys do, who are you? Obviously in the process, you know, the new business outcomes you could build on it, What's the FogHorn secret sauce that others Before I directly answer the question, if you don't mind, how constrained an environment you can operate in. but that's the kind of environment we're talking about. So that's the kind of size we're talking about. on the other thing you said, with, and refining the gas and all of that. the Cloud if you needed to do retraining? Import and bring the model back If the model is running ultimately on the device, These days, most of the PLCs, programmable controllers, if it doesn't have the connectivity USB stick, bring it to the PLC device and upload the model. we destroyed the Iranian centrifuges. but the devices have the ability to connect to the Cloud. you don't want the Cloud to reach the jet engine. but the Cloud cannot reach the jet engine. So Sastry, as a CTO you meet with customers often. they're going to look at you like, and that it doesn't have the analytics built in -- or the real-time, or the low footprint. So that when you subset, when you determine If you have a little bit of a local storage, So, you had also said something interesting So the way we start, where rather than trying that tells you the optimal answer? and the problem then was, "Hey, predict the failures, and sharing what it is that you guys at FogHorn are doing, Hopefully this was helpful, and yeah. We want to thank you for watching theCUBE,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
George Gilbert | PERSON | 0.99+ |
George | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Saudi Arabia | LOCATION | 0.99+ |
Sastry Malladi | PERSON | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
one hour | QUANTITY | 0.99+ |
Sastry | PERSON | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
GE | ORGANIZATION | 0.99+ |
100 megabytes | QUANTITY | 0.99+ |
Lisa | PERSON | 0.99+ |
Bill Joy | PERSON | 0.99+ |
two | QUANTITY | 0.99+ |
FogHorn | ORGANIZATION | 0.99+ |
last week | DATE | 0.99+ |
Mountain View | LOCATION | 0.99+ |
30 more seconds | QUANTITY | 0.99+ |
David Floor | PERSON | 0.99+ |
one question | QUANTITY | 0.99+ |
Hortonworks | ORGANIZATION | 0.99+ |
San Jose | LOCATION | 0.99+ |
30 plus years | QUANTITY | 0.99+ |
SiliconANGLE Media | ORGANIZATION | 0.99+ |
three plus years ago | DATE | 0.99+ |
one customer | QUANTITY | 0.98+ |
one | QUANTITY | 0.98+ |
second | QUANTITY | 0.98+ |
C plus plus | TITLE | 0.98+ |
One | QUANTITY | 0.98+ |
theCUBE | ORGANIZATION | 0.98+ |
150 megabytes | QUANTITY | 0.98+ |
two ways | QUANTITY | 0.97+ |
Strata Data Conference | EVENT | 0.97+ |
Iranian | OTHER | 0.97+ |
five levels | QUANTITY | 0.95+ |
millions of elevators | QUANTITY | 0.95+ |
about less than 100 | QUANTITY | 0.95+ |
one part | QUANTITY | 0.94+ |
Vel | OTHER | 0.94+ |
One thing | QUANTITY | 0.92+ |
dozens of machinery models | QUANTITY | 0.92+ |
each | QUANTITY | 0.91+ |
Intel | ORGANIZATION | 0.91+ |
FogHorn | PERSON | 0.86+ |
2018 | DATE | 0.85+ |
first thing | QUANTITY | 0.85+ |
single-core | QUANTITY | 0.85+ |
NiFi | ORGANIZATION | 0.82+ |
Power by the Hour | ORGANIZATION | 0.81+ |
about three years ago | DATE | 0.81+ |
Forager Tasting R | ORGANIZATION | 0.8+ |
a ton | QUANTITY | 0.8+ |
CTO | PERSON | 0.79+ |
multibillion dollar | QUANTITY | 0.79+ |
Data | EVENT | 0.79+ |
Bell-Heads | ORGANIZATION | 0.78+ |
every single day | QUANTITY | 0.76+ |
The Cube | ORGANIZATION | 0.75+ |
Cloud | COMMERCIAL_ITEM | 0.73+ |
Dozens of machinery algorithms | QUANTITY | 0.71+ |
Pi | COMMERCIAL_ITEM | 0.71+ |
petabytes | QUANTITY | 0.7+ |
raspberry | ORGANIZATION | 0.69+ |
Big Data | ORGANIZATION | 0.68+ |
Cloud | TITLE | 0.67+ |
dual-core | QUANTITY | 0.65+ |
Sastry | ORGANIZATION | 0.62+ |
Net | ORGANIZATION | 0.61+ |
Chris Selland, Unifi Software | Big Data SV 2018
>> Voiceover: Live from San Jose, it's The Cube. Presenting Big Data Silicon Valley, brought to you by SiliconANGLE Media and its ecosystem partners. >> Welcome back to The Cube, our continuing coverage of our event, Big Data SV. We're on day two of this event. I'm Lisa Martin, with George Gilbert. We've had a great day yesterday learning a lot and really peeling back the layers of big data, looking at it from different perspectives, from challenges to opportunities. Joining us next is one of our Cube alumni, Chris Selland, the VP of Strategic Alliances from Unifi Software. Chris, great to meet you, welcome back! >> Thank you Lisa, it's great to be here. I have to say, as a alumni and a many time speaker, this venue is spectacular. Congratulations on the growth of The Cube, and this is an awesome venue. I've been on The Cube a bunch of times and this is as nice as I've ever seen it, >> Yeah, this is pretty cool, >> Onward and upward. This place is great. Isn't it cool? >> It really is. This is our 10th Big Data event, we've been having five now in San Jose, do our fifth one in New York City in the fall, and it's always interesting because we get the chance, George and I, and the other hosts, to really look at what is going on from different perspectives in the industry of big data. So before we kind of dig into that, tell us a little bit about Unifi Software, what do you guys do, what is unique and differentiating about Unifi. >> Sure, yeah, so I joined Unifi a little over a year ago. You know, I was attracted to the company because it really, I think, is aligned with where the market is going, and Peter talked this morning, Peter Burris was talking this morning about networks of data. Unifi is fundamentally a data catalog and data preparation platform, kind of combined or unified together. So, you know, so people say, "What do you do?" We're a data catalog with integrated data preparation. And the idea behind that, to go to Peter's, you know, mention of networks of data, is that data is becoming more and more distributed in terms of where it is, where it lives, where it sits. This idea of we're going to put everything in the data warehouse, and then we're going to put everything in the data lake, well, in reality, some of the data's in the warehouse, some of the data's in the lake, some of the data's in SAS applications, some of the data's in blob storage. And where is all of that data, what is it, and what can I do with it, that's really the fundamental problem that we solve. And, by the way, solve it for business people, because it's not just data scientists anymore, it's really going out into the entire business community now, you know, marketing people, operations people, finance people, they need data to do their jobs. Their jobs are becoming more data driven, but they're not necessarily data people. They don't know what schemas are, or joins are, but they know, "I need better data "to be able to do my job more effectively." So that's really what we're helping with. So, Chris, this is, it's kind of interesting, if you distill, you know, the capability down to the catalog and the prep-- >> Chris: Yep. So that it's ready for a catalog, but that sort of thing is, it's like investment in infrastructure, in terms of like building the highway system, but there're going to be, you know, for those early highways, there's got to be roots that you, a reason to build them out. What are some of those early use cases that justifies the investment in data infrastructure? >> There absolutely are, I mean, and by the way, those roots don't go away, those roots, you know, just like cities, right? New roots get built on top of them. So we're very much, you know, about, there's still data sitting in mainframes and legacy systems and you know, that data is absolutely critical for many large organizations. We do a lot of working in banking and financial services, and healthcare. They're still-- >> George: Are there common use cases that they start with? >> A lot of times-- >> Like, either by industry or just cross-sectional? >> Well, it's interesting, because, you know, analysts like yourselves have tended to put data catalog, which is a relatively new term, although some other big analyst firm that's having another conference this week, they were telling us recently that, starts with a "G," right? They were telling us that data catalog is now the number one search term they're getting. But it's been, by many annals, also kind of lumped in, lumped in's the wrong word, but incorporated with data governance. So traditionally, governance, another word that starts with "G," it's been the term. So, we often, we're not a traditional data governance platform, per se, but cataloging data has to have a foundation of security in governance. You know, think about what's going on in the world right now, both in the court of law and the court of public opinion, things like GDPR, right? So GDPR sort of says any customer data you have needs to be managed a certain way, with a certain level of sensitivity, and then there's other capabilities you need to open up to customers, like the right to be forgotten, so that means I need to have really good control, first of all, knowledge of, control over, and governance over my customer data. I talked about all those business people before. Certainly marketers are a great example. Marketers want all the customer data they can get, right? But there's social security numbers, PII, who should be able to see and use what? Because, if this data is used inappropriately, then it can cause a lot of problems. So, IT kind of sits in a-- they want to enable the business, but at the same time, there's a lot of risk there. So, anyway, going back to your question, you know, the catalog market is kind of evolved out of the governance market with more of a focus on kind of, you know, enabling the business, but making sure that it's done in a secure and well-governed way. >> George: Guard rails. >> Yes, guard rails, exactly, good way to say it. So, yep, that's good, I said about 500 words, and you distilled it to about two, right? Perfect, yep. >> So, in terms of your role in strategic alliances, tell us a little about some of the partnerships that Unifi is forging, to help customers understand where all this data is, to your point earlier, the different lines of business that need it to drive, identify where's their value, and drive the business forward, can actually get it. >> Absolutely, well, certainly to your point, our customers are our partners, and we can talk about some of them. But also, strategic alliances, we work very closely with a number of, you know, larger technology companies, Microsoft is a good example. We were actually part of the Microsoft Accelerator Program, which I think they've now rebranded Microsoft for Startups, but we've really been given tremendous support by that team, and we're doing a lot of work to, kind of, we're to some degree cloud agnostic, we support AWS, we support Azure, we support Google Cloud, but we're doing a lot of our development also on the Azure cloud platform. But you know, customers use all of the above, so we need to support all of the above. So Microsoft's a very close partner of ours. Another, I'll be in two weeks, and we've got some interesting news pending, which unfortunately I can't get into today, but maybe in a couple weeks, with Adobe. We're working very closely with them on their marketing cloud, their experience cloud, which is what they call their enterprise marketing cloud, which obviously, big, big focus on customer data, and then we've been working with a number of organizations and the sort of professional services system integration. We've had a lot of success with a firm called Access Group. We announced the partnership with them about two weeks ago. They've been a great partner for us, as well. So, you know, it's all about an ecosystem. Making customers successful is about getting an ecosystem together, so it's a really exciting place to be. >> So, Chris, it's actually interesting, it sounds like there's sort of a two classic routes to market. One is essentially people building your solution into theirs, whether it's an application or, you know, >> Chris: An enabling layer. >> Yes. >> Chris: Yes. >> Even higher layer. But with corporate developers, you know, it's almost like we spent years experimenting with these data lakes. But they were a little too opaque. >> Chris: Yes. >> And you know, it's not just that you provide the guard rails, but you also provide, sort of some transparency-- >> Chris: Yes. >> Into that. Have you seen a greater success rate within organizations who curate their data lakes, as opposed to those who, you know, who don't? >> Yes, absolutely. I think Peter said it very well in his presentation this morning, as well. That, you know, generally when you see data lake, we associate it with Hadoop. There are use cases that Hadoop is very good for, but there are others where it might not be the best fit. Which, to the early point about networks of data and distributed data, so companies that have, or organizations that have approached Hadoop with a "let's use it what it's good for," as opposed to "let's just dump "everything in there and figure it out later," and there have been a lot of the latter, but the former have done, generally speaking, a lot better, and that's what you're seeing. And we actually use Hadoop as a part of our platform, at least for the data preparation and transformation side of what we do. We use it in its enabling technology, as well. >> You know, it's funny, actually, when you talk about, as Peter talked about, networks of data versus centralized repositories. Scott Gnau, CTO of Hortonworks, was on yesterday, and he was talking about how he had originally come from Teradata, and that they had tried to do work, that he had tried to push them in the direction of recognizing that not all the analytic data was going to be in Teradata, you know, but they had to look more broadly with Hadapt, and I forgot what the rest of, you know-- >> Chris: Right, Aster, and-- >> Aster, yeah. >> Chris: Yes, exactly, yep. >> But what was interesting is that Hortonworks was moving towards the "we believe "everything is going to be in the data lake," but now, with their data plane service, they're talking about, you know, "We have to give you visibility and access." You mediate access to data everywhere. >> Chris: Right. >> So maybe help, so for folks who aren't, like, all bought into Hortonworks, for example, how much, you know, explain how you work relative to data plane service. >> Well, you know, maybe I could step back and give you a more general answer, because I agree with that philosophically, right? That, as I think we've been talking about here, with the networks of data, that goes back to my prior statement that there's, you know, there's different types of data platforms that have different use cases, and different types of solutions should be built on top of them, so things are getting more distributed. I think that, you know, Hortonworks, like every company, has to make the investments that are, as we are, making their customers successful. So, using Hadoop, and Hortonworks is one of our supported Hadoop platforms, we do work with them on engagements, but you know, it's all about making customers successful, ultimately. It's not about a particular product, it's about, you know, which data belongs in which location, and for what use case and what purpose, and then at the same time, when we're taking all of these different data sets and data sources, and cataloging them and preparing them and creating our output, where should we put that and catalog that, so we can create kind of a continuous improvement cycle, as well, and for those types-- >> A flywheel. >> A flywheel, exactly, continuous improvement flywheel, and for those types of purposes, you know, that's actually great use case for, you know, Hortonworks, Hadoop. That's a lot of what we typically use it for. We can actually put the data any place our customers define, but that's very often what we do with it, and then, but doing it in a very structured and organized way. As opposed to, you know, a lot of the early Hadoop, and not specific to any particular distro that went bad, were, it was just like, "Let's just dump it all "into Hadoop because it's cheaper." You know, "Let's, 'cause it's cheaper than the warehouse, "so let's just put it all in there, "and we'll figure what to do with it later." That's bad, but if you're using it in a structured way, it can be extremely useful. At the same point, and at the same time, not everything's going to go there belongs there, if you're being thoughtful about it. So you're seeing a lot more thoughtfulness these days, which is good. Which is good for customers, and it's good for us in the vendor side. Us, Hortonworks, everybody, so. >> So is there, maybe you can tell us of the different approaches to, like, the advantage of integrating the data prep with the catalogized service, because as soon as you're done with data prep it's visible within the catalog. >> Chris: Absolutely, that's one, yep. >> When, let's say when people do derive additional views into the data, how are they doing that in a way that then gets also registered back in the catalog, for further discovery? >> Yeah, well, having the integrated data preparation which is a huge differentiator from us, there are a lot of data catalog products out there, but our huge differentiator, one of them, is the fact that we have integrated data preparation. We don't have to hand off to another product, so that, as you said, gives us the ability to then catalog our output and build that flywheel, that continuous improvement flywheel, and it also just basically simplifies things for customers, hence our name. So, you know, it really kind of starts there. I think I, the second part of your question I didn't really, rewind back on that for me, it was-- >> Go ahead. >> Well, I'm not sure I remember it, right now, either. >> We all need more coffee. >> Exactly, we all need more coffee. >> So I'll ask you this last question, then. >> Yes, please. >> What are, so here we are in March 2018, what are you looking forward to, in terms of momentum and evolution of Unifi this year? >> Well, a lot of it, and tying into my role, I mentioned we will be at Adobe Summit in two weeks, so if you're going to be at Adobe Summit, come see us there, some of the work that we're doing with our partner, some of the events we're doing with people like Microsoft and Access, but really it's also just customer success, I mean, we're seeing tremendous momentum on the customer side, working with our customers, working with our partners, and again, as I mentioned, we're seeing so much more thoughtfulness in the market, these days, and less talk about, you know, the speeds and feeds, and more around business solutions. That's really also where our professional services, system integration partners, many of whom I've been with this week, really help, because they're building out solutions. You know, GDPR is coming in May, right? And you're starting to really see a groundswell of, okay, you know, and that's not about, you know, speeds and feeds. That's ultimately about making sure that I'm compliant with, you know, this huge regulatory environment. And at the same time, the court of public opinion is just as important. You know, we want to make sure that we're doing the right thing with data. Spread it throughout organization, make ourselves successful and make our customers successful. So, it's a lot of fun. >> That's, fun is good. >> Exactly, fun is good. >> Well, we thank you so much, Chris, for stopping back by The Cube and sharing your insights, what you're hearing in the big data industry, and some of the momentum that you're looking forward to carrying throughout the year. >> It's always a pleasure, and you, too. So, love the venue. >> Lisa: All right. >> Thank you, Lisa, thank you, George. >> Absolutely. We want to thank you for watching The Cube. You're watching our coverage of our event, Big Data SV, hashtag BigDataSV, for George, I almost said George Martin. For George Gilbert. >> George: I wish. >> George R.R., yeah. You would not be here if you were George R.R. Martin. >> George: No, I wouldn't. >> That was a really long way to say thank you for watching. I'm Lisa Martin, for this George. Stick around, we'll be right back with our next guest. (techno music)
SUMMARY :
brought to you by SiliconANGLE Media and really peeling back the layers of big data, Thank you Lisa, it's great to be here. Onward and upward. George and I, and the other hosts, So, you know, so people say, "What do you do?" you know, for those early highways, and legacy systems and you know, with more of a focus on kind of, you know, and you distilled it to about two, right? and drive the business forward, can actually get it. So, you know, it's all about an ecosystem. or, you know, But with corporate developers, you know, as opposed to those who, you know, who don't? That, you know, generally when you see data lake, and I forgot what the rest of, you know-- yeah. "We have to give you visibility and access." how much, you know, explain how you work to my prior statement that there's, you know, and for those types of purposes, you know, So is there, maybe you can tell us So, you know, it really kind of starts there. and that's not about, you know, speeds and feeds. Well, we thank you so much, Chris, So, love the venue. We want to thank you for watching The Cube. You would not be here if you were George R.R. That was a really long way to say thank you for watching.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
George Gilbert | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Chris | PERSON | 0.99+ |
Peter | PERSON | 0.99+ |
Chris Selland | PERSON | 0.99+ |
George | PERSON | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Scott Gnau | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
March 2018 | DATE | 0.99+ |
Adobe | ORGANIZATION | 0.99+ |
San Jose | LOCATION | 0.99+ |
Peter Burris | PERSON | 0.99+ |
Unifi | ORGANIZATION | 0.99+ |
New York City | LOCATION | 0.99+ |
George R.R. Martin | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Unifi Software | ORGANIZATION | 0.99+ |
May | DATE | 0.99+ |
Teradata | ORGANIZATION | 0.99+ |
George Martin | PERSON | 0.99+ |
George R.R. | PERSON | 0.99+ |
Hortonworks | ORGANIZATION | 0.99+ |
Access Group | ORGANIZATION | 0.99+ |
yesterday | DATE | 0.99+ |
both | QUANTITY | 0.99+ |
five | QUANTITY | 0.99+ |
GDPR | TITLE | 0.99+ |
SiliconANGLE Media | ORGANIZATION | 0.98+ |
this week | DATE | 0.98+ |
Hadapt | ORGANIZATION | 0.98+ |
fifth one | QUANTITY | 0.98+ |
about 500 words | QUANTITY | 0.98+ |
Hadoop | TITLE | 0.98+ |
Adobe Summit | EVENT | 0.98+ |
one | QUANTITY | 0.98+ |
two weeks | QUANTITY | 0.98+ |
One | QUANTITY | 0.96+ |
Aster | PERSON | 0.96+ |
this morning | DATE | 0.96+ |
this year | DATE | 0.95+ |
two weeks ago | DATE | 0.95+ |
today | DATE | 0.95+ |
The Cube | ORGANIZATION | 0.95+ |
Cube | ORGANIZATION | 0.93+ |
Big Data | EVENT | 0.91+ |
day two | QUANTITY | 0.91+ |
Access | ORGANIZATION | 0.9+ |
Yuanhao Sun, Transwarp | Big Data SV 2018
>> Announcer: Live, from San Jose, it's The Cube (light music) Presenting Big Data Silicon Valley. Brought to you by Silicon Angle Media, and its ecosystem partners. >> Hi, I'm Peter Burris and welcome back to Big Data SV, The Cube's, again, annual broadcast of what's happening in the big data marketplace here at, or adjacent to Strada here in San Jose. We've been broadcasting all day. We're going to be here tomorrow as well, over at the Forager eatery and place to come meander. So come on over. Spend some time with us. Now, we've had a number of great guests. Many of the thought leaders that are visiting here in San Jose today were on the big data marketplace. But I don't think any has traveled as far as our next guest. Yuanhao Sun is the ceo of Transwarp. Come all the way from Shanghai Yuanhao. It's once again great to see you on The Cube. Thank you very much for being here. >> Good to see you again. >> So Yuanhao, the Transwarp as a company has become extremely well known for great technology. There's a lot of reasons why that's the case, but you have some interesting updates on how the technology's being applied. Why don't you tell us what's going on? >> Okay, so, recently we announced the first order to the TPC-DS benchmark result. Our product, calling scepter, that is, SQL engine on top of Hadoop. We already add quite a lot of features, like dissre transactions, like a full SQL support. So that it can mimic, like oracle or the mutual, and also traditional database features so that we can pass the whole test. This single is also scalable, because it's distributed, scalable. So the large benchmark, like TPC-DS. It starts from 10 terabytes. SQL engine can pester without much trouble. >> So I know that there have been other firms that have claimed to pass TPCC-DS, but they haven't been audited. What does it mean to say you're audited? I'd presume that as a result, you've gone through some extremely stringent and specific tests to demonstrate that you can actually pass the entire suite. >> Yes, actually, there is a third party auditor. They already audit our test process and it results for the passed six, uh, five months. So it is fully audited. The reason why we can pass the test is because, actually, there's two major reasons for traditional databases. They are not scalable to the process large dataset. So they could not pass the test. For (mumbles) vendors, because the SQL engine, the features to reach enough to pass all the test. You know, there several steps in the benchmark, and the SQL queries, there are 99 queries, the syntax is not supported by all howve vendors yet. And also, the benchmark required to upload the data, after the queries, and then we run the queries for multiple concurrent users. That means you have to support disputed transactions. You have to make the upload data consistent. For howve vendors, the SQL engine on Hadoop. They haven't implemented the de-switch transaction capabilities. So that's why they failed to pass the benchmark. >> So I had the honor of traveling to Shanghai last year and going and speaking at your user conference and was quite impressed with the energy that was in the room as you announced a large number of new products. You've been very focused on taking what open source has to offer but adding significant value to it. As you said, you've done a lot with the SQL interfaces and various capabilities of SQL on top of Hadoop. Where is Transwarp going with its products today? How is it expanding? How is it being organizing? How is it being used? >> We group these products into three catalog, including big data, cloud, AI and the machine learning. So there are three categories. The big data, we upgrade the SQL engine, the stream engine, and we have a set of tools called adjustable studio to help people to streamline the big data operations. And the second part I lie is data cloud. We call it transwarp data cloud. So this product is going to be raised in early in May this year. So this product we build this product on top of common idiots. We provide how to buy the service, get a sense as service, air as a service to customers. A lot of people took credit multiple tenets. And they turned as isolated by network, storage, cpu. They free to create a clusters and speeding up on turning it off. So it can also scale hundreds of cost. So this is the, I think this is the first we implement, like, a network isolation and sweaty percendency in cobinets. So that it can support each day affairs and all how to components. And because it is elastic, just like car computing, but we run on bare model, people can consult the data, consult the applications in one place. Because all application and Hadoop components are conternalized, that means, we are talking images. We can spend up a very quickly and scale through a larger cluster. So this data cloud product is very interesting for large company, because they usually have a small IT team. But they have to provide a (mumbles), and a machine only capability to larger groups, like one found the people. So they need a convenient way to manage all these bigger clusters. And they have to isolate the resources. Even they need a bidding system. So this product is, we already have few big names in China, like China Post, Picture Channel, and Secret of Source Channel. So they are already applying this data cloud for their internal customers. >> And China has a, has a few people, so I presume that, you know, China Post for example, is probably a pretty big implementation. >> Yes so, they have a, but the IT team is, like less than 100 people, but they have to support thousands of users. So that's why they, you usually would deploy 100 cluster for each application, right, but today, for large organization, they have lots of applications. They hope to leverage big data capability, but a very small team, IT team, can also part of so many applications. So they need a convenient the way like a, just like when you put Hadoop on public cloud. We provide a product that allows you to provide a hardware service in private cloud on bare model machines. So this is the second product category. And the third is the machine learning and artificial intelligence. We provide a data sales platform, a machine learning tool, that is, interactive tools that allows people to create the machine only pipelines and models. We even implemented some automatic modeling capability that allow you to, to fisher in youring automatically or seeming automatically and to select the best items for you so that the machine learning can be, so everyone can be at Los Angeles. So they can use our tool to quickly create a models. And we also have some probuter models for different industry, like financial service, like banks, security companies, even iot. So we have different probuter machine only models for them. We just need to modify the template, then apply the machine only models to the applications very quickly. So that probably like a lesson, for example, for a bank customer, they just use it to deploy a model in one week. This is very quick for them. Otherwise, in the past, they have a company to build that application, to develop much models. They usually takes several months. Today it is much faster. So today we have three categories, particularly like cloud and machine learning. >> Peter Burris: Machine learning and AI. >> And so three products. >> And you've got some very, very big implementations. So you were talking about a couple of banks, but we were talking, before we came on, about some of the smart cities. >> Yuanhao Sun: Right. Kinds of things that you guys are doing at enormous scale. >> Yes, so we deploy our streaming productor for more than 300 cities in China. So this cluster is like connected together. So we use streaming capability to monitor the traffic and send the information from city to the central government. So all the, the sort of essential repoetry. So whenever illegal behavior on the road is detected, that information will be sent to the policeman, or the central repoetry within two second. Whenever you are seen by the camera in any place in China, their loads where we send out within two seconds. >> So the bad behavior is detected. It's identified as the location. The system also knows where the nearest police person is. And it sends a message and says, this car has performed something bad. >> Yeah and you should stop that car in the next station or in the next crossroad. Today there are tens of thousands policeman. They depends on this system for their daily work. >> Peter Burris: Interesting. >> So, just a question on, it sounds like one of your, sort of nearest competitors, in terms of, let's take the open source community, at least the APIs, and in their case open source, Waway. Have their been customers that tried to do a POC with you and with Waway, and said, well it took four months using the pure open source stuff, and it took, say, two weeks with your stack having, being much broader and deeper? Are any examples like that? >> There are quite a lot. We have more macro-share, like in financial services, we have about 100 bank users. So if we take all banks into account, for them they already use Hadoop. So we, our macro-share is above 60%. >> George Gilbert: 60. >> Yeah, in financial services. We usually do POC and, like run benchmarks. They are real workloads and usually it takes us three days or one week. They can found, we can speed up their workload very quickly. For Bank of China, they might go to their oracle workload to our platform. And they test our platform and the huave platform too. So the first thing is they cannot marry the whole oracle workload to open source Hadoop, because the missing features. We are able to support all this workloads with very minor modifications. So the modification takes only several hours. And we can finish the whole workload within two hours, but originally they take, usually take oracle more than one day, >> George Gilbert: Wow. >> more than ten hours to finish the workload. So it is very easy to see the benefits quickly. >> Now the you have a streaming product also with that same SQL interface. Are you going to see a migration of applications that used to be batch to more near real time or continuous, or will you see a whole new set of applications that weren't done before, because the latency wasn't appropriate? >> For streaming applications, real time cases they are mostly new applications, but if we are using storm api or spark streaming api, it is not so easy to develop your applications. And another issue is once you detect one new rule, you had to add those rules dynamically to your cluster. So to add to your printer, they do not have so many knowledge of writing scholar codes. They only know how to configure. Probably they are familiar with c-code. They just need to add one SQL statement to add a new rule. So that they can. >> In your system. >> Yeah, in our system. So it is much easier for them to program streaming applications. And for those customers who they don't have real time equations, they hope to do, like a real time data warehousing. They collect all this data from websites from their censors, like Petrol Channel, an oil company, the large oil company. They collect all the (mumbles) information directly to our streaming product. In the past, they just accredit to oracle and around the dashboard. So it only takes hours to see the results. But today, the application can be moved through our streaming product with only a few modifications, because they are all SQL statements. And this application becomes the real time. They can see the real time dashboard results in several seconds. >> So Yuanhao, you're number one in China. You're moving more aggressively to participate in the US market. What's the, last question, what's the biggest difference between being number one in China, the way that big data is being done in China versus the way you're encountering big data being done here, certainly in the US, for example? Is there a difference? >> I think there are some difference. Some a seem, katsumoto usually request a POC. But in China, they usually, I think they focus more on the results. They focus on what benefit they can gain from your product. So we have to prove them. So we have to hip them to my great application to see the benefits. I think in US, they focus more on technology than Chinese customers. >> Interesting, so they're more on technology here in the US, more in the outcome in China. Once again, Yuanhao Sun, from, ceo of Transwarp, thank you very much for being on The Cube. >> Thank you. And I'm Peter Burris with George Gilbert, my co-host, and we'll be back with more from big data SV, in San Jose. Come on over to the Forager, and spend some time with us. And we'll be back in a second. (light music)
SUMMARY :
Brought to you by Silicon Angle Media, over at the Forager eatery and place to come meander. So Yuanhao, the Transwarp as a company has become So that it can mimic, like oracle or the mutual, to demonstrate that you can actually pass the entire suite. And also, the benchmark required to upload the data, So I had the honor of traveling to Shanghai last year So this product is going to be raised you know, China Post for example, and to select the best items for you So you were talking about a couple of banks, Kinds of things that you guys are doing at enormous scale. from city to the central government. So the bad behavior is detected. or in the next crossroad. and it took, say, two weeks with your stack having, So if we take all banks into account, So the first thing is they cannot more than ten hours to finish the workload. Now the you have a streaming product also So to add to your printer, So it only takes hours to see the results. to participate in the US market. So we have to prove them. in the US, more in the outcome in China. Come on over to the Forager, and spend some time with us.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Peter Burris | PERSON | 0.99+ |
Shanghai | LOCATION | 0.99+ |
George Gilbert | PERSON | 0.99+ |
US | LOCATION | 0.99+ |
China | LOCATION | 0.99+ |
99 queries | QUANTITY | 0.99+ |
three days | QUANTITY | 0.99+ |
two weeks | QUANTITY | 0.99+ |
Silicon Angle Media | ORGANIZATION | 0.99+ |
five months | QUANTITY | 0.99+ |
San Jose | LOCATION | 0.99+ |
China Post | ORGANIZATION | 0.99+ |
Picture Channel | ORGANIZATION | 0.99+ |
one week | QUANTITY | 0.99+ |
six | QUANTITY | 0.99+ |
four months | QUANTITY | 0.99+ |
Los Angeles | LOCATION | 0.99+ |
10 terabytes | QUANTITY | 0.99+ |
last year | DATE | 0.99+ |
today | DATE | 0.99+ |
Today | DATE | 0.99+ |
tomorrow | DATE | 0.99+ |
more than one day | QUANTITY | 0.99+ |
more than 300 cities | QUANTITY | 0.99+ |
second part | QUANTITY | 0.99+ |
two hours | QUANTITY | 0.99+ |
less than 100 people | QUANTITY | 0.99+ |
more than ten hours | QUANTITY | 0.99+ |
Waway | ORGANIZATION | 0.99+ |
Bank of China | ORGANIZATION | 0.99+ |
third | QUANTITY | 0.99+ |
Hadoop | TITLE | 0.99+ |
Petrol Channel | ORGANIZATION | 0.99+ |
three products | QUANTITY | 0.98+ |
one new rule | QUANTITY | 0.98+ |
hundreds | QUANTITY | 0.98+ |
three categories | QUANTITY | 0.98+ |
SQL | TITLE | 0.98+ |
single | QUANTITY | 0.98+ |
Transwarp | ORGANIZATION | 0.98+ |
first | QUANTITY | 0.98+ |
tens of thousands policeman | QUANTITY | 0.98+ |
Yuanhao Sun | ORGANIZATION | 0.98+ |
each application | QUANTITY | 0.98+ |
two seconds | QUANTITY | 0.98+ |
100 cluster | QUANTITY | 0.97+ |
first thing | QUANTITY | 0.97+ |
about 100 bank users | QUANTITY | 0.97+ |
two second | QUANTITY | 0.97+ |
each day | QUANTITY | 0.97+ |
Big Data SV | ORGANIZATION | 0.97+ |
The Cube | ORGANIZATION | 0.96+ |
two major reasons | QUANTITY | 0.95+ |
one | QUANTITY | 0.95+ |
above 60% | QUANTITY | 0.95+ |
early in May this year | DATE | 0.94+ |
Source Channel | ORGANIZATION | 0.93+ |
Big Data | ORGANIZATION | 0.92+ |
Chinese | OTHER | 0.9+ |
Strada | LOCATION | 0.89+ |
second product category | QUANTITY | 0.88+ |
Seth Dobrin, IBM | Big Data SV 2018
>> Announcer: Live from San Jose, it's theCUBE. Presenting Big Data Silicon Valley, brought to you by SiliconANGLE Media and it's ecosystem partners. >> Welcome back to theCUBE's continuing coverage of our own event, Big Data SV. I'm Lisa Martin, with my cohost Dave Vellante. We're in downtown San Jose at this really cool place, Forager Eatery. Come by, check us out. We're here tomorrow as well. We're joined by, next, one of our CUBE alumni, Seth Dobrin, the Vice President and Chief Data Officer at IBM Analytics. Hey, Seth, welcome back to theCUBE. >> Hey, thanks for having again. Always fun being with you guys. >> Good to see you, Seth. >> Good to see you. >> Yeah, so last time you were chatting with Dave and company was about in the fall at the Chief Data Officers Summit. What's kind of new with you in IBM Analytics since then? >> Yeah, so the Chief Data Officers Summit, I was talking with one of the data governance people from TD Bank and we spent a lot of time talking about governance. Still doing a lot with governance, especially with GDPR coming up. But really started to ramp up my team to focus on data science, machine learning. How do you do data science in the enterprise? How is it different from doing a Kaggle competition, or someone getting their PhD or Masters in Data Science? >> Just quickly, who is your team composed of in IBM Analytics? >> So IBM Analytics represents, think of it as our software umbrella, so it's everything that's not pure cloud or Watson or services. So it's all of our software franchise. >> But in terms of roles and responsibilities, data scientists, analysts. What's the mixture of-- >> Yeah. So on my team I have a small group of people that do governance, and so they're really managing our GDPR readiness inside of IBM in our business unit. And then the rest of my team is really focused on this data science space. And so this is set up from the perspective of we have machine-learning engineers, we have predictive-analytics engineers, we have data engineers, and we have data journalists. And that's really focus on helping IBM and other companies do data science in the enterprise. >> So what's the dynamic amongst those roles that you just mentioned? Is it really a team sport? I mean, initially it was the data science on a pedestal. Have you been able to attack that problem? >> So I know a total of two people that can do that all themselves. So I think it absolutely is a team sport. And it really takes a data engineer or someone with deep expertise in there, that also understands machine-learning, to really build out the data assets, engineer the features appropriately, provide access to the model, and ultimately to what you're going to deploy, right? Because the way you do it as a research project or an activity is different than using it in real life, right? And so you need to make sure the data pipes are there. And when I look for people, I actually look for a differentiation between machine-learning engineers and optimization. I don't even post for data scientists because then you get a lot of data scientists, right? People who aren't really data scientists, and so if you're specific and ask for machine-learning engineers or decision optimization, OR-type people, you really get a whole different crowd in. But the interplay is really important because most machine-learning use cases you want to be able to give information about what you should do next. What's the next best action? And to do that, you need decision optimization. >> So in the early days of when we, I mean, data science has been around forever, right? We always hear that. But in the, sort of, more modern use of the term, you never heard much about machine learning. It was more like stats, math, some programming, data hacking, creativity. And then now, machine learning sounds fundamental. Is that a new skillset that the data scientists had to learn? Did they get them from other parts of the organization? >> I mean, when we talk about math and stats, what we call machine learning today has been what we've been doing since the first statistics for years, right? I mean, a lot of the same things we apply in what we call machine learning today I did during my PhD 20 years ago, right? It was just with a different perspective. And you applied those types of, they were more static, right? So I would build a model to predict something, and it was only for that. It really didn't apply it beyond, so it was very static. Now, when we're talking about machine learning, I want to understand Dave, right? And I want to be able to predict Dave's behavior in the future, and learn how you're changing your behavior over time, right? So one of the things that a lot of people don't realize, especially senior executives, is that machine learning creates a self-fulfilling prophecy. You're going to drive a behavior so your data is going to change, right? So your model needs to change. And so that's really the difference between what you think of as stats and what we think of as machine learning today. So what we were looking for years ago is all the same we just described it a little differently. >> So how fine is the line between a statistician and a data scientist? >> I think any good statistician can really become a data scientist. There's some issues around data engineering and things like that but if it's a team sport, I think any really good, pure mathematician or statistician could certainly become a data scientist. Or machine-learning engineer. Sorry. >> I'm interested in it from a skillset standpoint. You were saying how you're advertising to bring on these roles. I was at the Women in Data Science Conference with theCUBE just a couple of days ago, and we hear so much excitement about the role of data scientists. It's so horizontal. People have the opportunity to make impact in policy change, healthcare, etc. So the hard skills, the soft skills, mathematician, what are some of the other elements that you would look for or that companies, enterprises that need to learn how to embrace data science, should look for? Someone that's not just a mathematician but someone that has communication skills, collaboration, empathy, what are some of those, openness, to not lead data down a certain, what do you see as the right mix there of a data scientist? >> Yeah, so I think that's a really good point, right? It's not just the hard skills. When my team goes out, because part of what we do is we go out and sit with clients and teach them our philosophy on how you should integrate data science in the enterprise. A good part of that is sitting down and understanding the use case. And working with people to tease out, how do you get to this ultimate use case because any problem worth solving is not one model, any use case is not one model, it's many models. How do you work with the people in the business to understand, okay, what's the most important thing for us to deliver first? And it's almost a negotiation, right? Talking them back. Okay, we can't solve the whole problem. We need to break it down in discreet pieces. Even when we break it down into discreet pieces, there's going to be a series of sprints to deliver that. Right? And so having these soft skills to be able to tease that in a way, and really help people understand that their way of thinking about this may or may not be right. And doing that in a way that's not offensive. And there's a lot of really smart people that can say that, but they can come across at being offensive, so those soft skills are really important. >> I'm going to talk about GDPR in the time we have remaining. We talked about in the past, the clocks ticking, May the fines go into effect. The relationship between data science, machine learning, GDPR, is it going to help us solve this problem? This is a nightmare for people. And many organizations aren't ready. Your thoughts. >> Yeah, so I think there's some aspects that we've talked about before. How important it's going to be to apply machine learning to your data to get ready for GDPR. But I think there's some aspects that we haven't talked about before here, and that's around what impact does GDPR have on being able to do data science, and being able to implement data science. So one of the aspects of the GDPR is this concept of consent, right? So it really requires consent to be understandable and very explicit. And it allows people to be able to retract that consent at any time. And so what does that mean when you build a model that's trained on someone's data? If you haven't anonymized it properly, do I have to rebuild the model without their data? And then it also brings up some points around explainability. So you need to be able to explain your decision, how you used analytics, how you got to that decision, to someone if they request it. To an auditor if they request it. Traditional machine learning, that's not too much of a problem. You can look at the features and say these features, this contributed 20%, this contributed 50%. But as you get into things like deep learning, this concept of explainable or XAI becomes really, really important. And there were some talks earlier today at Strata about how you apply machine learning, traditional machine learning to interpret your deep learning or black box AI. So that's really going to be important, those two things, in terms of how they effect data science. >> Well, you mentioned the black box. I mean, do you think we'll ever resolve the black box challenge? Or is it really that people are just going to be comfortable that what happens inside the box, how you got to that decision is okay? >> So I'm inherently both cynical and optimistic. (chuckles) But I think there's a lot of things we looked at five years ago and we said there's no way we'll ever be able to do them that we can do today. And so while I don't know how we're going to get to be able to explain this black box as a XAI, I'm fairly confident that in five years, this won't even be a conversation anymore. >> Yeah, I kind of agree. I mean, somebody said to me the other day, well, it's really hard to explain how you know it's a dog. >> Seth: Right (chuckles). But you know it's a dog. >> But you know it's a dog. And so, we'll get over this. >> Yeah. >> I love that you just brought up dogs as we're ending. That's my favorite thing in the world, thank you. Yes, you knew that. Well, Seth, I wish we had more time, and thanks so much for stopping by theCUBE and sharing some of your insights. Look forward to the next update in the next few months from you. >> Yeah, thanks for having me. Good seeing you again. >> Pleasure. >> Nice meeting you. >> Likewise. We want to thank you for watching theCUBE live from our event Big Data SV down the street from the Strata Data Conference. I'm Lisa Martin, for Dave Vellante. Thanks for watching, stick around, we'll be rick back after a short break.
SUMMARY :
brought to you by SiliconANGLE Media Welcome back to theCUBE's continuing coverage Always fun being with you guys. Yeah, so last time you were chatting But really started to ramp up my team So it's all of our software franchise. What's the mixture of-- and other companies do data science in the enterprise. that you just mentioned? And to do that, you need decision optimization. So in the early days of when we, And so that's really the difference I think any good statistician People have the opportunity to make impact there's going to be a series of sprints to deliver that. in the time we have remaining. And so what does that mean when you build a model Or is it really that people are just going to be comfortable ever be able to do them that we can do today. I mean, somebody said to me the other day, But you know it's a dog. But you know it's a dog. I love that you just brought up dogs as we're ending. Good seeing you again. We want to thank you for watching theCUBE
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Seth | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Seth Dobrin | PERSON | 0.99+ |
20% | QUANTITY | 0.99+ |
50% | QUANTITY | 0.99+ |
TD Bank | ORGANIZATION | 0.99+ |
San Jose | LOCATION | 0.99+ |
two people | QUANTITY | 0.99+ |
tomorrow | DATE | 0.99+ |
IBM Analytics | ORGANIZATION | 0.99+ |
two things | QUANTITY | 0.99+ |
SiliconANGLE Media | ORGANIZATION | 0.99+ |
one model | QUANTITY | 0.99+ |
five years | QUANTITY | 0.98+ |
20 years ago | DATE | 0.98+ |
Big Data SV | EVENT | 0.98+ |
five years ago | DATE | 0.98+ |
GDPR | TITLE | 0.98+ |
theCUBE | ORGANIZATION | 0.98+ |
one | QUANTITY | 0.98+ |
Strata Data Conference | EVENT | 0.97+ |
today | DATE | 0.97+ |
first statistics | QUANTITY | 0.95+ |
CUBE | ORGANIZATION | 0.94+ |
Women in Data Science Conference | EVENT | 0.94+ |
both | QUANTITY | 0.94+ |
Chief Data Officers Summit | EVENT | 0.93+ |
Big Data SV 2018 | EVENT | 0.93+ |
couple of days ago | DATE | 0.93+ |
years | DATE | 0.9+ |
Forager Eatery | ORGANIZATION | 0.9+ |
first | QUANTITY | 0.86+ |
Watson | TITLE | 0.86+ |
Officers Summit | EVENT | 0.74+ |
Data Officer | PERSON | 0.73+ |
SV | EVENT | 0.71+ |
President | PERSON | 0.68+ |
Strata | TITLE | 0.67+ |
Big Data | ORGANIZATION | 0.66+ |
earlier today | DATE | 0.65+ |
Silicon Valley | LOCATION | 0.64+ |
years | QUANTITY | 0.6+ |
Chief | EVENT | 0.44+ |
Kaggle | ORGANIZATION | 0.43+ |
Dr. Tendu Yogurtcu, Syncsort | Big Data SV 2018
>> Announcer: Live from San Jose, it's theCUBE. Presenting data, Silicon Valley brought to you by Silicon Angle Media and it's ecosystem partners. >> Welcome back to theCUBE. We are live in San Jose at our event, Big Data SV. I'm Lisa Martin, my co-host is George Gilbert and we are down the street from the Strata Data Conference. We are at a really cool venue: Forager Eatery Tasting Room. Come down and join us, hang out with us, we've got a cocktail par-tay tonight. We also have an interesting briefing from our analysts on big data trends tomorrow morning. I want to welcome back to theCUBE now one of our CUBE VIP's and alumna Tendu Yogurtcu, the CTO at Syncsort, welcome back. >> Thank you. Hello Lisa, hi George, pleasure to be here. >> Yeah, it's our pleasure to have you back. So, what's going on at Syncsort, what are some of the big trends as CTO that you're seeing? >> In terms of the big trends that we are seeing, and Syncsort has grown a lot in the last 12 months, we actually doubled our revenue, it has been really an successful and organic growth path, and we have more than 7,000 customers now, so it's a great pool of customers that we are able to talk and see the trends and how they are trying to adapt to the digital disruption and make data as part of their core strategy. So data is no longer an enabler, and in all of the enterprise we are seeing data becoming the core strategy. This reflects in the four mega trends, they are all connected to enable business as well as operational analytics. Cloud is one, definitely. We are seeing more and more cloud adoption, even our financial services healthcare and banking customers are now, they have a couple of clusters running in the cloud, in public cloud, multiple workloads, hybrid seems to be the new standard, and it comes with also challenges. IT governance as well as date governance is a major challenge, and also scoping and planning for the workloads in the cloud continues to be a challenge, as well. Our general strategy for all of the product portfolio is to have our products following design wants and deploy any of our strategy. So whether it's a standalone environment on Linux or running on Hadoop or Spark, or running on Premise or in the Cloud, regardless of the Cloud provider, we are enabling the same education with no changes to run all of these environments, including hybrid. Then we are seeing the streaming trend, with the connected devices with the digital disruption and so much data being generated, being able to stream and process data on the age, with the Internet of things, and in order to address the use cases that Syncsort is focused on, we are really providing more on the Change Data Capture and near real-time and real-time data replication to the next generation analytics environments and big data environments. We launched last year our Change Data Capture, CDC, product offering with data integration, and we continue to strengthen that vision merger we had data replication, real-time data replication capabilities, and we are now seeing even Kafka database becoming a consumer of this data. Not just keeping the data lane fresh, but really publishing the changes from multiple, diverse set of sources and publishing into a Kafka database and making it available for applications and analytics in the data pipeline. So the third trend we are seeing is around data science, and if you noticed this morning's keynote was all about machine learning, artificial intelligence, deep learning, how to we make use of data science. And it was very interesting for me because we see everyone talking about the challenge of how do you prepare the data and how do you deliver the the trusted data for machine learning and artificial intelligence use and deep learning. Because if you are using bad data, and creating your models based on bad data, then the insights you get are also impacted. We definitely offer our products, both on the data integration and data quality side, to prepare the data, cleanse, match, and deliver the trusted data set for data scientists and make their life easier. Another area of focus for 2018 is can we also add supervised learning to this, because with the premium quality domain experts that we have now in Syncsort, we have a lot of domain experts in the field, we can infuse the machine learning algorithms and connect data profiling capabilities we have with the data quality capabilities recommending business rules for data scientists and helping them automate the mandate tasks with recommendations. And the last but not least trend is data governance, and data governance is almost a umbrella focus for everything we are doing at Syncsort because everything about the Cloud trend, the streaming, and the data science, and developing that next generation analytics environment for our customers depends on the data governance. It is, in fact, a business imperative, and the regulatory compliance use cases drives more importance today than governance. For example, General Data Protection Regulation in Europe, GDPR. >> Lisa: Just a few months away. >> Just a few months, May 2018, it is in the mind of every C-level executive. It's not just for European companies, but every enterprise has European data sourced in their environments. So compliance is a big driver of governance, and we look at governance in multiple aspects. Security and issuing data is available in a secure way is one aspect, and delivering the high quality data, cleansing, matching, the example Hilary Mason this morning gave in the keynote about half of what the context matters in terms of searches of her name was very interesting because you really want to deliver that high quality data in the enterprise, trust of data set, preparing that. Our Trillium Quality for big data, we launched Q4, that product is generally available now, and actually we are in production with very large deployment. So that's one area of focus. And the third area is how do you create visibility, the farm-to-table view of your data? >> Lisa: Yeah, that's the name of your talk! I love that. >> Yes, yes, thank you. So tomorrow I have a talk at 2:40, March 8th also, I'm so happy it's on the Women's Day that I'm talking-- >> Lisa: That's right, that's right! Get a farm-to-table view of your data is the name of your talk, track data lineage from source to analytics. Tell us a little bit more about that. >> It's all about creating more visibility, because for audit reasons, for understanding how many copies of my data is created, valued my data had been, and who accessed it, creating that visibility is very important. And the last couple of years, we saw everyone was focused on how do I create a data lake and make my data accessible, break the data silos, and liberate my data from multiple platforms, legacy platforms that the enterprise might have. Once that happened, everybody started worrying about how do I create consumable data set and how do I manage this data because data has been on the legacy platforms like Mainframe, IMBI series has been on relational data stores, it is in the Cloud, gravity of data originating in the Cloud is increasing, it's originating from mobile. Hadoop vendors like Hortonworks and Cloudera, they are creating visibility to what happens within the Hadoop framework. So we are deepening our integration with the Cloud Navigator, that was our announcement last week. We already have integration both with Hortonworks and Cloudera Navigator, this is one step further where we actually publish what happened to every single granular level of data at the field level with all of the transformations that data have been through outside of the cluster. So that visibility is now published to Navigator itself, we also publish it through the RESTful API, so governance is a very strong and critical initiative for all of the businesses. And we are playing into security aspect as well as data lineage and tracking aspect and the quality aspect. >> So this sounds like an extremely capable infrastructure service, so that it's trusted data. But can you sell that to an economic buyer alone, or do you go in in conjunction with anther solution like anti-money laundering for banks or, you know, what are the key things that they place enough value on that they would spend, you know, budget on it? >> Yes, absolutely. Usually the use cases might originate like anti-money laundering, which is very common, fraud detection, and it ties to getting a single view of an entity. Because in anti-money laundering, you want to understand the single view of your customer ultimately. So there is usually another solution that might be in the picture. We are providing the visibility of the data, as well as that single view of the entity, whether it's the customer view in this case or the product view in some of the use cases by delivering the matching capabilities and the cleansing capabilities, the duplication capabilities in addition to the accessing and integrating the data. >> When you go into a customer and, you know, recognizing that we still have tons of silos and we're realizing it's a lot harder to put everything in one repository, how do customers tell you they want to prioritize what they're bringing into the repository or even what do they want to work on that's continuously flowing in? >> So it depends on the business use case. And usually at the time that we are working with the customer, they selected that top priority use case. The risk here, and the anti-money laundering, or for insurance companies, we are seeing a trend, for example, building the data marketplace, as that tantalize data marketplace concept. So depending on the business case, many of our insurance customers in US, for example, they are creating the data marketplace and they are working with near real-time and microbatches. In Europe, Europe seems to be a bit ahead of the game in some cases, like Hadoop production was slow but certainly they went right into the streaming use cases. We are seeing more directly streaming and keeping it fresh and more utilization of the Kafka and messaging frameworks and database. >> And in that case, where they're sort of skipping the batch-oriented approach, how do they keep track of history? >> It's still, in most of the cases, microbatches, and the metadata is still associated with the data. So there is an analysis of the historical what happened to that data. The tools, like ours and the vendors coming to picture, to keep track, of that basically. >> So, in other words, by knowing what happened operationally to the data, that paints a picture of a history. >> Exactly, exactly. >> Interesting. >> And for the governance we usually also partner, for example, we partner with Collibra data platform, we partnered with ASG for creating that business rules and technical metadata and providing to the business users, not just to the IT data infrastructure, and on the Hadoop side we partner with Cloudera and Hortonworks very closely to complete that picture for the customer, because nobody is just interested in what happened to the data in Hadoop or in Mainframe or in my relational data warehouse, they are really trying to see what's happening on Premise, in the Cloud, multiple clusters, traditional environments, legacy systems, and trying to get that big picture view. >> So on that, enabling a business to have that, we'll say in marketing, 360 degree view of data, knowing that there's so much potential for data to be analyzed to drive business decisions that might open up new business models, new revenue streams, increase profit, what are you seeing as a CTO of Syncsort when you go in to meet with a customer, data silos, when you're talking to a Chief Data Officer, what's the cultural, I guess, not shift but really journey that they have to go on to start opening up other organizations of the business, to have access to data so they really have that broader, 360 degree view? What's that cultural challenge that they have to, journey that they have to go on? >> Yes, Chief Data Officers are actually very good partners for us, because usually Chief Data Officers are trying to break the silos of data and make sure that the data is liberated for the business use cases. Still most of the time the infrastructure and the cluster, whether it's the deployment in the Cloud versus on Premise, it's owned by the IT infrastructure. And the lines of business are really the consumers and the clients of that. CDO, in that sense, almost mitigates and connects to those line of businesses with the IT infrastructure with the same goals for the business, right? They have to worry about the compliance, they have to worry about creating multiple copies of data, they have to worry about the security of the data and availability of the data, so CDOs actually help. So we are actually very good partners with the CDOs in that sense, and we also usually have IT infrastructure owner in the room when we are talking with our customers because they have a big stake. They are like the gatekeepers of the data to make sure that it is accessed by the right... By the right folks in the business. >> Sounds like maybe they're in the role of like, good cop bad cop or maybe mediator. Well Tendu, I wish we had more time. Thanks so much for coming back to theCUBE and, like you said, you're speaking tomorrow at Strata Conference on International Women's Day: Get a farm-to-table view of your data. Love the title. >> Thank you. >> Good luck tomorrow, and we look forward to seeing you back on theCUBE. >> Thank you, I look forward to coming back and letting you know about more exciting both organic innovations and acquisitions. >> Alright, we look forward to that. We want to thank you for watching theCUBE, I'm Lisa Martin with my co-host George Gilbert. We are live at our event Big Data SV in San Jose. Come down and visit us, stick around, and we will be right back with our next guest after a short break. >> Tendu: Thank you. (upbeat music)
SUMMARY :
brought to you by Silicon Angle Media and we are down the street from the Strata Data Conference. Hello Lisa, hi George, pleasure to be here. Yeah, it's our pleasure to have you back. and in all of the enterprise we are seeing data and delivering the high quality data, Lisa: Yeah, that's the name of your talk! it's on the Women's Day that I'm talking-- is the name of your talk, track data lineage and make my data accessible, break the data silos, that they place enough value on that they would and the cleansing capabilities, the duplication So it depends on the business use case. It's still, in most of the cases, operationally to the data, that paints a picture And for the governance we usually also partner, and the cluster, whether it's the deployment Love the title. to seeing you back on theCUBE. and letting you know about more exciting and we will be right back with our next guest Tendu: Thank you.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Lisa Martin | PERSON | 0.99+ |
George | PERSON | 0.99+ |
May 2018 | DATE | 0.99+ |
George Gilbert | PERSON | 0.99+ |
Syncsort | ORGANIZATION | 0.99+ |
Lisa | PERSON | 0.99+ |
Europe | LOCATION | 0.99+ |
Hortonworks | ORGANIZATION | 0.99+ |
US | LOCATION | 0.99+ |
Hilary Mason | PERSON | 0.99+ |
San Jose | LOCATION | 0.99+ |
ASG | ORGANIZATION | 0.99+ |
2018 | DATE | 0.99+ |
Tendu | PERSON | 0.99+ |
Silicon Angle Media | ORGANIZATION | 0.99+ |
Cloudera | ORGANIZATION | 0.99+ |
360 degree | QUANTITY | 0.99+ |
tomorrow | DATE | 0.99+ |
Collibra | ORGANIZATION | 0.99+ |
more than 7,000 customers | QUANTITY | 0.99+ |
last week | DATE | 0.99+ |
last year | DATE | 0.99+ |
tomorrow morning | DATE | 0.99+ |
one aspect | QUANTITY | 0.99+ |
third area | QUANTITY | 0.99+ |
Linux | TITLE | 0.99+ |
Cloud Navigator | TITLE | 0.99+ |
2:40 | DATE | 0.98+ |
Women's Day | EVENT | 0.98+ |
Tendu Yogurtcu | PERSON | 0.98+ |
GDPR | TITLE | 0.98+ |
Spark | TITLE | 0.97+ |
tonight | DATE | 0.97+ |
Big Data SV | EVENT | 0.97+ |
Kafka | TITLE | 0.97+ |
International Women's Day | EVENT | 0.97+ |
both | QUANTITY | 0.97+ |
CDC | ORGANIZATION | 0.96+ |
Navigator | TITLE | 0.96+ |
Strata Data Conference | EVENT | 0.96+ |
single view | QUANTITY | 0.96+ |
Hadoop | TITLE | 0.95+ |
third trend | QUANTITY | 0.95+ |
one step | QUANTITY | 0.95+ |
single view | QUANTITY | 0.95+ |
Dr. | PERSON | 0.94+ |
theCUBE | ORGANIZATION | 0.94+ |
CUBE | ORGANIZATION | 0.94+ |
this morning | DATE | 0.94+ |
Cloud | TITLE | 0.92+ |
last 12 months | DATE | 0.91+ |
Change Data Capture | ORGANIZATION | 0.9+ |
today | DATE | 0.9+ |
European | OTHER | 0.88+ |
last couple of years | DATE | 0.88+ |
General Data Protection Regulation in Europe | TITLE | 0.86+ |
Strata Conference | EVENT | 0.84+ |
one | QUANTITY | 0.83+ |
one repository | QUANTITY | 0.83+ |
tons of silos | QUANTITY | 0.82+ |
one area | QUANTITY | 0.82+ |
Q4 | DATE | 0.82+ |
Big Data SV 2018 | EVENT | 0.81+ |
four mega trends | QUANTITY | 0.76+ |
March 8th | DATE | 0.76+ |
Scott Gnau, Hortonworks | Big Data SV 2018
>> Narrator: Live from San Jose, it's the Cube. Presenting Big Data Silicon Valley. Brought to you by SiliconANGLE Media and its ecosystem partners. >> Welcome back to the Cube's continuing coverage of Big Data SV. >> This is out tenth Big Data event, our fifth year in San Jose. We are down the street from the Strata Data Conference. We invite you to come down and join us, come on down! We are at Forager Tasting Room & Eatery, super cool place. We've got a cocktail event tonight, and a endless briefing tomorrow morning. We are excited to welcome back to the Cube, Scott Gnau, the CTO of Hortonworks. Hey, Scott, welcome back. >> Thanks for having me, and I really love what you've done with the place. I think there's as much energy here as I've seen in the entire show. So, thanks for having me over. >> Yeah! >> We have done a pretty good thing to this place that we're renting for the day. So, thanks for stopping by and talking with George and I. So, February, Hortonworks announced some news about Hortonworks DataFlow. What was in that announcement? What does that do to help customers simplify data in motion? What industries is it going to be most impactful for? I'm thinking, you know, GDPR is a couple months away, kind of what's new there? >> Well, yeah, and there are a couple of topics in there, right? So, obviously, we're very committed to, which I think is one of our unique value propositions, is we're committed to really creating an easy to use data management platform, as it were, for the entire lifecycle of data, from one data created at the edge and as data are streaming from one place to another place, and, at rest, analytics get run, analytics get pushed back out to the edge. So, that entire lifecycle is really the footprint that we're looking at, and when you dig a level into that, obviously, the data in motion piece is usually important, and So I think one a the things that we've looked at is we don't want to be just a streaming engine or just a tool for creating pipes and data flows and so on. We really want to create that entire experience around what needs to happen for data that's moving, whether it be acquisition at the edge in a protected way with provenance and encryption, whether it be applying streaming analytics as the data are flowing and everywhere kind of in between, and so that's what HDF represents, and what we released in our latest release, which, to your point, was just a few weeks ago, is a way for our customers to go build their data in motion applications using a very simple drag and drop GUI interface. So, they don't have to understand all of the different animals in the zoo, and the different technologies that are in play. It's like, "I want to do this." Okay, here's a GUI tool, you can have all of the different operators that are represented by the different underlying technologies that we provide as Hortonworks DataFlow, and you can stream them together, and then, you can make those applications and test those applications. One of the biggest enhancements that we did, is we made it very easy then for once those things are built in a laptop environment or in a dev environment, to be published out to production or to be published out to other developers who might want to enhance them and so on. So, the idea is to make it consumable inside of an enterprise, and when you think about data in motion and IOT and all those use cases, it's not going to be one department, one organization, or one person that's doing it. It's going to be a team of people that are distributed just like the data and the sensors, and, so, being able to have that sharing capability is what we've enhanced in the experience. >> So, you were just saying, before we went live, that you're here having speed dates with customers. What are some of the things... >> It's a little bit more sincere than that, but yeah. >> (laughs) Isn't speed dating sincere? It's 2018, I'm not sure. (Scott laughs) What are some of the things that you're hearing from customers, and how is that helping to drive what's coming out from Hortonworks? >> So, the two things that I'm hearing right, number one, certainly, is that they really appreciate our approach to the entire lifecycle of data, because customers are really experiencing huge data volume increases and data just from everywhere, and it's no longer just from the ERP system inside the firewall. It's from third party, it's from Sensors, it's from mobile devices, and, so, they really do appreciate kind of the territory that we cover with the tools and technologies we bring to market, and, so, that's been very rewarding. Clearly, customers who are now well into this path, they're starting to think about, in this new world, data governance, and data governance, I just took all of the energy out of the room, governance, it sounds like, you know, hard. What I mean by data governance, really, is customers need to understand, with all of this diverse, connected data everywhere, in the cloud, on PRIM, then Sensors, third party, partners, is, frankly, they need a trail of breadcrumbs that say what is it, where'd it come from, who had access to it, and then, what did they do with it? If you start to piece that together, that's what they really need to understand, the data estate that belongs to them, so they can turn that into refined product, and, so, when you then segway in one of your earlier questions, that GDPR is, certainly, a triggering point where if it's like, okay, the penalties are huge, oh my God, it's a whole new set of regulations that I have to comply with, and when you think about that trail of breadcrumbs that I just described, that actually becomes a roadmap for compliance under regulations like GDPR, where if a European customer calls up and says, "Forget my data.", the only way that you can guarantee that you forgot that person's data, is to actually understand where it all is, and that requires proper governance, tools, and techniques, and, so, when I say governance, it's, really, not like, you know, the governor and the government, and all that. That's an aspect, but the real, important part is how do I keep all of that connectivity so that I can understand the landscape of data that I've got access to, and I'm hearing a lot of energy around that, and when you think about an IOT kind of world, distributed processing, multiple hybrid cloud footprints, data is just everywhere, and, so, the perimeter is no longer fixed, it's kind of variable, and being able to keep track of that is a very important thing for our customers. >> So, continuing on that theme, Scott. Data lakes seem to be the first major new repository we added after we had data warehouses and data marts, and it looked like the governance solutions were sort of around that perimeter of the data lake. Tell us, you were alluding to, sort of, how many more repositories, whether at rest or in motion, there are for data. Do we have to solve the governance problem end-to-end before we can build meaningful applications? >> So, I would argue personally, that governance is one of the most strategic things for us as an industry, collectively, to go solve in a universal way, and what I mean by that, is throughout my career, which is probably longer than I'd like to admit, in an EDW centric world, where things are somewhat easier in terms of the perimeter and where the data came from, data sources were much more controlled, typically ERP systems, owned wholly by a company. Even in that era, true data governance, meta data management, and that provenance was never really solved adequately. There were 300 different solutions, none of which really won. They were all different, non-compatible, and the problem was easier. In this new world, with connected data, the problem is infinitely more difficult to go solve, and, so, that same kind of approach of 300 different proprietary solutions I don't think is going to work. >> So, tell us, how does that approach have to change and who can make that change? >> So, one of the things, obviously, that we're driving is we're leveraging our position in the open community to try to use the community to create that common infrastructure, common set of APIs for meta data management, and, of course, we call that Apache Atlas, and we work with a lot of partners, some of whom are customers, some of whom are other vendors, even some of whom could be considered competitors, to try to drive an Apache open source kind of project to become that standard layer that's common into which vendors can bring their applications. So, now, if I have a common API for tracking meta data in that trail of breadcrumbs that's commonly understood, I can bring in an application that helps customers go develop the taxonomy of the rules that they want to implement, and, then, that helps visualize all of the other functionality, which is also extremely important, and that's where I think specialization comes into play, but having that common infrastructure, I think, is a really important thing, because that's going to enable data, data lakes, IOT to be trusted, and if it's not trusted, it's not going to be successful. >> Okay, there's a chicken and an egg there it sounds like, potentially. >> Am I the chicken or the egg? >> Well, you're the CTO. (Lisa laughs) >> Okay. >> The thing I was thinking of was, the broader the scope of trust that you're trying to achieve at first, the more difficult the problem, do you see customers wanting to pick off one high value application, not necessarily that's about managing what's in Atlas, in the meta data, so much as they want to do an IOT app and they'll implement some amount of governance to solve that app. In other words, which comes first? Do they have to do the end-to-end meta data management and governance, or do they pick a problem off first? >> In this case, I think it's chicken or egg. I mean, you could start from either point. I see customers who are implementing applications in the IOT space, and they're saying, "Hey, this requires a new way to think of governance, "so, I'm going to go and build that out, but I'm going to "think about it being pluggable into the next app." I also see a lot of customers, especially in highly regulated industries, and especially in highly regulated jurisdictions, who are stepping back and saying, "Forget the applications, this is a data opportunity, "and, so, I want to go solve my data fabric, "and I want to have some consistency across "that data fabric into which I can publish data "for specific applications and guarantee "that, wholistically, I am compliant "and that I'm sitting inside of our corporate mission "and all of those things." >> George: Okay. >> So, one of the things you mention, and we talk about this a lot, is the proliferation of data. It's so many, so many different sources, and companies have an opportunity, you had mentioned the phrase data opportunity, there is massive opportunity there, but you said, you know, from even a GDR perspective alone, I can't remove the data if I don't know where it is to the breadcrumbs. As a marketer, we use terms like get a 360 degree view of your customer. Is that actually really something that customers can achieve leveraging a data. Can they actually really get, say a retailer, a 360, a complete view of their customer? >> Alright, 358. >> That's pretty good! >> And we're getting there. (Lisa laughs) Yeah, I mean, obviously, the idea is to get a much broader view, and 360 is a marketing term. I'm not a marketing person, >> Yes. But it, certainly, creates a much broader view of highly personalized information that help you interact with your customer better, and, yes, we're seeing customers do that today and have great success with it and actually change and build new business models based on that capability, for sure. The folks who've done that have realized that in this new world, the way that that works is you have to have a lot of people have access to a lot of data, and that's scary, because that's not the way it used to be, right? >> Right. >> It used to be you go to the DBA and you ask for access, and then, your boss has to sign off and say it's what you asked for. In this world, you need to have access to all of it. So, when you think about this new governance capability where as part of the governance integrated with security, personalized information can be encrypted, it can be blurred out, but you still have access to the data to look at the relationships to be found in the data to build out those sophisticated models. So, that's where not only is it a new opportunity for governance just because the sources, the variety at the different landscape, but it's, ultimately, very much required, because if you're the CSO, you're not going to give access to the marketing team all of its customer data unless you understand that, right, but it has to be, "I'm just giving it to you, "and I know that it's automatically protected." versus, "I'm going to let you ask for it." to be successful. >> Right. >> I guess, following up on that, it sounds like what we were talking about, chicken or egg. Are you seeing an accelerating shift from where data is sort of collected, centrally, from applications, or, what we hear on Amazon, is the amount coming off the edge is accelerating. >> It is, and I think that that is a big drive to, frankly, faster clouded option, you know, the analytic space, particularly, has been a laggard in clouded option for many reasons, and we've talked about it previously, but one of the biggest reasons, obviously, is that data has gravity, data movement is expensive, and, so, now, when you think about where data is being created, where it lives, being further out on the edge, and may live its entire lifecycle in the cloud, you're seeing a reversal of gravity more towards cloud, and that, again, creates more opportunities in terms of driving a more varied perimeter and just keeping track of where all the assets are. Finally, I think it also leads to this notion of managing entire lifecycle of data. One of the implications of that is if data is not going to be centralized, it's going to live in different places, applications have to be portable to move to where the data exists. So, when I think about that landscape of creating ubiquitous data management within Hortonworks' portfolio, that's one of the big values that we can create for our customers. Not only can we be an on-ramp to their hybrid architecture, but as we become that on-ramp, we can also guarantee the portability of the applications that they've built out to those cloud footprints and, ultimately, even out to the edge. >> So, a quick question, then, to clarify on that, or drill down, would that mean you could see scenarios where Hortonworks is managing the distribution of models that do the inferencing on the edge, and you're collecting, bringing back the relevant data, however that's defined, to do the retraining of any models or recreation of new models. >> Absolutely, absolutely. That's one of the key things about the NiFi project in general and Hortonworks DataFlow, specifically, is the ability to selectively move data, and the selectivity can be based on analytic models as well. So, the easiest case to think about is self-driving cars. We all understand how that works, right? A self-driving car has cameras, and it's looking at things going on. It's making decisions, locally, based on models that have been delivered, and they have to be done locally, because of latency, right, but, selectively, hey, here's something that I saw as an image I didn't recognize. I need to send that up, so that it can be added to my lexicon of what images are and what action should be taken. So, of course, that's all very futuristic, but we understand how that works, but that has application in things that are very relevant today. Think about jet engines that have diagnostics running. Do I need to send that terabyte of data an hour over an expensive thing? No, but I have a model that runs locally that says, "Wow, this thing looks interesting. "Let me send a gigabyte now for immediate action." So, that decision making capability is extremely important. >> Well, Scott, thanks so much for taking some time to come chat with us once again on the Cube. We appreciate your insights. >> Appreciate it, time flies. This is great. >> Doesn't it? When you're having fun! >> Yeah. >> Alright, we want to thank you for watching the Cube. I'm Lisa Martin with George Gilbert. We are live at Forager Tasting Room in downtown San Jose at our own event, Big Data SV. We'd love for you to come on down and join us tonight, today, tonight, and tomorrow. Stick around, we'll be right back with our next guest after a short break. (techno music) >> Narrator: Since the dawn of the cloud, the Cube
SUMMARY :
Brought to you by SiliconANGLE Media Welcome back to the Cube's We are down the street from the Strata Data Conference. as I've seen in the entire show. What does that do to help customers simplify data in motion? So, the idea is to make it consumable What are some of the things... It's a little bit more from customers, and how is that helping to drive what's that I have to comply with, and when you think and it looked like the governance solutions the problem is infinitely more difficult to go solve, So, one of the things, obviously, Okay, there's a chicken and an egg there it sounds like, Well, you're the CTO. of governance to solve that app. "so, I'm going to go and build that out, but I'm going to So, one of the things you mention, is to get a much broader view, that help you interact with your customer better, in the data to build out those sophisticated models. off the edge is accelerating. if data is not going to be centralized, of models that do the inferencing on the edge, is the ability to selectively move data, to come chat with us once again on the Cube. This is great. Alright, we want to thank you for watching the Cube.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
George | PERSON | 0.99+ |
Scott | PERSON | 0.99+ |
Hortonworks | ORGANIZATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
George Gilbert | PERSON | 0.99+ |
Scott Gnau | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
San Jose | LOCATION | 0.99+ |
February | DATE | 0.99+ |
360 degree | QUANTITY | 0.99+ |
2018 | DATE | 0.99+ |
tomorrow | DATE | 0.99+ |
358 | OTHER | 0.99+ |
GDPR | TITLE | 0.99+ |
today | DATE | 0.99+ |
tomorrow morning | DATE | 0.99+ |
fifth year | QUANTITY | 0.99+ |
tonight | DATE | 0.99+ |
Lisa | PERSON | 0.99+ |
SiliconANGLE Media | ORGANIZATION | 0.99+ |
first | QUANTITY | 0.99+ |
Hortonworks' | ORGANIZATION | 0.99+ |
one department | QUANTITY | 0.99+ |
one organization | QUANTITY | 0.99+ |
two things | QUANTITY | 0.99+ |
360 | QUANTITY | 0.98+ |
one person | QUANTITY | 0.98+ |
one | QUANTITY | 0.98+ |
Cube | ORGANIZATION | 0.97+ |
Strata Data Conference | EVENT | 0.96+ |
300 different solutions | QUANTITY | 0.96+ |
an hour | QUANTITY | 0.95+ |
One | QUANTITY | 0.95+ |
tenth | QUANTITY | 0.95+ |
300 different proprietary solutions | QUANTITY | 0.95+ |
Big Data SV 2018 | EVENT | 0.93+ |
few weeks ago | DATE | 0.92+ |
one data | QUANTITY | 0.87+ |
Atlas | TITLE | 0.86+ |
Hortonworks DataFlow | ORGANIZATION | 0.85+ |
Big Data | EVENT | 0.85+ |
Cube | COMMERCIAL_ITEM | 0.84+ |
Silicon Valley | LOCATION | 0.83+ |
European | OTHER | 0.82+ |
DBA | ORGANIZATION | 0.82+ |
Apache | TITLE | 0.79+ |
Tasting | ORGANIZATION | 0.76+ |
Apache | ORGANIZATION | 0.73+ |
CTO | PERSON | 0.72+ |
Sensors | ORGANIZATION | 0.71+ |
downtown San Jose | LOCATION | 0.7+ |
Forager Tasting Room | LOCATION | 0.67+ |
SV | EVENT | 0.66+ |
terabyte of data | QUANTITY | 0.66+ |
NiFi | ORGANIZATION | 0.64+ |
Forager | LOCATION | 0.62+ |
Narrator: | TITLE | 0.6+ |
Big Data | ORGANIZATION | 0.55+ |
Room | LOCATION | 0.52+ |
Eatery | ORGANIZATION | 0.45+ |
Paul Appleby, Kinetica | Big Data SV 2018
>> Announcer: From San Jose, it's theCUBE. (upbeat music) Presenting Big Data, Silicon Valley, brought to you by Silicon Angle Media and its ecosystem partners. >> Welcome back to theCUBE. We are live on our first day of coverage of our event, Big Data SV. This is our tenth Big Data event. We've done five here in Silicon Valley. We also do them in New York City in the fall. We have a great day of coverage. We're next to where the Startup Data conference is going on at Forger Tasting Room and Eatery. Come on down, be part of our audience. We also have a great party tonight where you can network with some of our experts and analysts. And tomorrow morning, we've got a breakfast briefing. I'm Lisa Martin with my co-host, Peter Burris, and we're excited to welcome to theCUBE for the first time the CEO of Kinetica, Paul Appleby. Hey Paul, welcome. >> Hey, thanks, it's great to be here. >> We're excited to have you here, and I saw something marketer, and terms, I grasp onto them. Kinetica is the insight engine for the extreme data economy. What is the extreme data economy, and what are you guys doing to drive insight from it? >> Wow, how do I put that in a snapshot? Let me share with you my thoughts on this because the fundamental principals around data have changed. You know, in the past, our businesses are really validated around data. We reported out how our business performed. We reported to our regulators. Over time, we drove insights from our data. But today, in this kind of extreme data world, in this world of digital business, our businesses need to be powered by data. >> So what are the, let me task this on you, so one of the ways that we think about it is that data has become an asset. >> Paul: Oh yeah. >> It's become an asset. But now, the business has to care for, has to define it, care for it, feed it, continue to invest in it, find new ways of using it. Is that kind of what you're suggesting companies to think about? >> Absolutely what we're saying. I mean, if you think about what Angela Merkel said at the World Economic Forum earlier this year, that she saw data as the raw material of the 21st century. And talking about about Germany fundamentally shifting from being an engineering, manufacturing centric economy to a data centric economy. So this is not just about data powering our businesses, this is about data powering our economies. >> So let me build on that if I may because I think it gets to what, in many respects Kinetica's Core Value proposition is. And that is, is that data is a different type of an asset. Most assets are characterized by, you apply it here, or you apply it there. You can't apply it in both places at the same time. And it's one of the misnomers of the notion of data as fuels. Because fuel is still an asset that has certain specificities, you can't apply it to multiple places. >> Absolutely. >> But data, you can, which means that you can copy it, you can share it. You can combine it in interesting ways. But that means that the ... to use data as an asset, especially given the velocity and the volume that we're talking about, you need new types of technologies that are capable of sustaining the quality of that data while making it possible to share it to all the different applications. Have I got that right? And what does Kinetica do in that regard? >> You absolutely nailed it because what you talked about is a shift from predictability associated with data, to unpredictability. We actually don't know the use cases that we're going to leverage for our data moving forward, but we understand how valuable an asset it is. And I'll give you two examples of that. There's a company here, based in the Bay Area, a really cool company called Liquid Robotics. And they build these autonomous aquatic robots. And they've carried a vast array of senses and now we're collecting data. And of course, that's hugely powerful to oil and gas exploration, to research, to shipping companies, etc. etc. etc. Even homeland security applications. But what they did, they were selling the robots, and what they realized over time is that the value of their business wasn't the robots. It was the data. And that one piece of data has a totally different meaning to a shipping company than it does to a fisheries companies. But they could sell that exact same piece of data to multiple companies. Now, of course, their business has grown on in Scaldon. I think they were acquired by Bowing. But what you're talking about is exactly where Kinetica sits. It's an engine that allows you to deal with the unpredictability of data. Not only the sources of data, but the uses of data, and enables you to do that in real time. >> So Kinetica's technology was actually developed to meet some intelligence needs of the US Army. My dad was a former army ranger airborne. So tell us a little bit about that and kind of the genesis of the technology. >> Yeah, it's a fascinating use case if you think about it, where we're all concerned, globally, about cyber threat. We're all concerned about terrorist threats. But how do you identity terrorist threats in real time? And the only way to do that is to actually consume vast amount of data, whether it's drone footage, or traffic cameras. Whether it's mobile phone data or social data. but the ability to stream all of those sources of data and conduct analytics on that in real time was, really, the genesis of this business. It was a research project with the army and the NSA that was aimed at identifying terrorist threats in real time. >> But at the same time, you not only have to be able to stream all the data in and do analytics on it, you also have to have interfaces and understandable approaches to acquiring the data, because I have a background, some background in that as well, to then be able to target the threat. So you have to be able to get the data in and analyze it, but also get it out to where it needs to be so an action can be taken. >> Yeah, and there are two big issues there. One issue is the inter-offer ability of the platform and the ability for you to not only consume data in real time from multiple sources, but to push that out to a variety of platforms in real time. That's one thing. The other thing is to understand that in this world that we're talking about today, there are multiple personas that want to consume that data, and many of them are not data scientists. They're not IT people, they're business people. They could be executives, or they could be field operatives in the case of intelligence. So you need to be able to push this data out in real time onto platforms that they consume, whether it's via mobile devices or any other device for that matter. >> But you also have to be able to build applications on it, right? >> Yeah, absolutely. >> So how does Kinetica facilitate that process? Because it looks more like a database, which is, which is, it's more than that, but it satisfies some of those conventions so developers have an afinity for it. >> Absolutely, so in the first instance, we provide tools ourselves for people to consume that data and to leverage the power of that data in real time in an incredibly visual way with a geospatial platform. But we also create the ability for a, to interface with really commonly used tools, because the whole idea, if you think about providing some sort of ubiquitous access to the platform, the easiest way to do that is to provide that through tools that people are used to using, whether that's something like Tablo, for example, or Esri, if you want to talk about geospatial data. So the first instance, it's actually providing access, in real time, through platforms that people are used to using. And then, of course, by building our technology in a really, really open framework with a broadly published set of APIs, we're able to support, not only the ability for our customers to build applications on that platform, and it could well be applications associated with autonomous vehicles. It could well be applications associated with Smart City. We're doing some incredible things with some of the bigger cities on the planet and leveraging the power of big data to optimize transportation, for example, in the city of London. It's those sorts of things that we're able to do with the platform. So it's not just about a database platform or an insights engine for dealing with these complex, vast amounts of data, but also the tools that allow you to visualize and utilize that data. >> Turn that data into an action. >> Yeah, because the data is useless until you're doing something with it. And that's really, if you think about the promise of things like smart grid. Collecting all of that data from all of those smart sensors is absolutely useless until you take an action that is meaningful for a consumer or meaningful in terms of the generational consumption of power. >> So Paul, as the CEO, when you're talking to customers, we talk about chief data officer, chief information officer, chief information security officer, there's a lot, data scientist engineers, there's just so many stakeholders that need access to the data. As businesses transform, there's new business models that can come into development if, like you were saying, the data is evaluated and it's meaningful. What are the conversations that you're having, I guess I'm curious, maybe, which personas are the table (Paul laughs) when you're talking about the business values that this technology can deliver? >> Yeah, that's a really, really good question because the truth is, there are multiple personas at the table. Now, we, in the technology industry, are quite often guilty of only talking to the technology personas. But as I've traveled around the world, whether I'm meeting with the world's biggest banks, the world's biggest Telco's, the world's biggest auto manufacturers, the people we meet, more often than not, are the business leaders. And they're looking for ways to solve complex problems. How do you bring the connected card alive? How do you really bring it to life? One car traveling around the city for a full day generates a terabyte of data. So what does that really mean when we start to connect the billions of cars that are in the marketplace in the framework of connected car, and then, ultimately, in a world of autonomous vehicles? So, for us, we're trying to navigate an interesting path. We're dragging the narrative out of just a technology-based narrative speeds and feeds, algorithms, and APIs, into a narrative about, well what does it mean for the pharmaceutical industry, for example? Because when you talk to pharmaceutical executives, the holy grail for the pharma industry is, how do we bring new and compelling medicines to market faster? Because the biggest challenge for them is the cycle times to bring new drugs to market. So we're helping companies like GSK shorten the cycle times to bring drugs to market. So they're the kinds of conversations that we're having. It's really about how we're taking data to power a transformational initiative in retail banking, in retail, in Telco, in pharma, rather than a conversation about the role of technology. Now, we always needs to deal with the technologists. We need to deal with the data scientists and the IT executives, and that's an important part of the conversation. But you would have seen, in recent times, the conversation that we're trying to have is far more of a business conversation. >> So if I can build on that. So do you think, in your experience, and recognizing that you have a data management tool with some other tools that helps people use the data that gets into Kinetica, are we going to see the population of data scientists increase fast enough so our executives don't have to become familiar with this new way of thinking, or are executives going to actually adopt some of these new ways of thinking about the problem from a data risk perspective? I know which way I think. >> Paul: Wow, >> Which way do you think? >> It's a loaded question, but I think if we're going to be in a world where business is powered by data, where our strategy is driven by data, our investment decisions are driven by data, and the new areas of business that we explored to creat new paths to value are driven by data, we have to make data more accessible. And if what you need to get access to the data is a whole team of data scientists, it kind of creates a barrier. I'm not knocking data scientists, but it does create a barrier. >> It limits the aperture. >> Absolutely, because every company I talk to says, "Our biggest challenge is, we can't get access to the data scientists that we need." So a big part of our strategy from the get go was to actually build a platform with all of these personas in mind, so it is built on this standard principle, the common principles of a relational database, that you're built around anti-standard sequel. >> Peter: It's recognizable. >> And it's recognizable, and consistent with the kinds of tools that executives have been using throughout their careers. >> Last question, we've got about 30 seconds left. >> Paul: Oh, okay. >> No pressure. >> You have said Kinetica's plan is to measure the success of the business by your customers' success. >> Absolutely. >> Where are you on that? >> We've begun that journey. I won't say we're there yet. We announced three weeks ago that we created a customer success organization. We've put about 30% of the company's resources into that customer success organization, and that entire team is measured not on revenue, not on project delivered on time, but on value delivered to the customer. So we baseline where the customer is at. We agree what we're looking to achieve with each customer, and we're measuring that team entirely against the delivery of those benefits to the customer. So it's a journey. We're on that journey, but we're committed to it. >> Exciting. Well, Paul, thank you so much for stopping by theCUBE for the first time. You're now a CUBE alumni. >> Oh, thank you, I've had a lot of fun. >> And we want to thank you for watching theCUBE. I'm Lisa Martin, live in San Jose, with Peter Burris. We are at the Forger Tasting Room and Eatery. Super cool place. Come on down, hang out with us today. We've got a cocktail party tonight. Well, you're sure to learn lots of insights from our experts, and tomorrow morning. But stick around, we'll be right back with our next guest after a short break. (CUBE theme music)
SUMMARY :
brought to you by Silicon Angle Media the CEO of Kinetica, Paul Appleby. We're excited to have you here, You know, in the past, our businesses so one of the ways that we think about it But now, the business has to care for, that she saw data as the raw material of the 21st century. And it's one of the misnomers of the notion But that means that the ... is that the value of their business wasn't the robots. and kind of the genesis of the technology. but the ability to stream all of those sources of data So you have to be able to get the data in of the platform and the ability for you So how does Kinetica facilitate that process? but also the tools that allow you to visualize Yeah, because the data is useless that need access to the data. is the cycle times to bring new drugs to market. and recognizing that you have a data management tool and the new areas of business So a big part of our strategy from the get go and consistent with the kinds of tools is to measure the success of the business the delivery of those benefits to the customer. for stopping by theCUBE for the first time. We are at the Forger Tasting Room and Eatery.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Paul | PERSON | 0.99+ |
Peter Burris | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Peter | PERSON | 0.99+ |
Angela Merkel | PERSON | 0.99+ |
San Jose | LOCATION | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
Kinetica | ORGANIZATION | 0.99+ |
Paul Appleby | PERSON | 0.99+ |
Silicon Angle Media | ORGANIZATION | 0.99+ |
London | LOCATION | 0.99+ |
New York City | LOCATION | 0.99+ |
Telco | ORGANIZATION | 0.99+ |
tomorrow morning | DATE | 0.99+ |
One issue | QUANTITY | 0.99+ |
US Army | ORGANIZATION | 0.99+ |
NSA | ORGANIZATION | 0.99+ |
21st century | DATE | 0.99+ |
Liquid Robotics | ORGANIZATION | 0.99+ |
tonight | DATE | 0.99+ |
first instance | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
Bay Area | LOCATION | 0.99+ |
CUBE | ORGANIZATION | 0.99+ |
five | QUANTITY | 0.99+ |
two examples | QUANTITY | 0.99+ |
first day | QUANTITY | 0.99+ |
both places | QUANTITY | 0.99+ |
billions of cars | QUANTITY | 0.99+ |
GSK | ORGANIZATION | 0.98+ |
One car | QUANTITY | 0.98+ |
three weeks ago | DATE | 0.98+ |
each customer | QUANTITY | 0.98+ |
two big issues | QUANTITY | 0.98+ |
first time | QUANTITY | 0.97+ |
earlier this year | DATE | 0.97+ |
tenth | QUANTITY | 0.96+ |
Bowing | ORGANIZATION | 0.96+ |
Startup Data | EVENT | 0.96+ |
one | QUANTITY | 0.96+ |
Esri | TITLE | 0.95+ |
Big Data | EVENT | 0.94+ |
about 30 seconds | QUANTITY | 0.93+ |
about 30% | QUANTITY | 0.93+ |
Tablo | TITLE | 0.93+ |
World Economic Forum | EVENT | 0.92+ |
one thing | QUANTITY | 0.92+ |
theCUBE | ORGANIZATION | 0.88+ |
2018 | DATE | 0.87+ |
Big Data SV | EVENT | 0.84+ |
a terabyte | QUANTITY | 0.81+ |
one piece of data | QUANTITY | 0.77+ |
Forger Tasting Room and | ORGANIZATION | 0.73+ |
Big Data SV | ORGANIZATION | 0.72+ |
Eatery | LOCATION | 0.7+ |
Tasting | ORGANIZATION | 0.67+ |
Germany | LOCATION | 0.67+ |
data | QUANTITY | 0.65+ |
Forger | LOCATION | 0.65+ |
Room | LOCATION | 0.56+ |
CEO | PERSON | 0.55+ |
Kinetica | COMMERCIAL_ITEM | 0.45+ |
Eatery | ORGANIZATION | 0.43+ |
Scaldon | ORGANIZATION | 0.38+ |
Jacque Istok, Pivotal | Big Data SV 2018
>> Announcer: Live from San Jose, it's The Cube. Presenting Big Data, Silicon Valley. Brought to you by SiliconANGLE Media and its ecosystem partners. >> Welcome back to The Cube, we are live in San Jose at Forager Eatery, a really cool place down the street from the Strata Data Conference. This is our 10th big data event, we call this BigData SV, we've done five here, five in New York, and this is our day one of coverage, I'm Lisa Martin with my co-host George Gilbert, and we're joined by a Cube alumni, Jacque Istok, the head of data from Pivotal. Welcome back to the cube, Jacque. >> Thank you, it's great to be here. >> So, just recently you guys announced, Pivotal announced, the GA of your Kubernetes-based Pivotal container service, PKS following this initial beta that you guys released last year, tell us about that, what's the main idea behind PKS? >> So, as we were talking about earlier, we've had this opinionated platform as a service for the last couple of years, it's taken off, but it really requires a very specific methodology for deploying microservices and kind of next gen applications, and what we've seen with the ground swell behind Kubernetes is a very seamless way where we can not just do our opinionated applications, we can do any applications leveraging Kubernetes. In addition, it actually allows us to again, kind of have an opinionated way to work with stateful, stateful data, if you will. And so, what you'll see is two of the main things we have going on, again, if you look at both of those products they're all managed by a thing we call Bosch and Bosch allows for not just the ease of installation, but also the actual operation of the entire platform. And so, what we're seeing is the ability to do day two operations not just around just the apps, not just the platform, but also the data products that run within it. And you'll see later this year as we continue to evolve our data products running on top of either the PKS product or the PCF product. >> Quick question before you jump in George, so you talk about some of the technology benefits and reasoning for that, from a customer perspective, what are some of the key benefits that you've designed this for, or challenges to solve? >> I'd say the key benefits, one is convenience and ease of installation, and operationalization. Kubernetes seems to have basically become the standard for being able to deploy containers, whether its on Pram or off Pram, and having an enterprise solution to do that is something that customers are actually really looking towards, in fact, we had sold about a dozen of these products even before it was GA there was so much excitement around it. But, beyond that, I think we've been really focused on this idea of digital transformation. So Pivotal's whole talk track really is changing how companies build software. And I think the introduction of PKS really takes us to the next level, which is that there's no digital transformation without data, and basically Kubernetes and PKS allow us to implement that and perform for our customers. >> This is really a facilitator of a company's digital transformation journey. >> Correct. In a very easy and convenient way, and I think, you know, whether it's our generation, or, you know, what's going on in just technology, but everybody is so focused on convenience, push button, I just want it to work. I don't want to have to dig into the details. >> So this picks up on a theme we've been pounding on for a couple of years on our side, which is the infrastructure was too hard to stand up and operate >> Male Speaker: Yeah. >> But now that we're beginning to solve some of those problems, talk about some of the use case. Let's pick GE because that's a flagship customer, start with some of the big outcomes, some of the big business outcomes they're shooting for and then how some of the pivotal products map into that. >> Sure, so there's a lot of use cases. Obviously, GE is both a large organization, as well as an investor inside of Pivotal. A lot of different things we can talk about one that comes to mind out of the gate is we've got a data suite we sell in addition to PKS and PCF, and within that data suite there are a couple of products, green plum being one of them. Green plum is this open source MPP data platform. Probably one of the most successful implementations within GE is this ability to actually consolidate a bunch of different ERP data and have people be able to querey it, again, cheaply, easily, effectively and there are a lot of different ways you can implement a solution like that. I think what's attractive to these guys specifically around green plum is that it leverages, you know, standard ANSI SQL, it scales to pedobytes of data, we have this ability to do on pram and off pram I was actually at the Gartner Conference earlier this week and walking around the show it was actually somewhat eye opening to me to be able to see that if you look at just that one product, there really isn't a competitive product that was being showcased that was open source, multi cloud, analytical in nature, et cetera. And so I think, again, to get back to the GE scenario, what was attractive to them was everything they're doing on pram can move to the cloud, whether it's Google, Azure, Amazon they can literally run the exact same product and the exact same queries. If you extend it beyond that particular use case, there are other use cases that are more real time, and again, inside of the data suite, we've got another product called gem fire, which is an in-memory data grid that allows for this rapid ingest, so you can kind of think and imagine whether it's jet engines, or whether it's wind turbines data is constantly being generated, and our ability to take that data in real time, ingest it, actually perform analytics on it as it comes in, so, again, kind of a loose example would be if you know the heat tolerance of a wind turbine is between this temperature and this temperature, do something: send an alarm, shut it down, et cetera. If you can do that in real time, you can actually save millions of dollars by not letting that turbine fail. >> Okay, it sounds here like the gem fire product and the green plum DBMS are very complimentary. You know, one is speed, and one is sort of throughput. And we've seen almost like with Hadupen overreaction in turning a coherent platform into a bunch of building blocks. >> Male Speaker: Yes. >> And with green plum you have everything packaged together. Would it be proper to think of green plum as combining the best of the data link and the data warehouse where you've got the data scientists and data engineers with what would have been another product and the business analysts and the BI crowd satisfied with the same product, but what would have been another? >> Male Speaker: So, I'd say you're spot on. What is super interesting to me is, one, I've been doing data warehousing now for, I don't know, 20 years, and for the last five, I've kind of felt like data warehouse, just the term, was equivalent to the mainframe. So, I actually kind of relegated it the I'm not going to use that term anymore, but with the advent of the cloud and with other products that are out there we're seeing this resurgence where the data warehouse is cool again, and I think part of it is because we had this shift where we had really expensive products doing the classic EDW and it was too rigid, and it was too expensive, and Haduke sort of came on and everyone was like hey this is really easy, this is really cheap, we can store whatever we want, we can do any kind of analytics, and I think, I was saying before, the love affair with piecing all of that together is kind of over and I also think, it's funny, it was really hard for organizations to successfully stand up a Haduke platform, and I think the metric we hear is fifty percent of them fail, right, so part of that, I believe is because there just aren't enough people to be able to do what needed to be done. So, interestingly enough, because of those failures, because the Haduke ecosystem didn't quite integrate into the classic enterprise, products like green plum are suddenly very popular. I was just seeing our downloads for the open source part of green plum, and we're literally, at this juncture seeing 1500 distinct customers leveraging the open source product, so I feel like we're on kind of this upswing of getting everybody to understand that you don't have to go to Haduke to be able to do structured to unstructured data at scale. You can actually use some of these other products. >> Female Speaker: Sorry George, quickly, being in the industry for 20 years, we talk about, you know, culture a lot, and we say cultural shift. People started embracing Haduke, we can dump everything that data lake turned into swamps. I'm curious though, what is that, maybe it's not a cultural shift, maybe it's a cultural roller coaster, like, mainframes are cool again. Give us your perspective on how you've helped companies like GE sort of as technology waves come really kind of help design and maybe drive a culture that embraces the velocity of this change. >> Sure, so one of the things we do a lot is help our customers better leverage technology, and really kind of train it. So, we have a couple different aspects to pivotal. One of them is our labs aspect, and effectively that is our ability to teach people how to better build applications, how to better do data science, how to better do data engineering. Now, when we come in, we have an opinionated way to do all those things, and when a customer embraces it it actually opens up a lot of doors. So we're somewhat technology agnostic, which aids in your question, right, so we can come in, we're not trying to push a specific technology, we're trying to push a methodology and an end goal and solution. And I think, you know, often times of course that end goal and solution is best met by our products, but to your point about the roller coaster, it seems as though as we have evolved there is a notion that data will, from an organization, will all come together in a common object store, and then the ability to quickly be able to spin up an analytical or a programmmatic interface within that data is super important and that's where we're kind of leaning, and that's where I think this idea of convenience being able to push button instantiate a green plum cluster, push button instantiate a gem fire grid so that you can do analytics or you can take actions on it is so super important. >> Male Speaker: You said something that sounds really important which is we want to get it sounded like you were alluding to a single source of truth, and then you spin up whatever compute, you bring it to the data. But there's an emerging, still early school of thought which is maybe the single source of truth should be a hub centered around real time streams. >> Male Speaker: Sure. Yeah. >> How does Pivotal play in that role? >> So, there are a lot of products that can help facilitate that including our own. I would say that there is a broad ecosystem that kind of says, if I was going to start an organization today there are a number of vertical products I would need in order to be successful with data. One of the would be just a standard relational database. And if I pause there for a second, if you look at it, there is definitely a move toward building microservices so that you can glue all those pieces together. Those microservices require smaller, simpler relational type databases, or you know, SQL type databases on the front end, but they become simpler and simpler where I think if I was Oracle or some of the more stalwart on the relational side, it's not about how many widgets you can put into the database, it's really about it's simplicity and performance. From there, having some kind of message queue or system to be able to take the changes and the updates of the data down the line so that, not so much IT providing it to an end user, but more self service, being able to subscribe to the data that I care about. And again, going back to the simplicity, me as an end user being able to take control of my destiny and use whatever product or technology makes the most sense to me and if I sort of dovetail on the side of that, we've focused so much this year on convenience and flexibility that I think it is now at a spot where all of the innovations that we're doing in the Amazon marketplace on green plum, all of those innovations are actually leading us to the same types of innovations in data deployments on top of Kubernetes. And so two of them that come to mind, I felt like, I was in front of a group last week and we were presenting some of the things we had done, and one of them was self-healing of green plum and so it's often been said that these big analytical solutions are really hard to operate and through our innovations we're able to have, if a segment goes down or a host goes down, or network problems, through the implementation the system will actually self heal itself, so all of a sudden the operational needs become quite a bit less. In addition, we've also created this automatic snapshotting capability which allows, I think our last benchmark we did about a pedobyte of data in less than three minutes, so suddenly you've got this operational stalwart, almost a database as a service without really being a service really just this living breathing thing. And that kind of dovetails back to where we're trying to make all of our products perform in a way that customers can just use them and not worry about the nuts and bolts of it. >> Female Speaker: So last question, we've got about 30 seconds left. You mentioned a lot of technologies but you mentioned methodology. Is that approach from Pivotal one of the defining competitive advantages that you deliver to the market? >> Male Speaker: It is 100 per cent one of our defining our defining things. Our methodology is what is enabling our customers to be successful and it actually allows me to say we've partnered with postcrestkampf and green plum summit this year is next month in April and the theme of that is hashtag data tells the story. And so, from our standpoint, green plum is continuing to take off, gem fire is continuing to take off, Kubernetes is continuing to take off, PCF is continuing to take off, but we believe that digital transformation doesn't happen without data. We think data tells a story. I'm here to encourage everyone to come to green plum summit, I'm also here to encourage everyone to share their stories with us on twitter, hashtag data tells a story, so that we can continue to broaden this ecosystem. >> Female Speaker: Hahtag data tells a story. Jacque, thanks so much for carving out some time this week to come back to the cube and share what's new and differentiating at Pivotal. >> Thank you. >> We want to thank you for watching The Cube. I'm Lisa Martin with my co-host George Gilbert. We are live at Big Data SV, our tenth big data event come down here, see us, we're in San Jose at Forrager eatery, we've got a great party tonight and also tomorrow morning at eight am we've got a breakfast briefing you wont' want to miss. Stick around, we'll be back with our next guest after a short break.
SUMMARY :
Brought to you by SiliconANGLE Media Welcome back to The Cube, we are live in San Jose and Bosch allows for not just the ease of installation, and having an enterprise solution to do that This is really a facilitator of a company's you know, whether it's our generation, But now that we're beginning to solve and again, inside of the data suite, we've got and the green plum DBMS are very complimentary. and the business analysts and the BI crowd of getting everybody to understand a culture that embraces the velocity of this change. and then the ability to quickly be able to Male Speaker: You said something that And that kind of dovetails back to where we're competitive advantages that you deliver to the market? and it actually allows me to say and share what's new and differentiating at Pivotal. we've got a breakfast briefing you wont' want to miss.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
George Gilbert | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Bosch | ORGANIZATION | 0.99+ |
George | PERSON | 0.99+ |
two | QUANTITY | 0.99+ |
GE | ORGANIZATION | 0.99+ |
San Jose | LOCATION | 0.99+ |
fifty percent | QUANTITY | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Jacque | PERSON | 0.99+ |
New York | LOCATION | 0.99+ |
Jacque Istok | PERSON | 0.99+ |
20 years | QUANTITY | 0.99+ |
last year | DATE | 0.99+ |
Pivotal | ORGANIZATION | 0.99+ |
one | QUANTITY | 0.99+ |
100 per cent | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
last week | DATE | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
less than three minutes | QUANTITY | 0.99+ |
five | QUANTITY | 0.99+ |
The Cube | TITLE | 0.99+ |
Haduke | ORGANIZATION | 0.99+ |
next month | DATE | 0.99+ |
SiliconANGLE Media | ORGANIZATION | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
both | QUANTITY | 0.99+ |
PKS | ORGANIZATION | 0.98+ |
Cube | ORGANIZATION | 0.98+ |
Strata Data Conference | EVENT | 0.98+ |
One | QUANTITY | 0.97+ |
tonight | DATE | 0.97+ |
millions of dollars | QUANTITY | 0.97+ |
this week | DATE | 0.97+ |
tomorrow morning at | DATE | 0.97+ |
about a dozen | QUANTITY | 0.96+ |
Azure | ORGANIZATION | 0.96+ |
earlier this week | DATE | 0.96+ |
today | DATE | 0.96+ |
SQL | TITLE | 0.96+ |
1500 distinct customers | QUANTITY | 0.96+ |
this year | DATE | 0.95+ |
later this year | DATE | 0.95+ |
last couple of years | DATE | 0.95+ |
Gartner Conference | EVENT | 0.95+ |
single source | QUANTITY | 0.94+ |
Kubernetes | ORGANIZATION | 0.94+ |
one product | QUANTITY | 0.93+ |
Pram | ORGANIZATION | 0.93+ |
10th | QUANTITY | 0.93+ |
PCF | ORGANIZATION | 0.92+ |
tenth | QUANTITY | 0.92+ |
Kubernetes | TITLE | 0.88+ |
about 30 seconds | QUANTITY | 0.86+ |
April | DATE | 0.86+ |
Hahtag | ORGANIZATION | 0.85+ |
about a pedobyte of data | QUANTITY | 0.84+ |
2018 | DATE | 0.84+ |
ORGANIZATION | 0.84+ | |
One of them | QUANTITY | 0.84+ |
gem fire | ORGANIZATION | 0.83+ |
The Cube | ORGANIZATION | 0.82+ |
Murthy Mathiprakasam, Informatica | Big Data SV 2018
>> Narrator: Live from San Jose, it's theCUBE. Presenting big data silicon valley, brought to you be Siliconangle Media and its ecosystem partner. >> Welcome back to theCUBE we are live in San Jose, at Forger Eatery, super cool place. Our first day of our two days of coverage at our event called Big Data SV. Down the street is the Strata Data Conference, and we've got some great guests today that are going to share a lot insight and different perspectives on Big Data. This is our 10th big data event on theCUBE, our fifth in San Jose. We invite you to come on down to Forger Eatery and we also invite you to come down this evening. We've got a party going on and we've got a really cool breakfast presentation on the analysis site in the morning. Our first guest is, needs no introduction to theCUBE, he's a Cube Alumni, Murthy Mathiprakasam, did I get that right? >> Murthy: Absolutely. >> Murthy, awesome, as we're going to call him. The director of product marketing for Informatica, welcome back to theCUBE, it's great to have you back. >> Thanks for having me back, and congratulations on the 10 year anniversary. >> Yeah! So, interesting, exciting news from Informatica in the last two days, tell us about a couple of those big announcements that you guys just released. >> Absolutely, yes. So this has been very exciting year for us lots of, you know product, innovations and announcements, so just this week alone, actually there's one announcement that's probably going out right now as we speak, around API management, so one of the things, probably taking about before we started interviews you know around the trend toward cloud, lots of people doing a lot more data integration and application integration in the cloud space. But they face all the challenges that we've always seen in the data management space. Around developer productivity, and hand coding, just a lot of complexity that organizations have around maintenance. So one of the things at Informatica always brought to every domain that we cover is this ability to kind of abstract the underlying complexity, use a graphical user interface, make things at the logical level instead of the physical level. So we're bringing that entire kind of paradigm to the API management space. That's going to be very exciting, very game changing on the kind of app-to-app integration side of things. Back on the data world of course, which is what we're, you know, mainly talking about here today. We're doing a lot there as well. So we announced kind of a next generation of our data management platforms for the big data world, part of that is also a lot of cloud capabilities. 'Cause again, one of the bigger trends. >> Peter: Have you made a big bet there? >> Absolutely, and I mean this is the investment, return on investments over like 10 years, right? We were started in a kind of cloud game about 10 years ago with our platform as a service offering. So that has been continuously innovated on and we've been engineering, re-imagining that, to now include more of the big data stuff in it too, because more and more people are building data lakes in the cloud. So it's actually quite surprising, you know the rate at which the data lake kind of projects are now either migrating or just starting in the cloud environments. So given that being the trend, we were kind of innovating against that as well. So now our platform is service offerings supports the ability to connect to data sources in the cloud natively. You can do processing and gestion in the cloud. So there's a lot of really cool capabilities, again it's kind of bringing the Informatica ease of use, and kind of acceleration that comes to platform approach to the cloud environment. And there's a whole bunch of other announcements too, I mean I could spend 20 minutes, just on different innovations, but you know bringing artificial intelligence into the platform so we can talk more about that. >> Well I want to connect what you just announced with the whole notion of the data lake, 'cause it's really Informatica strength has always been in between. And it turns out that where a lot the enterprise problems have been. So the data lake has been there, but it's been big, it's been large, it was big data and the whole notion is make this as big as you can and we'll figure out what to do with it later. >> Murthy: Right. >> And now you're doing the API which is just a indication that we're seeing further segmentation and a specificity, a targeting of how we're going to use data, the value that we create out of data and apply it to business problems. But really Informatica strength is been in between. >> Murthy: Absolutely. >> It's been in, knowing where you data is, it's been in helping to build those pipelines and managing those pipelines. How have the investments that you've made over the last few years, made it possible for you to actually deliver an API orientation, that will actually work for Enterprises? >> Yeah, absolutely, and I would actually phrase it as sort of platform orientation, but you're exactly right. So what's happening is, I view this as sort of maturation of a lot of these new technologies. You know Hadoop was a very very, as you were saying kind of experimental technology four or five years ago. And we had customers too who were kind of in that experimental phase. But what's happening now is, big data isn't just a conversation with data engineers and developers, we're talking to CDO's, and Chief Data Officers, and VP's of data infrastructures about using Hadoop for Enterprise scale projects, now the minute you start having a conversation with a Chief Data Officer, you're not just talking about simple tools for ingestion and stuff like that. You're talking about security, you're talking about compliance, you're talking about GDPR if you're in Europe. So there's a whole host of sort of data management challenges, that are now relevant for the big data world, just because the big data world has become main stream. And so this is exactly to your point, where the investments that I think Informatica has been making and bringing our kind of comprehensive platform oriented approach to this space are paying off. Because for Chief Data Officer, they can't really do big data without those features. They can't not deal with security and compliance, they can't not deal with not knowing what the data is. 'Cause they're accountable for knowing what the data is, right? And so, there's a number of things that by virtue of the maturation of the industry, I think that trends are pointing toward, the enterprises kind of going more toward that platform approach. >> On that platform approach Informatica's really one of the only vendors that's talking about that, and delivering it. So that clearly is an area of differentiation. Why do you think that's nascent. This platform approach verses a kind of fit-for-purpose approach. >> Yeah, absolutely. And we should be careful with even the phrase fit-for-purpose too, 'cause I think that word gets thrown around a lot as it's one of those buzz words in the industry. Because it's sort of the positive way of saying incomplete, you know? And so, I think there are vendors who have tried to kind of address, know you one aspect of sort of one feature of the entire problem, that a Chief Data Officer would care about. They might call it fit-for-purpose, but you have to actually solve a problem at the end of the day. The Chief Data Officer's are trying to build enterprise data pipelines. You know you've got raw information from all sorts of data sources, on premise, in the cloud. You need to push that through a process, like at manufacturing process of being able to ingest it, repair it, cleanse it, govern it, secure it, master it, all the stuff has to happen in order to serve all the various communities that a Chief Data Officer has to serve. And so you're either doing all that or you're not. You know, that's the problem, that way we see the problem. And so the platform approach is a way of addressing the comprehensive set of problems that a Chief Data Officer, or these kind of of Data Executives care about, but also do it in a way, that fosters productivity and re-usability. Because the more you sort of build things in a kind of infrastructure level way, as soon as the infrastructure changes you're hosed, right? So you're seeing a lot of this in the industry now too, where somebody built something in Mapreduce three years ago, as soon as Spark came out, they're throwing all that stuff away. And it's not just, you know, major changes like that, even versions of Spark, or versions of Hadoop, can sometimes trigger a need to recode and throw away stuff. And organization can't afford this. When you're talking about 40 to 50% growth in the data overall. The last thing you want to do is make an investment that you're going to end up throwing away. And so, the platform approach to go back to your question, is the sort of most efficient pathway from an investment stand point, that an enterprise can take, to build something now that they can actually reuse and maintain and kind of scale in a very very pragmatic way. >> Well, let me push you on that a little bit. >> Murthy: Yeah. >> 'Cause what we would say is that, the fit-to-purpose is okay so long as you're true about the purpose, and you understand what it means to fit. What a lot of the open source, a lot of companies have done, is they've got a fit-to-purpose but then they do make promises that they say, oh this is fit-to-purpose, but it's really a platform. And as a consequence you get a whole bunch of, you know, duck-like solutions, (laughing) That are, you know, are they swimming, or are they flying, kind of problems. So, I think that what we see clients asking for, and this is one of my questions, what we see clients asking for is, I want to invest in technologies that allow me to sustain my investments, including perhaps some of my mistakes, if they are generating business value. >> Murthy: Right. >> So it's not a rip and replace, that's not what you're suggesting, what you're suggesting I think is, you know, use what you got, if it's creating value continue to use it, and then over time, invest the platform approach that's able to generate additional returns on top of it. Have I got that right? >> Absolutely. So it goes back to flexibility, that's the key word, I think that's kind of on the minds of a lot of Chief Data Officers. I don't want to build something today, that I know I'm going to throw away a year from now. >> Peter: I want to create options for the future. >> Create options. >> When I build them today. >> Exactly. So even the cloud, you're bringing up earlier on, right? Not everybody knows exactly what their cloud strategy is. And it's changing extremely rapidly, right? We had almost, we were seeing very few big data customers in the cloud maybe even a year or two ago? Now we're close to almost 50% of our big data business is people deploying off premise, I mean that's amazing, you know in a period of just a year or two. So Chief Data Officers are having to operate in these extreme kind of high velocity environments. The last thing you want to do is make a bet today, with the knowledge that you're going to end up having to throw away that bet in six months or a year. So the platform approach is sort of like your insurance policy because it enables you to design for today's requirements, but then very very quickly migrate or modify for new requirements that maybe be six months, a year or two down the line. >> On that front, I'd love for you to give us an example of a customer that has maybe in the last year, since you've seen so much velocity, come to you. But also had other technologies and their environment that from a cost perspective, I mean but at Peter's point there's still generating value, business value. How do you help customers that have multiple different products maybe exploring different multi-calibers, how to they come and start working with Informatica and not have to rip out other stuff, but be able to move forward and achieve ROI? >> So, it's really interesting kind of how people think about the whole rip and replace concept. So we actually had a customer dinner last night and I'm sitting next to a guy, and I was kind of asking very similar question. Tell me about your technology landscape, you know where are things going, where have things gone in the past, and he basically said there's a whole portfolio of technologies that they plan to obsolete. 'Cause they just know that, like they're probably, they don't even bother thinking about sustainability, to your point. They just want to use something just to kind of try it out. It's basically like a series of like three month trails of different technologies. And that's probably why we such proliferation of different technologies, 'cause people are just kind of trying stuff out, but it's like, I know I'm going to throw this stuff out. >> Yeah but that's, I mean, let me make sure I got that. 'Cause I want to reconcile a point. That's if they're in pilot and the pilot doesn't work. But the minute it goes into production and values being created they want to be able to sustain that stream of value. >> This is production environment. I'm glad you asked that question. So this is a customer that, and I'll tell you where I'm going to the point. So they've been using Informatica for over four years, for big data which is essentially almost the entire time big data's been around. So the reason this customers making the point is, Informatica's the only technology that is actually sustained precisely for the point that you're bringing up, because their requirements have changed wildly during this time. Even the internal politics of who needs access to data, all of that has changed radically over these four years. But the platform has enabled them to actually make those changes, and it's you know, been able to give them that flexibly. Everything else as far as, you know, developer tools, you know, visualization tools, like every year there's some kind of new thing that sort of comes out. And I don't want to be terribly harsh, there's probably one or two kind of vendors that have also persisted in those other areas. But, the point that they were trying to make to your original point is, is the point about sustainability. Like, at some point to avoid complete and utter chaos, you got to have like some foundation in the data environment. Something actually has to be something you can invest in today, knowing that as these changes internally externally are happening, you can kind of count on it and you can go to cloud you can be on Premise, you can have structured data, unstructured data, you know, for any type of data, any type of user, any type of deployment environment. I need something that I can count on, that's actually existing for four or more years. And that's where Informatica fits in. And meanwhile there's going to be a lot of other tools that, like this guy was saying, they're going to try out for three month or six months and that's great, but they're almost using it with the idea that they're going to throw it away. >> Couple questions here; What are some of the business values that you were, stating like this gentlemen, that you ere talking to last night. What's the industry that's he in and also, are there any like stats or ranges you can give us. Like, reduction in TCO, or new business models opening up. What's the business impact that Informatica is helping these customers achieve. >> Yeah, absolutely, I'll use this example, he's, I can't mention the name of the company but it's an insurance company. >> Lisa: Lot's of data. >> Lots of data, right. Not only do they have a lot of data, but there's a lot of sensitivity around the data. Because basically the only way they grow is by identifying patterns in consumers and they want to look at it if somebody's using car insurance in, maybe it for so long they're ready to get married, they need home insurance, they have these like really really sophisticated models around human behavior. So they know when to go and position new forms of insurance. There's also obviously security government types of issues that are at play as well. So the sensitivity around data is very very important. So for them, the business value is increased revenue, and you know ability to meet kind of regulatory pressure. I think that's generally, I mean every industry has some variant of that. >> Right. >> Cost production, increase revenue, you know meeting regulatory pressures. And so Informatica facilitates that, because instead of having to hire armies of people, and having to change them out maybe every three months or six months 'cause the underlying infrastructures changing, there's this one team, the Informatica team that's actually existed for this entire journey. They just keep changing, used cases, and projects, and new data sets, new deployment models, but the platform is sort of fixed and it's something that they can count on it's robust, it enables that kind of. >> Peter: It's an asset. >> It's an asset that delivers that sustainable value that you were taking about. >> Last question, we've got about a minute left, in terms of delivering value, Informatica not the only game in town, your competitors are kind of going with this MNA partnership approach. What makes Informatica stand out, why should companies consider Informatica? >> So they say like, what there's a quote about it. Imitation is the most sincere from of flattery. Yeah! (laughing) I guess we should feel as a little bit flattered, you know, by what we're seeing in the industry, but why from a customers stand point should they, you know continue to rely on Informatica. I mean we keep pushing the envelope on innovations, right? So, one the other areas that we innovated on is machine learning within the platform, because ultimately if one of the goals of the platform is to eliminate manual labor, a great way to do that is to just not have people doing it in the first place. Have machines doing it. So we can automatically understand the structure of data without any human intervention, right? We can understand if there's a file and it's got costumer names and you know, cost and skews, it must be an order. You don't actually have to say that it's an order. We can infer all this, because of the machine learning them we have. We can give recommendations to people as they're using our platform, if you're using a data set and you work with another person, we can go to you and say hey, maybe this is a data set that you would be interesting in. So those types of recommendations, predictions, discovery, totally changes the economic game for an organization. 'Cause the last thing you want is to have 40 to 50% growth in data translate into 40 to 50% of labor. Like you just can't afford it. It's not sustainable, again, to go back to your original point. The only sustainable approach to managing data for the future, is to have a machine learning based approach and so that's why, to your question, I think just gluing a bunch of stuff together still doesn't actually get to nut of sustainability. You actually have to have, the glue has to have something in it, you know? And in our case it's the machine learning approach that ties everything together that brings a data organization together, so they can actually deliver the maximum business value. >> Literally creates a network of data that delivers business value. >> You got it. >> Well Murthy, Murthy Awesome, thank you so much for coming back to theCUBE. >> Thank you! >> And sharing what's going on the Informatica and what's differentiating you guys. We wish you a great rest of the Strata Conference. >> Awesome, you as well. Thank you. >> Absolutely, we want to thank you for watching theCUBE. I'm Lisa Martin with Peter Burris, we are live in San Jose at the Forger Eatery, come down here and join us, we've got a really cool space, we've got a part-tay tonight, so come join us. And we've got a really interesting breakfast presentation tomorrow morning, stick around and we'll be right back, with our next guest for this short break. (fun upbeat music)
SUMMARY :
brought to you be Siliconangle Media and we also invite you to come down this evening. welcome back to theCUBE, it's great to have you back. and congratulations on the 10 year anniversary. big announcements that you guys just released. of our data management platforms for the big data world, and kind of acceleration that comes to platform approach So the data lake has been there, and apply it to business problems. for you to actually deliver an API orientation, now the minute you start having a conversation Informatica's really one of the only vendors And so, the platform approach to go back to your question, about the purpose, and you understand what it means to fit. you know, use what you got, that I know I'm going to throw away a year from now. So even the cloud, you're bringing up earlier on, right? that has maybe in the last year, of technologies that they plan to obsolete. But the minute it goes into production But the platform has enabled them to actually make What are some of the business values that you were, he's, I can't mention the name of the company and you know ability to meet kind of regulatory pressure. and it's something that they can count on it's robust, that you were taking about. Informatica not the only game in town, the glue has to have something in it, you know? that delivers business value. thank you so much for coming back to theCUBE. and what's differentiating you guys. Awesome, you as well. Absolutely, we want to thank you for watching theCUBE.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
40 | QUANTITY | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Peter Burris | PERSON | 0.99+ |
Europe | LOCATION | 0.99+ |
Lisa | PERSON | 0.99+ |
Peter | PERSON | 0.99+ |
six months | QUANTITY | 0.99+ |
San Jose | LOCATION | 0.99+ |
Murthy Mathiprakasam | PERSON | 0.99+ |
three month | QUANTITY | 0.99+ |
Informatica | ORGANIZATION | 0.99+ |
Murthy | PERSON | 0.99+ |
20 minutes | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
two days | QUANTITY | 0.99+ |
four | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
first day | QUANTITY | 0.99+ |
a year | QUANTITY | 0.99+ |
tomorrow morning | DATE | 0.99+ |
Siliconangle Media | ORGANIZATION | 0.99+ |
four | DATE | 0.99+ |
one team | QUANTITY | 0.99+ |
MNA | ORGANIZATION | 0.99+ |
last year | DATE | 0.99+ |
last night | DATE | 0.99+ |
today | DATE | 0.99+ |
fifth | QUANTITY | 0.99+ |
50% | QUANTITY | 0.98+ |
first guest | QUANTITY | 0.98+ |
four years | QUANTITY | 0.98+ |
tonight | DATE | 0.98+ |
Cube | ORGANIZATION | 0.98+ |
three years ago | DATE | 0.98+ |
Big Data SV | EVENT | 0.97+ |
five years ago | DATE | 0.97+ |
first | QUANTITY | 0.96+ |
Couple questions | QUANTITY | 0.96+ |
Strata Data Conference | EVENT | 0.95+ |
almost 50% | QUANTITY | 0.95+ |
10 year anniversary | QUANTITY | 0.94+ |
over four years | QUANTITY | 0.94+ |
two kind | QUANTITY | 0.94+ |
one announcement | QUANTITY | 0.94+ |
GDPR | TITLE | 0.94+ |
two ago | DATE | 0.92+ |
Action Item | Big Data SV Preview Show - Feb 2018
>> Hi, I'm Peter Burris and once again, welcome to a Wikibon Action Item. (lively electronic music) We are again broadcasting from the beautiful theCUBE Studios here in Palo Alto, California, and we're joined today by a relatively larger group. So, let me take everybody through who's here in the studio with us. David Floyer, George Gilbert, once again, we've been joined by John Furrier, who's one of the key CUBE hosts, and on the remote system is Jim Kobielus, Neil Raden, and another CUBE host, Dave Vellante. Hey guys. >> Hi there. >> Good to be here. >> Hey. >> So, one of the things we're, one of the reasons why we have a little bit larger group here is because we're going to be talking about a community gathering that's taking place in the big data universe in a couple of weeks. Large numbers of big data professionals are going to be descending upon Strata for the purposes of better understanding what's going on within the big data universe. Now we have run a CUBE show next to that event, in which we get the best thought leaders that are possible at Strata, bring them in onto theCUBE, and really to help separate the signal from the noise that Strata has historically represented. We want to use this show to preview what we think that signal's going to be, so that we can help the community better understand what to look for, where to go, what kinds of things to be talking about with each other so that it can get more out of that important event. Now, George, with that in mind, what are kind of the top level thing? If it was one thing that we'd identify as something that was different two years ago or a year ago, and it's going to be different from this show, what would we say it would be? >> Well, I think the big realization that's here is that we're starting with the end in mind. We know the modern operational analytic applications that we want to build, that anticipate or influence a user interaction or inform or automate a business transaction. And for several years, we were experimenting with big data infrastructure, but that was, it wasn't solution-centric, it was technology-centric. And we kind of realized that the do it yourself, assemble your own kit, opensource big data infrastructure created too big a burden on admins. Now we're at the point where we're beginning to see a more converged set of offerings take place. And by converged, I mean an end to end analytic pipeline that is uniform for developers, uniform for admins, and because it's pre-integrated, is lower latency. It helps you put more data through one single analytic latency budget. That's what we think people should look for. Right now, though, the hottest new tech-centric activity is around Machine Learning, and I think the big thing we have to do is recognize that we're sort of at the same maturity level as we were with big data several years ago. And people should, if they're going to work with it, start with the knowledge, for the most part, that they're going to be experimenting, 'cause the tooling isn't quite mature enough, we don't have enough data scientists for people to be building all these pipelines bespoke. And the third-party applications, we don't have a high volume of them where this is embedded yet. >> So if I can kind of summarize what you're saying, we're seeing bifurcation occur within the ecosystem associated with big data that's driving toward simplification on the infrastructure side, which increasingly is being associated with the term big data, and new technologies that can apply that infrastructure and that data to new applications, including things like AI, ML, DL, where we think about modeling and services, and a new way of building value. Now that suggests that one or the other is more or less hot, but Neil Raden, I think the practical reality is that here in Silicon Valley, we got to be careful about getting too far out in front of our skis. At the end of the day, there's still a lot of work to be done inside how you simply do things like move data from one place to the other in a lot of big enterprises. Would you agree with that? >> Oh absolutely. I've been talking to a lot clients this week and, you know, we don't talk about the fact that they're still running their business on what we would call legacy systems, and they don't know how to, you know, get out of them or transform from them. So they're still starting to plan for this, but the problem is, you know, it's like talking about the 27 rocket engines on the whatever it was that he launched into space, launching a Tesla into space. But you can talk about the engineering of those engines and that's great, but what about all the other things you're going to have to do to get that (laughs) car into space? And it's the same thing. A year ago, we were talking about Hadoop and big data and, to a certain extent, Machine Learning, maybe more data science. But now people are really starting to say, How do we actually do this, how do we secure it, how do we govern it, how do we get some sort of metadata or semantics on the data we're working with so people know what they're using. I think that's where we are in a lot of companies. >> Great, so that's great feedback, Neil. So as we look forward, Jim Kobielus, the challenges associated with what it means to better improve the facilities of your infrastructure, but also use that as a basis for increasing your capability on some of the new applications services, what are we looking for, what should folks be looking for as they explore the show in the next couple of weeks on the ML side? What new technologies, what new approaches? Going back to what George said, we're in experimentation mode. What are going to be the experiments that are going to generate greatest results over the course of the next year? >> Yeah, for the data scientists, who flock to Strata and similar conferences, automation of the Machine Learning pipeline is super hot in terms of investments by the solution providers. Everybody from Google to IBM to AWS, and others, are investing very heavily in automation of, not just the data engine, that problem's been had a long time ago. It's automation of more of the feature engineering and the trending. These very manual, often labor intensive, jobs have to be sped up and automated to a great degree to enable the magic of productivity by the data scientists in the new generation of app developers. So look for automation of Machine Learning to be a super hot focus. Related to that is, look for a new generation of development suites that focus on DevOps, speeding the Machine Learning in DL and AI from modeling through training and evaluation deployment in iteration. We've seen a fair upswing in the number of such toolkits on the market from a variety of startup vendors, like the DataRobots of the world. But also coming to say, AWS with SageMaker, for example, that's hot. Also, look for development toolkits that automate more of the cogeneration, you know, a low-code tools, but the new generation of low-code tools, as highlighted in a recent Wikibons study, use ML to drive more of the actual production of fairly decent, good enough code, as a first rough prototype for a broad range of applications. And finally we're seeing a fair amount of ML-generated code generation inside of things like robotic process automation, RPA, which I believe will probably be a super hot theme at Strata and other shows this year going forward. So there's a, you mentioned the idea of better tooling for DevOps and the relationship between big data and ML, and what not, and DevOps. One of the key things that we've been seeing over the course of the last few years, and it's consistent with the trends that we're talking about, is increasing specialization in a lot of the perspectives associated with changes within this marketplace, so we've seen other shows that have emerged that have been very, very important, that we, for example, are participating in. Places like Splunk, for example, that is the vanguard, in many respects, of a lot of these trends in big data and how big data can applied to business problems. Dave Vellante, I know you've been associated with a number of, participating in these shows, how does this notion of specialization inform what's going to happen in San Jose, and what kind of advice and counsel should we tell people to continue to explore beyond just what's going to happen in San Jose in a couple weeks? >> Well, you mentioned Splunk as an example, a very sort of narrow and specialized company that solves a particular problem and has a very enthusiastic ecosystem and customer base around that problem. LAN files to solve security problems, for example. I would say Tableau is another example, you know, heavily focused on Viz. So what you're seeing is these specialized skillsets that go deep within a particular domain. I think the thing to think about, especially when we're in San Jose next week, is as we talk about digital disruption, what are the skillsets required beyond just the domain expertise. So you're sort of seeing this bifurcated skillsets really coming into vogue, where if somebody understands, for example, traditional marketing, but they also need to understand digital marketing in great depth, and the skills that go around it, so there's sort of a two-tool player. We talk about five-tool player in baseball. At least a multidimensional skillset in digital. >> And that's likely to occur not just in a place like marketing, but across the board. David Floyer, as folks go to the show and start to look more specifically about this notion of convergence, are there particular things that they should think about that, to come back to the notion of, well, you know, hardware is going to make things more or less difficult for what the software can do, and software is going to be created that will fill up the capabilities of hardware. What are some of the underlying hardware realities that folks going to the show need to keep in mind as they evaluate, especially the infrastructure side, these different infrastructure technologies that are getting more specialized? >> Well, if we look historically at the big data area, the solution has been to put in very low cost equipment as nodes, lots of different nodes, and move the data to those nodes so that you get a parallelization of the, of the data handling. That is not the only way of doing it. There are good ways now where you can, in fact, have a single version of that data in one place in very high speed storage, on flash storage, for example, and where you can allow very fast communication from all of the nodes directly to that data. And that makes things a lot simpler from an operational point of view. So using current Batch Automation techniques that are in existence, and looking at those from a new perspective, which is I do IUs apply these to big data, how do I automate these things, can make a huge difference in just the practicality in the elapsed time for some of these large training things, for example. >> Yeah, I was going to say that to many respects, what you're talking about is bringing things like training under a more traditional >> David: Operational, yeah. >> approach and operational set of disciplines. >> David: Yes, that's right. >> Very, very important. So John Furrier, I want to come back to you, or I want to come to you, and say that there are some other technologies that, while they're the bright shiny objects and people think that they're going to be the new kind of Harry Potter technologies of magic everywhere, Blockchain is certainly going to become folded into this big data concept, because Blockchain describes how contracts, ownership, authority ultimately get distributed. What should folks look for as the, as Blockchain starts to become part of these conversations? >> That's a good point, Peter. My summary of the preview for BigData SV Silicon Valley, which includes the Strata show, is two things: Blockchain points to the future and GDPR points to the present. GDPR is probably the most, one of the most fundamental impacts to the big data market in a long time. People have been working on it for a year. It is a nightmare. The technical underpinnings of what companies have to do to comply with GDPR is a moving train, and it's complete BS. There's no real solutions out there, so if I was going to tell everyone to think about that and what to look for: What is happening with GDPR, what's the impact of the databases, what's the impact of the architectures? Everyone is faking it 'til they make it. No one really has anything, in my opinion from what I can see, so it's a technical nightmare. Where was that database? So it's going to impact how you store the data, and the sovereignty issue is another issue. So the Blockchain then points to the sovereignty issue of the data, both in terms of the company, the country, and the user. These things are going to impact software development, application development, and, ultimately, cloud choice and the IoT. So to me, GDPR is not just a one and done thing and Blockchain is kind of a future thing to look at. So I would look out of those two lenses and say, Do you have a direction or a narrative that supports me today with what GDPR will impact throughout the organization. And then, what's going on with this new decentralized infrastructure and the role of data, and the sovereignty of that data, with respect to company, country, and user. So to me, that's the big issue. >> So George Gilbert, if we think about this question of these fundamental technologies that are going to become increasingly important here, database managers are not dead as a technology. We've seen a relative explosion over the last few years in at least invention, even if it hasn't been followed with, as Neil talked about, very practical ways of bringing new types of disciplines into a lot of enterprises. What's going to happen with the database world, and what should people be looking for in a couple of weeks to better understand how some of these data management technologies are going to converge and, or involve? >> It's a topic that will be of intense interest and relevance to IT professionals, because it's become the common foundation of all modern apps. But I think what we can do is we can see, for instance, a leading indicator of what's going to happen with the legacy vendors, where we have in-memory technologies from both transaction processing and analytics, and we have more advanced analytics embedded in the database engine, including Machine Learning, the model training, as well as model serving. But the, what happened in the big data community is that we disassembled the DBMS into the data manipulation language, which is an analytic language, like, could be Spark, could be Flink, even Hive. We had the Catalog, which I think Jim has talked about or will be talking about, where we're not looking, it's not just a dictionary of what's in one DBMS, but it's a whole way of tracking and governing data across many stores. And then there's the Storage Manager, could be the file system, an object store, could be just something like Kudu, which is a MPP way of, in parallel, performing a bunch of operations on data that's stored. The reason I bring all this up is, following on David's comment about the evolution of hardware, databases are fundamentally meant to expose capabilities in the hardware and to mediate access to data, using these hardware capabilities. And now that we have this, what's emerging as this unigrid, with memory-intensive architectures and super low latency to get from any point or node on that cluster to any other node, like with only a five microsecond lag, relative to previous architectures. We can now build databases that scale up with the same knowledge base that we built databases... I'm sorry, that scale out, that we used to build databases that scale up. In other words, it democratizes the ability to build databases of enormous scale, and that means that we can have analytics and the transactions working together at very low latency. >> Without binding them. Alright, so I think it's time for the action items. We got a lot to do, so guys, keep it really tight, really simple. David Floyer, let me start with you. Action item. >> So action item on big data should be focus on technologies that are going to reduce the elapse time of solutions in the data center, and those are many and many of them, but it's a production problem, it's becoming a production problem, treat it as a production problem, and put it in the fundamental procedures and technologies to succeed. >> And look for vendors >> Who can do that, yes. >> that do that. George Gilbert, action item. >> So I talked about convergence before. The converged platform now is shifting, it's center of gravity is shifting to continuous processing, where the data lake is a reference data repository that helps inform the creation of models, but then you run the models against the streaming continuous data for the freshest insights-- >> Okay, Jim Kobielus, action item. >> Yeah, focus on developer productivity in this new era of big data analytics. Specifically focus on the next generation of developers, who are data scientists, and specifically focus on automating most of what they do, so they can focus on solving problems and sifting through data. Put all the grunt work or training, and all that stuff, take and carry it by the infrastructure, the tooling. >> Peter: Neil Raden, action item. >> Well, one thing I learned this week is that everything we're talking about is about the analytical problem, which is how do you make better decisions and take action? But companies still run on transactions, and it seems like we're running on two different tracks and no one's talking about the transactions anymore. We're like the tail wagging the dog. >> Okay, John Furrier, action item. >> Action item is dig into GDPR. It is a really big issue. If you're not proactive, it could be a nightmare. It's going to have implications that are going to be far-reaching in the technical infrastructure, and it's the Sarbanes-Oxley, what they did for public companies, this is going to be a nightmare. And evaluate the impact of Blockchains. Two things. >> David Vellante, action item. >> So we often say that digital is data, and just because your industry hasn't been upended by digital transformations, don't think it's not coming. So it's maybe comfortable to sit back and say, Well, we're going to wait and see. Don't sit back and wait and see. All industries are susceptible to digital transformation. >> Alright, so I'll give the action item for the team. We've talked a lot about what to look for in the community gathering that's taking place next week in Silicon Valley around strata. Our observations as the community, it descends upon us, and what to look for is, number one, we're seeing a bifurcation in the marketplace, in the thought leadership, and in the tooling. One set of group, one group is going more after the infrastructure, where it's focused more on simplification, convergence; another group is going more after the developer, AI, ML, where it's focused more on how to create models, training those models, and building applications with the services associated with those models. Look for that. Don't, you know, be careful about vendors who say that they do it all. Be careful about vendors that say that they don't have to participate in a converged approach to doing this. The second thing I think we need to look for, very importantly, is that the role of data is evolving, and data is becoming an asset. And the tooling for driving velocity of data through systems and applications is going to become increasingly important, and the discipline that is necessary to ensure that the business can successfully do that with a high degree of predictability, bringing new production systems are also very important. A third area that we take a look at is that, ultimately, the impact of this notion of data as an asset is going to really come home to roost in 2018 through things like GDPR. As you scan the show, ask a simple question: Who here is going to help me get up to compliance and sustain compliance, as the understanding of privacy, ownership, etc. of data, in a big data context, starts to evolve, because there's going to be a lot of specialization over the next few years. And there's a final one that we might add: When you go to the show, do not just focus on your favorite brands. There's a lot of new technology out there, including things like Blockchain. They're going to have an enormous impact, ultimately, on how this marketplace unfolds. The kind of miasma that's occurred in big data is starting to specialize, it's starting to break down, and that's creating new niches and new opportunities for new sources of technology, while at the same time, reducing the focus that we currently have on things like Hadoop as a centerpiece. A lot of convergence is going to create a lot of new niches, and that's going to require new partnerships, new practices, new business models. Once again, guys, I want to thank you very much for joining me on Action Item today. This is Peter Burris from our beautiful Palo Alto theCUBE Studio. This has been Action Item. (lively electronic music)
SUMMARY :
We are again broadcasting from the beautiful and it's going to be different from this show, And the third-party applications, we don't have Now that suggests that one or the other is more or less hot, but the problem is, you know, it's like talking about the What are going to be the experiments that are going to in a lot of the perspectives associated with I think the thing to think about, that folks going to the show need to keep in mind and move the data to those nodes and people think that they're going to be So the Blockchain then points to the sovereignty issue What's going to happen with the database world, in the hardware and to mediate access to data, We got a lot to do, so guys, focus on technologies that are going to that do that. that helps inform the creation of models, Specifically focus on the next generation of developers, and no one's talking about the transactions anymore. and it's the Sarbanes-Oxley, So it's maybe comfortable to sit back and say, and sustain compliance, as the understanding of privacy,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
David | PERSON | 0.99+ |
Jim Kobielus | PERSON | 0.99+ |
George | PERSON | 0.99+ |
David Floyer | PERSON | 0.99+ |
George Gilbert | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Neil Raden | PERSON | 0.99+ |
Neil | PERSON | 0.99+ |
Peter Burris | PERSON | 0.99+ |
David Vellante | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
San Jose | LOCATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
Peter | PERSON | 0.99+ |
Feb 2018 | DATE | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
Jim | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
2018 | DATE | 0.99+ |
ORGANIZATION | 0.99+ | |
GDPR | TITLE | 0.99+ |
next week | DATE | 0.99+ |
two things | QUANTITY | 0.99+ |
Palo Alto, California | LOCATION | 0.99+ |
Splunk | ORGANIZATION | 0.99+ |
both | QUANTITY | 0.99+ |
A year ago | DATE | 0.99+ |
two lenses | QUANTITY | 0.99+ |
a year ago | DATE | 0.99+ |
two years ago | DATE | 0.99+ |
this week | DATE | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
first | QUANTITY | 0.99+ |
third area | QUANTITY | 0.98+ |
CUBE | ORGANIZATION | 0.98+ |
one group | QUANTITY | 0.98+ |
second thing | QUANTITY | 0.98+ |
27 rocket | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
next year | DATE | 0.98+ |
Two things | QUANTITY | 0.97+ |
theCUBE Studios | ORGANIZATION | 0.97+ |
two-tool player | QUANTITY | 0.97+ |
five microsecond | QUANTITY | 0.96+ |
One set | QUANTITY | 0.96+ |
Tableau | ORGANIZATION | 0.94+ |
a year | QUANTITY | 0.94+ |
single version | QUANTITY | 0.94+ |
one | QUANTITY | 0.94+ |
Wikibons | ORGANIZATION | 0.91+ |
Wikibon | ORGANIZATION | 0.91+ |
two different tracks | QUANTITY | 0.91+ |
five-tool player | QUANTITY | 0.9+ |
several years ago | DATE | 0.9+ |
this year | DATE | 0.9+ |
Strata | TITLE | 0.87+ |
Harry Potter | PERSON | 0.85+ |
one thing | QUANTITY | 0.84+ |
years | DATE | 0.83+ |
one place | QUANTITY | 0.82+ |
Action Item Quick Take | John Furrier - Feb 2018 (Segment 2)
>> Hi, I'm Peter Burris, and welcome to Wikibon Action Item Quick Take. Big Data SV is one of our important shows where we bring thought leadership around big data to the Cube and have great conversations about what's happening in the Big Data universe. John Furrier, what are we looking for in the next couple of weeks? >> Big Data Silicon Valley known as Big Data SV as we have NYC for New York City, two events that we co-produce in conjunction with Strata Conference going on side-by-side where we do the following. We have three days: Tuesday, Wednesday, and Thursday, the sixth, seventh, and eighth. We're going to be in San Jose. And we have a great lineup. And it's pretty much sold out, but we added Thursday for more live interviews, where we extract the signal from the noise. So we have more opportunities to interview more people, and also, we're opening up more sponsorship slots. So if you want to get your company's name out there, get above the noise, and get those thought leadership interviews out, we have just released extra sponsorship opportunities for Thursday for live interviews and conversations on the Cube, a new format. As you know, it's proven as a conversational great way to get the word out in an informative, inspirational way. Of course, that's the Cube mission, Peter, as you know. And we love doing what we do. We love the support of our sponsors. So if you want to be a sponsor and have that conversation with us, we'd love to entertain that opportunity on Thursday, March 8th. >> Alright, John Furrier, your Cube host, who actually is going to be hosting Big Data SV. This has been a Wikibon Action Item Quick Take.
SUMMARY :
and have great conversations So we have more opportunities to interview more people, Alright, John Furrier, your Cube host,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Peter Burris | PERSON | 0.99+ |
Feb 2018 | DATE | 0.99+ |
San Jose | LOCATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
Thursday | DATE | 0.99+ |
Peter | PERSON | 0.99+ |
Thursday, March 8th | DATE | 0.99+ |
three days | QUANTITY | 0.99+ |
two events | QUANTITY | 0.99+ |
Tuesday | DATE | 0.99+ |
New York City | LOCATION | 0.99+ |
Wednesday | DATE | 0.99+ |
one | QUANTITY | 0.99+ |
NYC | LOCATION | 0.99+ |
Cube | ORGANIZATION | 0.97+ |
Wikibon | ORGANIZATION | 0.95+ |
eighth | QUANTITY | 0.94+ |
Silicon Valley | LOCATION | 0.91+ |
Strata Conference | EVENT | 0.88+ |
seventh | QUANTITY | 0.87+ |
sixth | QUANTITY | 0.78+ |
next couple of weeks | DATE | 0.71+ |
Big Data SV | EVENT | 0.71+ |
Big Data | ORGANIZATION | 0.56+ |
Segment 2 | OTHER | 0.54+ |
Cube | COMMERCIAL_ITEM | 0.33+ |
Abby Kearns | IBM Interconnect 2017
(bouncy electronic music) [Narrator] Live from Las Vegas, it's the CUBE. Covering InterConnect 2017 brought to you by IBM. >> Hey welcome back everyone, we are live in Las Vegas for IBM InterConnect 2017. This is the CUBE's coverage of IBM's Cloud and data show. I'm John Furrier, with my co-host Dave Vellante. Our next guest is Abby Kearns, Executive Director of Cloud Foundry Foundation. Welcome to the CUBE! >> Welcome, thank you! >> Thanks for joining us, so Cloud Foundry, you're new as the executive role, Sam had moved on to Microsoft? >> Abby: Google. >> Google, I'm sorry, Google, he was formerly at Microsoft, former Microsoft employee, but Google, Google Cloud Next was a recent show. >> Yeah. >> So, you're new. >> I'm new. >> John: To the reins but you're not new, new to the community. >> I've been a part of the community for several years prior to joining the foundation a year ago I was at Pivotal for a couple of years so I've been part of the Cloud Foundry community for several years and it's a technology that's near and dear to my heart and it's a community that I am very passionate about. >> And the emergence of Cloud Foundry if you think about it has really kind of changed the game it's really lifted all the boats, if you will, rising tide floats all boats. IBM uses it, you've got a lot of customers. Just go down the list of the notable folks working with Cloud Foundry. >> Well, look no further than those that are on our board and those that represent the strategic vision around the Cloud Foundry, so IBM, Pivotal, but, Dell EMC, and Cisco and SAP and VMware and Allianz and Swisscom. And of course, Pivotal. I think all of them really bring such a broad perspective to the table. But then broadening beyond that community, our community has grown so much. A lot of people don't realize that Cloud Foundry has only been an open-source project for just a little over two years, so January 2015 marked when it became an official open-source project. Prior to that it was part of Pivotal. And in that little-over-two years, we've grown to nearly 70 members in our community and are just excited to continue to grow and bring more perspectives to the table. >> So what has been the differences, a lot of people have been taking a different approach on for Bluemix, for instance, they have a good core at Cloud Foundry. Is it going the way you guys had thought as a community, that this was the plan all along? Because you see people really kind of making some good stuff out of the Cloud Foundry. Was that part of the plan, this open direction? >> Well I think part of the plan was really coalescing around the single vision of that abstraction And what's the whole vision of Cloud Foundry, it's to allow developers to create code faster. And whatever realm that takes. Our industry is evolving and it's evolving so quickly and exciting, all of these enterprise organizations that are becoming software companies. I mean how exciting is that? As we think about the abstraction that Cloud Foundry can provide for them and the automation it can provide, it allows them to focus on one thing and one thing only, creating code that changes their business. We're really focused myopically on ensuring that developers have the ability to quickly and easily create code and innovate quickly as an organization. >> So on the development side, sometimes standards can go fall down by forcing syntax or forcing certain things. You guys had a different approach, looking back now, what were the key things that were critical for Cloud Foundry to maintain its momentum? >> I think a couple of things. It's a complex distributed system but it is put together amazingly well. Quality was first and foremost, part of its origins. And it's continued to adhere to that quality and that control around the development process and around the release process. So Cloud Foundry as an open-source project is very much a governance by contribution. So we look for those in the organizations and different communities to be part of it and contribute. So we have the full-time committers that are basically doing this all day, every day, and then we have the contributors that are also part of the community providing feedback and value. >> And there was a big testimonial with American Airlines on stage, that's a big win. >> Abby: Yes, it is a big win. >> Give us some color on that deal. >> I can't give you any details on the deal that IBM has-- >> But that's a Cloud Foundry, IBM-- >> But it is Cloud Foundry, yes. >> You guys were part of the Bluemix thing? >> Yes. And American Airlines is a company that I have a lot of history with, They were a customer of mine for many years in the early 2000s, so I'm thrilled to see them innovating and taking advantage of a platform. >> So, help us unpack this conversation that's going on around PaaS, right? >> Some people say, "oh, PaaS is pase," but it's development tools and it's programming and it's a platform that you've created, so what do you make of that conversation? What implications does it have to your strategy and your ecosystem strategy? >> Well, I for one don't like the term PaaS anyway, so I'm happy to say PaaS is pase. Because I do think it's evolved, so when I talk about Cloud Foundry, I talk about it as a Cloud application platform. Because at the end of the day, our goal is to help organizations create code faster. The high degrees of automation, the abstraction that the platform brings to the table, it isn't just a platform, it is an enabler for that development. So we think about what that means, it's, can I create applications faster and do I have a proliferation of services to your ecosystem point that enable applications to grow and to scale and to change the way that organization works. Because it's a technology-enabled business transformation for many of these organizations. >> John: It's app-driven, too, that's the key to success. >> It's app-driven, which is why we talk so much about developers, is because that's the key, if I'm going to become a software company, what does that mean? I am writing code, and that code is changing the way I think about my business and my consumers. >> And the app landscape has certainly changed with UX creativity, but now you've got IoT, there's a real functional integration going on with the analog world going digital, it's like, "Whoa, "I've got all this stuff that's now instrumented "connected to the internet!" IoT, Internet of Things. That's going to be interesting, Cloud has to power that. >> I think it does, because what is IoT reliant on? Applications that take advantage of that data. That's what you're looking to gain, you're looking to have small applications streaming large amounts of data from sensors, be it from cars, or be it from a manufacturing plant, if you're thinking industrial IoT, so Cloud Foundry provides the platform for many of these applications to be developed, created, and scaled at the level that companies like GE, and Siemens, and others are looking to build out and tackle that IoT space. >> It's open, I mean we can all agree that Cloud Foundry's the most open platform to develop applications on, but developers have choices. You're seeing infrastructure as a service, plus you're seeing SAS kind of minus emerge. How should we be thinking about the evolution, you said earlier it evolved, where is it evolving to? Obviously you bet on open, good bet. Other more propriet... I don't even know what open is anymore sometimes (Abby laughs) >> But we can agree that Cloud Foundry's open. But how should we be thinking about the evolution going forward? >> Well that's the beauty of open, right? What is open-source, open-source brings together a diverse set of perspective and background to innovate faster. And that's where we are, we're seeing a lot of technology evolve. I mean, just think about all of the things that evolved the last two years. Where we've had technologies come up, some go down, but there's so much happening right now, because the time is now. For these companies that are trying to develop more applications, or trying to figure out ways to not only develop these applications, but develop them at scale and really grow those out and build those and IoT, and you're getting more data, and we're capturing those data and operationalizing that data and it comes back to one thing. Applications that can take advantage of that. And so I think there's the potential, as we build out and innovate both the ecosystem but the platform will naturally evolve and take advantage of those winds from these organizations that are driving this to scale. >> So scale is the linchpin. >> Abby: Yeah. >> If you think about traditional paths, environments, if I can use that term, they're limited in scale, and obviously simplicity. Is that another way to think about it? >> I think about it this way, the platform enables you to run fast. You're not running fast with scissors. You want to be able to run fast safely. And so it provides that abstraction and those guardrails so you can quickly iterate and develop and deploy code. If I look at what... HCSE as a company. They went from developing an application, it took them 35 people and nine months to create an app, right? And now with Cloud Foundry, they're able to do it with four people and six weeks. It changes the way you work as an organization. Just imagine as you scale that out, what that means. Imagine the changes that can bring in your organization when you're software-centric and you're customer-first and you're bringing that feedback loop in. >> And you guys do a lot of heavy lifting on behalf of the customer, but you're not hardening it to the point where they can't mold it and shape it to what they want is kind of what I'm-- >> No, we want to abstract away and automate as much as possible, the things you care about. Resiliency, auto-scaling, the ability to do security and compliance, because those are things you care about as an enterprise. Let's make that happen for you, but then give the control to the developer to self provision, to scale, to quickly deploy and iterate, do continuous delivery. All of those things that allow you to go from developing an app once a year to developing an app and iterating on that app constantly, all the time. >> So I've been wanting to ask you to kind of take a step back, and look at the community trends right now. PC Open Stack has a trajectory, it's becoming more of an infrastructure, as a service, kind of settling in there. That's gone through a lot of changes. Seeing a lot of growth in IoT, which we talked about. You're starting to see some movement in the open-source community. CNCF has got traction, The Linux Foundation, Cloud Native, you've got the Kubernetes, I call it the Cold War for orchestration going on right now so it's a really interesting time, microservices are booming. This is the holy grail for developers for the next gen. It's going to be awesome, like machine learning, everyone's getting intoxicated on that these days, so super cool things coming down the pike. >> For sure, I think we're in the coolest time. >> What's going on in the communities, is there any movement, is there trends, is there a sentiment among the developer communities that you see that you could... Any patterns developing around what people are gravitating to? >> I think developers want the freedom to create. They want the ability to create applications and see those come to fruition. I think a lot of things that were new and innovative a couple of years ago and even now, are becoming table stakes. For example, five years ago, having a mobile app as a bank was new and interesting and kind of fun. Now, it's table stakes. Are you going to go bank with a bank that doesn't have one? Are you going to bank with a bank that doesn't have it? It becomes table stakes or, who doesn't, if you don't have fraud detection which is basically event driven responses, right? And so you think about what table stakes are and what, as we think about the abstraction moving up, that's really where it's going to get interesting. >> But open-source community, is it going to move to these new ground, what I'm trying to get at is to see what's happening, what's the trend in the developer community. What's hot, what's fashionable. Is there new projects popping up that you could share that you think is cool and interesting? >> Well they're all cool and interesting. >> John: You'd rather not comment. (laughs) >> I think they're all cool and interesting, I think, you know, CNCF is a sister organization underneath The Linux Foundation. >> John: They kind of inherited that from Kub Con though. Kubernetes Con. >> Yeah, I think they're doing interesting things. I think any organizations that's promoting Cloud Native application architecture and the value of that, we all deserve to be part of the same conversation because to your point earlier, a rising tide lifts all boats. And if every organizations is doing Cloud Native application architectures and Cloud Native solutions, it's going to be super interesting. >> We just had STRAD at Duke, we ran our own event last week called Big Data SV, and it's very clear to us that the big data world industry and Cloud are coming together and the forcing function is machine learning, IoT, and then AI is the appeal, that's the big trend that's kind of, puts a mental model around but IoT is driving this data and the Cloud horsepower is forcing this to move faster. It seems to be very accelerated. >> But, it also enables so much, I mean if you can operationalize this data that you're aggregating and turn it into actionable apps that do things for your business, save money, improve logistics, reach your users better and faster, you start to see the change and the shift that that can bring. You have the data married with the apps, married with the in point sensors and all of a sudden this gets to be a really interesting evolution of technology. >> So what's your hundred day plan, well you're in the hundred day plan already. So what's your plan for this year as new Executive Director for Cloud Foundry, what's on the agenda, what's your top three things you're going to chip away at this year for objectives? >> Developers, developers, developers, does that count as top three? >> More, more, more? Increase the developer count? (laughs) >> Just really, reaching out to the developers and ensuring that they're able to be successful in Cloud Foundry. So I think you'll hear more from us in the next couple of weeks about that. But, ensuring-- >> John: The proof points, basically? >> The proof points, but just ensuring they can be successful and ensuring that scale is affable for them, and then really, our summits are even changing. We've actually added developer tracks to our summit, to make them a place not only where you can learn about Cloud Foundry, but also where you can work with other developers and learn from them and learn about specific languages, but also, how to enable those into Cloud Native application architectures and I think our goal this year is to really enrich that development community and build that pipeline and help fill those gaps. >> And celebrate the wins like the American Airlines of the world, and as IBM and others are successful, then it gets to be less... You don't want to have cognitive dissonance as a developer, that's the worst thing, developers want to make sure they're on a good bus with good people. >> You've obviously got some technology titans behind you, IBM the most prominent, I would say, but obviously guys like VMware, and Cisco, and others, but you've also got [Interference] organizations, guys like Allianz, VW, Allstate I think was early-on in the program. >> JPMC, Citibank. >> Yeah, I shouldn't have started, 'cause I know I'd leave some out, but you're the Executive Director, so you have to fill in the gaps. That's somewhat unique, in a consortium like this. Somewhat, but that many is somewhat unique. Is there more traction there? What's their motivation? >> Abby: As a user? >> Yeah. >> Well, to your earlier point, we're an open-source, right? And what's the value, if I'm an enterprise and I'm looking to take advantage of a platform, but also an open-source platform, open-source allows me to be part of that conversation. I can be a contributor, I can be part of the direction, I can influence where it's going and I think that is a powerful sentiment, for many of these organizations that are looking to evolve and become more software-centric, and this is a good way for them to give back and be part of that momentum. >> And Cloud's exploding, more open-source is needed, it's just a great mission. Congratulations on the new job, and good luck this year. We'll keep in touch, and certainly see you at the Cloud Foundry Summit, that's in San Fransisco again this year? >> Santa Clara, June 13th through 15th. >> John: So every year, you guys always have the fire code problem. (laughs) >> Well I think I'm going to go on record now and officially say this, this will be our last year there, which I think everyone's excited about, 'cause I think we're all over Santa Clara right now. (laughs) >> Alright, well, we'll see you there. Abby Kearns, Executive Director of Cloud Foundry Foundation, here inside the CUBE, powering the Cloud, this is the CUBE's coverage of IBM InterConnect 2017. Stay with us, more coverage after this short break. (bouncy electronic music)
SUMMARY :
brought to you by IBM. This is the CUBE's coverage of IBM's Cloud and data show. Google, I'm sorry, Google, he was formerly at Microsoft, John: To the reins but you're not new, so I've been part of the Cloud Foundry community it's really lifted all the boats, if you will, and are just excited to continue to grow Is it going the way you guys had thought as a community, have the ability to quickly and easily create code So on the development side, sometimes standards can go and that control around the development process And there was a big testimonial with American Airlines in the early 2000s, so I'm thrilled to see them innovating that the platform brings to the table, about developers, is because that's the key, And the app landscape has certainly changed with the platform for many of these applications to be the most open platform to develop applications on, the evolution going forward? and it comes back to one thing. Is that another way to think about it? the platform enables you to run fast. give the control to the developer to self provision, and look at the community trends right now. What's going on in the communities, and see those come to fruition. is it going to move to these new ground, John: You'd rather not comment. I think they're all cool and interesting, I think, John: They kind of inherited that from Kub Con though. it's going to be super interesting. that the big data world industry and Cloud in point sensors and all of a sudden this gets to be for Cloud Foundry, what's on the agenda, what's your that they're able to be successful in Cloud Foundry. to make them a place not only where you can learn about And celebrate the wins like the American Airlines IBM the most prominent, I would say, but obviously the Executive Director, so you have to fill in the gaps. that are looking to evolve and become more software-centric, Congratulations on the new job, and good luck this year. the fire code problem. Well I think I'm going to go on record now here inside the CUBE, powering the Cloud,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Citibank | ORGANIZATION | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Abby Kearns | PERSON | 0.99+ |
JPMC | ORGANIZATION | 0.99+ |
VW | ORGANIZATION | 0.99+ |
Cisco | ORGANIZATION | 0.99+ |
Allianz | ORGANIZATION | 0.99+ |
John | PERSON | 0.99+ |
Swisscom | ORGANIZATION | 0.99+ |
January 2015 | DATE | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
American Airlines | ORGANIZATION | 0.99+ |
San Fransisco | LOCATION | 0.99+ |
CNCF | ORGANIZATION | 0.99+ |
GE | ORGANIZATION | 0.99+ |
Sam | PERSON | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
Siemens | ORGANIZATION | 0.99+ |
Pivotal | ORGANIZATION | 0.99+ |
six weeks | QUANTITY | 0.99+ |
nine months | QUANTITY | 0.99+ |
Allstate | ORGANIZATION | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
Santa Clara | LOCATION | 0.99+ |
Cloud Foundry Foundation | ORGANIZATION | 0.99+ |
four people | QUANTITY | 0.99+ |
35 people | QUANTITY | 0.99+ |
HCSE | ORGANIZATION | 0.99+ |
June 13th | DATE | 0.99+ |
SAP | ORGANIZATION | 0.99+ |
last week | DATE | 0.99+ |
Abby | PERSON | 0.99+ |
15th | DATE | 0.99+ |
five years ago | DATE | 0.99+ |
Cloud Native | ORGANIZATION | 0.99+ |
Las Vegas | LOCATION | 0.99+ |
CUBE | ORGANIZATION | 0.99+ |
Cold War | EVENT | 0.99+ |
Bluemix | ORGANIZATION | 0.99+ |
PaaS | TITLE | 0.99+ |
ORGANIZATION | 0.99+ | |
Cloud Foundry | TITLE | 0.99+ |
both | QUANTITY | 0.99+ |
early 2000s | DATE | 0.99+ |
this year | DATE | 0.98+ |
last year | DATE | 0.98+ |
hundred day | QUANTITY | 0.97+ |
a year ago | DATE | 0.97+ |
one thing | QUANTITY | 0.97+ |
nearly 70 members | QUANTITY | 0.97+ |
Dell EMC | ORGANIZATION | 0.97+ |
Cloud Foundry Summit | EVENT | 0.96+ |
Big Data SV | EVENT | 0.96+ |
first | QUANTITY | 0.95+ |
Josh Rogers, Syncsort - Big Data SV 17 - #BigDataSV - #theCUBE
>> Announcer: Live from San Jose, California, it's The Cube covering Big Data Silicon Valley 2017. (innovative music) >> Welcome back, everyone, Live in Silicon Valley is The Cube's coverage of Big Data SV, our event in Silicon Valley in conjunction with our Big Data NYC for New York City. Every year, twice a year, we get our event going around Strata Hadoop in conjunction with those guys. I'm John Furrier with SiliconANGLE with George Gilbert, our Wikibon (mumbles). Our next guest is Josh Rogers, the CEO of Syncsort, but on many times, Cube alumni, that firm that acquired Trillium, which we talked about yesterday. Welcome back to The Cube, good to see you. >> Good to see you, how are ya? >> So Syncsort is just one of those companies that's really interesting. We were talking about this. I want to get your thoughts on this because I'm not sure if it was in the plan or not, or really ingenius moves by you guys on the manager's side, but Legacy Business, lockdown legacy environments, like the mainframe, and then transform into a modern data company. Was that part of the plan or kind of on purpose by accident? Or what's-- >> Part of the plan. You think about what we've been doing for the last 40 years. We had specific capabilities around managing data at scale and around helping customers who process that data to give more value out of it through analytics, and we've just continually moved through the various kind of generations of technology to apply that same discipline in new environments and big data is frankly been a terrific opportunity for us to apply that same technical and talented DNA in that new environment. It's kind of been running the same game plan. (talking over each other) >> You guys have a good execution, but I think one of the things we were point out, and this is one of those things where, certainly, I live in Palo Alto in Silicon Valley. We love innovation. We love all the shiny, new toys, but you get tempted to go after something really compelling, cool, and relevant, and then go, "Whoa, I forgot about locking down "some of the legacy data stuff," and then you're kind of working down and you guys took a different approach. You going in to the trends from a solid foundation. That's a different execution approach and, like you said, by design, so that's working. >> Yeah, it's definitely working and I think it's also kind of focused on an element that maybe is under-reported, which is a lot of these legacy systems aren't going away, and so one of the big challenges-- >> And this is for record, by the way. >> Right (talking over each other). How do I integrate those legacy environments with these next-generation environments and to do that you have to have expertise on both side, and so one of the things I think we've done a good job is developing that big data expertise and then turning around and saying we can solve that challenge for you, and obviously, the big iron, the big data solutions we bring to market are a perfect example of that, but there's additional solutions that we can provide customers, and we'll talk more about those in a few-- >> Talk about the Trillium acquisition. I want to just, you take a minute to describe that you bought a company called Trillium. What is it, just take a minute to explain what it is and why is it relevant? >> Trillium is a really special company. They are the independent leader in data quality and have been for many years. They've been in the top-right of the gartner magic quadrant for more than a decade, and really, when you look at large, complex, global enterprises, they are the kind of gold-standard in data quality, and when I say data quality, what I mean is an ability to take a dataset, understand the issues with that dataset, and then establish business rules to improve the quality of that data so you can actually trust that data. Obviously that's relevant in a near-adjacency to the data movement and transformation that Syncsort's been known for for so long. What's interesting about it is you think about the development and the maturity of big data environments, specifically Hadoop, you know, people have a desire to obviously do analytics in that data and implicit in that is the ability to trust that data and the way you get there is being able to apply profiling equality rules in that environment, and that's an underserved market today. When we thought about the Trillium acquisition, it was partly, "Hey, this is a great firm "that has so much respect and the space, "and so much talented capability, a powerful capability "and market-leading data quality talent, "but also, we have an ability to apply it "in this next generation environment "much like we did on the ETL and data movement space." And I think that the industry is at a point where enterprises are realizing, "I'm going to need to apply the same "data management disciplines to make use of my data "in my next generation analytics environment "that I did in my data warehouse environment." Obviously, there's different technologies involved. There's different types of data involved. But those disciplines don't go away and being able to improve the quality and be able to kind of build integrity in your datasets is critical, and Trillium is best in market capabilities in that respect. >> Josh, you were telling us earlier about sort of the strategy of knocking down the pins one by one as, you know, it's become clear that we sort of took, first the archive from the data warehouse, and then ETL off-loaded, now progressively more of the business intelligence. What are some of the, besides data quality, what are some of the other functions you have to-- >> There's the whole notion of metadata management, right? And that's incredibly important to support a number of key business initiatives that people want to leverage. There's different styles of movement of data so a thing you'll hear a lot about is change data capture, right, so if I'm moving datasets from source systems into my Hadoop environment, I can move the whole set, but how do I move the incremental changes on a ongoing basis at the speed of business. There's notions of master data management, right? So how do I make sure that I understand and have a gold kind of standard of reference data that I can use to try my own analytic capabilities, and then of course, there's all the analytics that people want to do both in terms of visualization and predictive analytics, but you can think about all these is various engines that I need to apply the data to get maximum value. And it's not so much that these engines aren't important anymore. It's I can now apply them in a different environment that gives me a lot more flexibility, a lot more scale, a better cost structure, and an ability to kind of harness broader datasets. And so that's really our strategy is bring those engines to this new environment. There's two ways to do that. One is build it from scratch, which is kind of a long process to get it right when you're thinking about complex, global, large enterprise requirements. The other is to take existing, tested, proven, best-in-market engines and integrate it deeply in this environment and that's the strategy we've taken. We think that offers a much faster time to value for customers to be able to maximize their investments in this next generation analytics infrastructure. >> So who shares that vision and sort of where are we in the race? >> I think we're fairly unique in our approach of taking that approach. There's certainly other large platform players. They have a broad (mumbles) ability and I think they're working on, "How do I kind of take that architecture and make it relevant?" It ends up creating a co-generation approach. I think that approach has limitations, and I think if you think about taking the core engine and integrate it deeply within the Hadoop ecosystem and Hadoop capabilities, you get a faster time to market and a more manageable solution going forward, and also one that gives you kind of a future pre-shoot from underlying changes that we'll continue to see in the Hadoop component, sort of the big data components, I guess is a better articulation. >> Josh, what's the take on the show this year and the trends, (mumbles) will become a machine learning, and I've seen that. You guys look at your execution plan. What's the landscape happening out there in the show this year? I mean, we're starting to see more business outcome conversations about machine-learning in AI. It's really putting pressure on the companies, and certainly IOT in the cloud-growth as a forcing function. Do you see the same thing? What's your thoughts? >> So machine-learning's a really powerful capability and I think as it relates to the data integration kind of space, there's a lot of benefit to be had. Think about quality. If I have to establish a set of business rules to improve the quality of my data, wouldn't it be great if those little rules could learn as they actually process datasets and see how they change over time, so there's really interesting opportunities there. We're seeing a lot of adoption of cloud. More and more customers are looking at "How do I live in a world where I've got a piece "of my operations on premise, "I've got a piece of operations in cloud, "manage those together and gradually "probably shift more into cloud over time." So I'm doing a lot of work in that space. There's some basic fundamental recognitions that have happened, which is, if I stand up a Hadoop cluster, I am going to have to buy a series of tools to make to get value out of that data in that cluster. That's a good step forward in my perspective because this notion of I'm going to stand up a team off-shore and they're just going to build all these things. >> Cost of ownership goes through the roof. >> Yeah, so I think the industry's moved past this concept of "I make an investment in Hadoop. "I don't need additional solutions." >> It highlights something that we were talking about at Google Next last week about enterprise-ready, and I want to get your thoughts 'cause you guys have a lot of experience, something that's, get in your wheelhouse, how you guys have attacked the market's been pretty impressive and not obvious, and on paper, it looks pretty boring, but you're doing great! I mean, you've done the right strategy, it works. Mainframe, locking in the mainframe, system of record. We've talked this on The Cube. Lots of videos going back three years, but enterprise-ready is a term now that's forcing people, even the best at Google, to be like like, look in the mirror and saying, "Wait a minute. "We have a blind spot." Best tech doesn't always win. You've got table steps; you've got SLAs; you've got mission data quality. One piece of bad data that should be clean could really screw up something. So what's your thoughts on enterprise-ready right now? >> I think that people are recognizing that to get a payoff on a lot of these investments in next generation analytic infrastructure, they're going to need to build, run mission-critical workloads there and take on mission-critical kind of business initiatives and prove out the value. To do that you have to be able to manage the environment, achieve the up-times, have the reliability resiliency that, quite frankly, we've been delivering for four years, and so I think that's another kind of point in our value proposition that frankly seems to be so unique, which is hey, we've been doing this for thousands of customers, the most sophisticated-- >> What are one of the ones that are going to be fatal flaws for people if they don't pay attention to? >> Well, security is huge. I think the manageability, right. So look, if I have to upgrade 25 components in my Hadoop cluster to get to the next version and I need to upgrade all the tools, I've got to have a way to do that that allows me to not only get to the next level of capability that the vendors are providing, but also to do that in a way that doesn't maybe bring down all these mission-critical workloads that have to be 24 by seven. Those pieces are really important and having both the experience and understanding of what that means, and also being able to invest the engineering resources to be able to-- >> And don't forget the sales force. You've got the DNA and the people on the streets. Josh, thanks for coming to The Cube, really appreciate it, great insight. You guys have, just to give you a compliment, great strategy, and again, good execution on your side and as you guys, you're in new territory. Every time we talk to you, you're entering in something new every time, so great to see you. Syncsort here inside The Cube. Always back at sharing commentary on what's going on in the marketplace: AI machine-learning with the table stakes in the enterprise security and what not, still critical for execution and again, IOT is really forcing the function of (mumbles). You've got to focus on the data. Thanks so much. I'm (mumbles). We'll be back with more live coverage after this break. (upbeat innovative music)
SUMMARY :
Announcer: Live from Welcome back to The Cube, good to see you. Was that part of the plan or kind of generations of technology to apply You going in to the trends and to do that you have to a minute to describe and implicit in that is the from the data warehouse, and have a gold kind of and also one that gives you and certainly IOT in the cloud-growth lot of benefit to be had. Cost of ownership Yeah, so I think the even the best at Google, to be like like, and so I think that's of capability that the in the marketplace: AI
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Tristan | PERSON | 0.99+ |
George Gilbert | PERSON | 0.99+ |
John | PERSON | 0.99+ |
George | PERSON | 0.99+ |
Steve Mullaney | PERSON | 0.99+ |
Katie | PERSON | 0.99+ |
David Floyer | PERSON | 0.99+ |
Charles | PERSON | 0.99+ |
Mike Dooley | PERSON | 0.99+ |
Peter Burris | PERSON | 0.99+ |
Chris | PERSON | 0.99+ |
Tristan Handy | PERSON | 0.99+ |
Bob | PERSON | 0.99+ |
Maribel Lopez | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Mike Wolf | PERSON | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
Merim | PERSON | 0.99+ |
Adrian Cockcroft | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Brian | PERSON | 0.99+ |
Brian Rossi | PERSON | 0.99+ |
Jeff Frick | PERSON | 0.99+ |
Chris Wegmann | PERSON | 0.99+ |
Whole Foods | ORGANIZATION | 0.99+ |
Eric | PERSON | 0.99+ |
Chris Hoff | PERSON | 0.99+ |
Jamak Dagani | PERSON | 0.99+ |
Jerry Chen | PERSON | 0.99+ |
Caterpillar | ORGANIZATION | 0.99+ |
John Walls | PERSON | 0.99+ |
Marianna Tessel | PERSON | 0.99+ |
Josh | PERSON | 0.99+ |
Europe | LOCATION | 0.99+ |
Jerome | PERSON | 0.99+ |
ORGANIZATION | 0.99+ | |
Lori MacVittie | PERSON | 0.99+ |
2007 | DATE | 0.99+ |
Seattle | LOCATION | 0.99+ |
10 | QUANTITY | 0.99+ |
five | QUANTITY | 0.99+ |
Ali Ghodsi | PERSON | 0.99+ |
Peter McKee | PERSON | 0.99+ |
Nutanix | ORGANIZATION | 0.99+ |
Eric Herzog | PERSON | 0.99+ |
India | LOCATION | 0.99+ |
Mike | PERSON | 0.99+ |
Walmart | ORGANIZATION | 0.99+ |
five years | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Kit Colbert | PERSON | 0.99+ |
Peter | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Tanuja Randery | PERSON | 0.99+ |
Donna Prlich, Pentaho, Informatica - Big Data SV 17 - #BigDataSV - #theCUBE
>> Announcer: Live from San Jose, California, it's theCUBE. Covering Big Data Silicon Valley 2017. >> Okay, welcome back everyone. Here live in Silicon Valley this is theCUBE. I'm John Furrier, covering our Big Data SV event, #BigDataSV. Our companion event to Big Data NYC, all in conjunction Strata Hadoop, the Big Data World comes together, and great to have guests come by. Donna Prlich, who's the senior VP of products and solutions at Pentaho, a Hitachi company who we've been following before Hitachi had acquired you guys. But you guys are unique in the sense that you're a company within Hitachi left alone after the acquisition. You're now running all the products. Congratulations, welcome back, great to see you. >> Yeah, thank you, good to be back. It's been a little while, but I think you've had some of our other friends on here, as well. >> Yep, and we'll be at Pentaho World, you have Orlando, I think is October. >> Yeah, October, so I'm excited about that, too, so. >> I'm sure the agenda is not yet baked for that because it's early in the year. But what's going on with Hitachi? Give us the update, because you're now, your purview into the product roadmap. The Big Data World, you guys have been very, very successful taking this approach to big data. It's been different and unique to others. >> [Donna} Yep. What's the update? >> Yeah, so, very exciting, actually. So, we've seen, especially at the show that the Big Data World, we all know that it's here. It's monetizable, it's where we, actually, where we shifted five years ago, and it's been a lot of what Pentaho's success has been based on. We're excited because the Hitachi acquisition, as you mentioned, sets us up for the next bit thing, which is IOT. And I've been hearing non-stop about machine learning, but that's the other component of it that's exciting for us. So, yeah, Hitachi, we're-- >> You guys doing a lot of machine learning, a lot of machine learning? >> So we, announced our own kind of own orchestration capabilities that really target how do you, it's less about building models, and how do you enable the data scientists and data preparers to leverage the actual kind of intellectual properties that companies have in those models they've built to transform their business. So we have our own, and then the other exciting piece on the Hitachi side is, on the products, we're now at the point where we're running as Pentaho, but we have access to these amazing labs, which there's about 25 to 50 depending on where you are, whether you're here or in Japan. And those data scientists are working on really interesting things on the R & D side, when you apply those to the kind of use cases we're solving for, that's just like a kid in a candy store with technology, so that's a great-- >> Yeah, you had a built-in customer there. But before I get into Pentaho focusing on what's unique, really happening within you guys with the product, especially with machine learning and AI, as it starts to really get some great momentum. But I want to get your take on what you see happening in the marketplace. Because you've seen the early days and as it's now, hitting a whole another step function as we approach machine learning and AI. Autonomous vehicles, sensors, everything's coming. How are enterprises in these new businesses, whether they're people supporting smart cities or a smart home or automotive, autonomous vehicles. What's the trends you are seeing that are really hitting the pavement here. >> Yeah, I think what we're seeing is, and it's been kind of Pentaho's focus for a long time now, which is it's always about the data. You know, what's the data challenge? Some of the amounts of data which everybody talks about from IOT, and then what's interesting is, it's not about kind of the concepts around AI that have been around forever, but when you start to apply some of those AI concepts to a data pipeline, for instance. We always talk about that 6data pipeline. The reason it's important is because you're really bringing together the data and the analytics. You can't separate those two things, and that's been kind of not only a Pentaho-specific, sort of bent that I've had for years, but a personal one, as well. That, hey, when you start separating it, it makes it really hard to get to any kind of value. So I think what we're doing, and what we're going to be seeing going forward, is applying AI to some of the things that, in a way, will close the gaps between the process and the people, and the data and the analytics that have been around for years. And we see those gaps closing with some of the tools that are emerging around preparing data. But really, when you start to bring some of that machine learning into that picture, and you start applying math to preparing data, that's where it gets really interesting. And I think we'll see some of that automation start to happen. >> So I got to ask you, what is unique about Pentaho? Take a minute to share with the audience some of the unique things that you guys are doing that's different in this sea of people trying to figure out big data. You guys are doing well, an6d you wrote a blog post that I referenced earlier yesterday, around these gaps. How, what's unique about Pentaho and what are you guys doing with examples that you could share? >> Yeah, so I think the big thing about Pentaho that's unique is that it's solving that analytics workflow from the data side. Always from the data. We've always believed that those two things go together. When you build a platform that's really flexible, it's based on open source technology, and you go into a world where a customer says, "I not only want to manage and have a data lake available," for instance, "I want to be able to have that thing extend over the years to support different groups of users. I don't want to deliver it to a tool, I want to deliver it to an application, I want to embed analytics." That's where having a complete end-to-end platform that can orchestrate the data and the analytics across the board is really unique. And what's happened is, it's like, the time has come. Where all we're hearing is, hey, I used to think it was throw some data over and, "here you go, here's the tools." The tools are really easy, so that's great. Now we have all kinds of people that can do analytics, but who's minding the data? With that end-to-end platform, we've always been able to solve for that. And when you move in the open source piece, that just makes it much easier when things like Spark emerge, right. Spark's amazing, right? But we know there's other things on the horizon. Flink, Beam, how are you going to deal with that without being kind of open source, so this is-- >> You guys made a good bet there, and your blog post got my attention because of the title. It wasn't click bait either, it was actually a great article, and I just shared it on Twitter. The Holy Grail of analytics is the value between data and insight. And this is interesting, it's about the data, it's in bold, data, data, data. Data's the hardest part. I get that. But I got to ask you, with cloud computing, you can see the trends of commoditization. You're renting stuff, and you got tools like Kinesis, Redshift on Amazon, and Azure's got tools, so you don't really own that, but the data, you own, right? >> Yeah, that's your intellectual property, right? >> But that's the heart of your piece here, isn't it, the Holy Grail. >> Yes, it is. >> What is that Holy Grail? >> Yeah, that Holy Grail is when you can bring those two things together. The analytics and the data, and you've got some governance, you've got the control. But you're allowing the access that lets the business derive value. For instance, we just had a customer, I think Eric might have mentioned it, but they're a really interesting customer. They're one of the largest community colleges in the country, Ivy Tech, and they won an award, actually, for their data excellence. But what's interesting about them is, they said we're going to create a data democracy. We want data to be available because we know that we see students dropping out, we can't be efficient, people can't get the data that they need, we have old school reporting. So they took Pentaho, and they really transformed the way they think about running their organization and their community colleges. Now they're adding predictive to that. So they've got this data democracy, but now they're looking at things like, "Okay we an see where certain classes are over capacity, but what if we could predict, next year, not only which classes are over capacity, what's the tendency of a particular student to drop out?" "What could we do to intervene?" That's where the kind of cool machine learning starts to apply. Well, Pentaho is what enables that data democracy across the board. I think that's where, when I look at it from a customer perspective, it's really kind of, it's only going to get more interesting. >> And with RFID and smart phones, you could have attendance tracking, too. You know, who's not showing up. >> Yeah absolutely. And you bring Hitachi into the picture, and you think about, for instance, from an IOT perspective, you might be capturing data from devices, and you've got a digital twin, right? And then you bring that data in with data that might be in a data lake, and you can set a threshold, and say, "Okay, not only do we want to be able to know where that student is," or whatever, "we want to trigger something back to that device," and say, "hey, here's a workshop for you to login to right away, so that you don't end up not passing a class." Or whatever it is, it's a simplistic model, but you can imagine where that starts to really become transformative. >> So I asked Eric a question yest6erday. It was from Dave Valante, who's in Boston, stuck in the snowstorm, but he was watching, and I'll ask you and see how it matches. He wrote it differently on Crouch, it was public, but this is in my chat, "HDS is known for main frames, historically, and storage, but Hitachi is an industrial giant. How is Pentaho leveraging the Hitachi monster?" >> Yes, that's a great way to put it. >> Or Godzilla, because it's Japan. >> We were just comparing notes. We were like, "Well, is it an $88 billion company or $90 billion. According to the yen today, it's 88. We usually say 90, but close enough, right? But yeah, it's a huge company. They're in every industry. Make all kinds of things. Pretty much, they've got the OT of the world under their belt. How we're leveraging it is number one, what that brings to the table, in terms of the transformations from a software perspective and data that we can bring to the table and the expertise. The other piece is, we've got a huge opportunity, via the Hitachi channel, which is what's seeing for us the growth that we've had over the last couple of years. It's been really significant since we were acquired. And then the next piece is how do we become part of that bigger Hitachi IOT strategy. And what's been starting to happen there is, as I mentioned before, you can kind of probably put the math together without giving anything away. But you think about capturing, being able to capture device data, being able to bring it into the digital twin, all of that. And then you think about, "Okay, and what if I added Pentaho to the mix?" That's pretty exciting. You bring those things together, and then you add a whole bunch of expertise and machine learning and you're like, okay. You could start to do, you could start to see where the IOT piece of it is where we're really going to-- >> IOT is a forcing function, would you agree? >> Yes, absolutely. >> It's really forcing IT to go, "Whoa, this is coming down fast." And AI and machine learning, and cloud, is just forcing everyone. >> Yeah, exactly. And when we came into the big data market, whatever it was, five years ago, in the early market it's always hard to kind of get in there. But one of the things that we were able to do, when it was sort of, people were still just talking about BI would say, "Have you heard about this stuff called big data, it's going to be hard." You are going to have to take advantage of this. And the same thing is happening with IOT. So the fact that we can be in these environments where customers are starting to see the value of the machine generated data, that's going to be-- >> And it's transformative for the business, like the community college example. >> Totally transformative, yeah. The other one was, I think Eric might have mentioned, the IMS, where all the sudden you're transforming the insurance industry. There's always looking at charts of, "I'm a 17-year-old kid," "Okay, you're rate should be this because you're a 17-year-old boy." And now they're starting to track the driving, and say, "Well, actually, maybe not, maybe you get a discount." >> Time for the self-driving car. >> Transforming, yeah. >> Well, Donna, I appreciate it. Give us a quick tease here, on Pentaho World coming in October. I know it's super early, but you have a roadmap on the product side, so you can see a little bit around the corner. >> Donna: Yeah. >> What is coming down the pike for Pentaho? What are the things that you guys are beavering away at inside the product group? >> Yeah, I think you're going to see some really cool innovations we're doing. I won't, on the Spark side, but with execution engines, in general, we're going to have some really interesting kind of innovative stuff coming. More on the machine learning coming out, and if you think about, if data is, you know what, is the hard part, just think about applying machine learning to the data, and I think you can think of some really cool things, we're going to come up with. >> We're going to need algorithms for the algorithms, machine learning for the machine learning, and, of course, humans to be smarter. Donna, thanks so much for sharing here inside theCUBE, appreciate it. >> Thank you. >> Pentaho, check them out. Going to be at Pentaho World in October, as well, in theCUBE, and hopefully we can get some more deep dives on, with their analyst group, for what's going on with the engines of innovation there. More CUBE coverage live from Silicon Valley for Big Data SV, in conjunction with Strata Hadoop, I'm John Furrier. Be right back with more after this short break. (techno music)
SUMMARY :
it's theCUBE. and great to have guests come by. but I think you've had some you have Orlando, I think is October. Yeah, October, so I'm because it's early in the year. What's the update? that the Big Data World, and how do you enable the data scientists What's the trends you are seeing and the data and the analytics and what are you guys doing that can orchestrate the but the data, you own, right? But that's the heart of The analytics and the data, you could have attendance tracking, too. and you think about, for and I'll ask you and see how it matches. of the transformations And AI and machine learning, and cloud, And the same thing is happening with IOT. for the business, the IMS, where all the on the product side, so and I think you can think for the algorithms, Going to be at Pentaho
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Donna | PERSON | 0.99+ |
Hitachi | ORGANIZATION | 0.99+ |
Donna Prlich | PERSON | 0.99+ |
Dave Valante | PERSON | 0.99+ |
Eric | PERSON | 0.99+ |
Ivy Tech | ORGANIZATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
Boston | LOCATION | 0.99+ |
$88 billion | QUANTITY | 0.99+ |
$90 billion | QUANTITY | 0.99+ |
Japan | LOCATION | 0.99+ |
San Jose, California | LOCATION | 0.99+ |
October | DATE | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
next year | DATE | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
88 | QUANTITY | 0.99+ |
90 | QUANTITY | 0.99+ |
two things | QUANTITY | 0.99+ |
17-year | QUANTITY | 0.99+ |
NYC | LOCATION | 0.99+ |
Big Data SV | ORGANIZATION | 0.98+ |
Orlando | LOCATION | 0.98+ |
five years ago | DATE | 0.98+ |
Pentaho | ORGANIZATION | 0.98+ |
five years ago | DATE | 0.98+ |
today | DATE | 0.98+ |
#BigDataSV | EVENT | 0.98+ |
Informatica | ORGANIZATION | 0.97+ |
one | QUANTITY | 0.97+ |
Silicon Valley | LOCATION | 0.97+ |
Big Data SV | EVENT | 0.95+ |
Spark | TITLE | 0.94+ |
about 25 | QUANTITY | 0.93+ |
17-year-old | QUANTITY | 0.92+ |
Pentaho | PERSON | 0.91+ |
ORGANIZATION | 0.9+ | |
Big Data World | EVENT | 0.9+ |
Azure | ORGANIZATION | 0.88+ |
Big Data Silicon Valley 2017 | EVENT | 0.88+ |
Big Data World | ORGANIZATION | 0.85+ |
Big Data | EVENT | 0.84+ |
Pentaho World | ORGANIZATION | 0.81+ |
Pentaho | LOCATION | 0.8+ |
Kinesis | ORGANIZATION | 0.8+ |
Beam | PERSON | 0.78+ |
6data | QUANTITY | 0.78+ |
Redshift | ORGANIZATION | 0.74+ |
Pentaho World | LOCATION | 0.74+ |
Flink | ORGANIZATION | 0.67+ |
yen | OTHER | 0.66+ |
twin | QUANTITY | 0.63+ |
HDS | ORGANIZATION | 0.62+ |
Crouch | ORGANIZATION | 0.62+ |
earlier yesterday | DATE | 0.62+ |
CUBE | ORGANIZATION | 0.61+ |
last couple of years | DATE | 0.59+ |
Pentaho World | ORGANIZATION | 0.58+ |
50 | QUANTITY | 0.58+ |