Wendi Whitmore, Palo Alto Networks | Palo Alto Networks Ignite22
>>The Cube presents Ignite 22, brought to you by Palo Alto Networks. >>Welcome back to Vegas. Guys. We're happy that you're here. Lisa Martin here covering with Dave Valante, Palo Alto Networks Ignite 22. We're at MGM Grand. This is our first day, Dave of two days of cube coverage. We've been having great conversations with the ecosystem with Palo Alto executives, with partners. One of the things that they have is unit 42. We're gonna be talking with them next about cyber intelligence. And the threat data that they get is >>Incredible. Yeah. They have all the data, they know what's going on, and of course things are changing. The state of play changes. Hold on a second. I got a text here. Oh, my Netflix account was frozen. Should I click on this link? Yeah. What do you think? Have you had a, it's, have you had a little bit more of that this holiday season? Yeah, definitely. >>Unbelievable, right? A lot of smishing going on. >>Yeah, they're very clever. >>Yeah, we're very pleased to welcome back one of our alumni to the queue. Wendy Whitmore is here, the SVP of Unit 42. Welcome back, Wendy. Great to have >>You. Thanks Lisa. So >>Unit 42 created back in 2014. One of the things that I saw that you said in your keynote this morning or today was everything old is still around and it's co, it's way more prolific than ever. What are some of the things that Unit 42 is seeing these days with, with respect to cyber threats as the landscape has changed so much the last two years alone? >>You know, it, it has. So it's really interesting. I've been responding to these breaches for over two decades now, and I can tell you that there are a lot of new and novel techniques. I love that you already highlighted Smishing, right? In the opening gate. Right. Because that is something that a year ago, no one knew what that word was. I mean, we, it's probably gonna be invented this year, right? But that said, so many of the tactics that we have previously seen, when it comes to just general espionage techniques, right? Data act filtration, intellectual property theft, those are going on now more than ever. And you're not hearing about them as much in the news because there are so many other things, right? We're under the landscape of a major war going on between Russia and Ukraine of ransomware attacks, you know, occurring on a weekly basis. And so we keep hearing about those, but ultimately these nations aid actors are using that top cover, if you will, as a great distraction. It's almost like a perfect storm for them to continue conducting so much cyber espionage work that like we may not be feeling that today, but years down the road, they're, the work that they're doing today is gonna have really significant impact. >>Ransomware has become a household word in the last couple of years. I think even my mom knows what it is, to some degree. Yeah. But the threat actors are far more sophisticated than they've ever written. They're very motivated. They're very well funded. I think I've read a stat recently in the last year that there's a ransomware attack once every 11 seconds. And of course we only hear about the big ones. But that is a concern that goes all the way up to the board. >>Yeah. You know, we have a stat in our ransomware threat report that talks about how often victims are posted on leak sites. And I think it's once every seven minutes at this point that a new victim is posted. Meaning a victim has had their data, a victim organization had their data stolen and posted on some leak site in the attempt to be extorted. So that has become so common. One of the shifts that we've seen this year in particular and in recent months, you know, a year ago when I was at Ignite, which was virtual, we talked about quadruple extortion, meaning four different ways that these ransomware actors would go out and try to make money from these attacks in what they're doing now is often going to just one, which is, I don't even wanna bother with encrypting your data now, because that means that in order to get paid, I probably have to decrypt it. Right? That's a lot of work. It's time consuming. It's kind of painstaking. And so what they've really looked to do now is do the extortion where they simply steal the data and then threaten to post it on these leak sites, you know, release it other parts of the web and, and go from there. And so that's really a blending of these techniques of traditional cyber espionage with intellectual property theft. Wow. >>How trustworthy are those guys in terms of, I mean, these are hackers, right? In terms of it's really the, the hacker honor system, isn't it? I mean, if you get compromised like that, you really beholden to criminals. And so, you >>Know, so that's one of the key reasons why having the threat intelligence is so important, right? Understanding which group that you're dealing with and what their likelihood of paying is, what's their modus operandi. It's become even more important now because these groups switch teams more frequently than NFL trades, you know, free agents during the regular season, right? Or players become free agents. And that's because their infrastructure. So the, you know, infrastructure, the servers, the systems that they're using to conduct these attacks from is actually largely being disrupted more from law enforcement, international intelligence agencies working together with public private partnerships. So what they're doing is saying, okay, great. All that infrastructure that I just had now is, is burned, right? It's no longer effective. So then they'll disband a team and then they'll recruit a new team and it's constant like mixing and matching in players. >>All that said, even though that's highly dynamic, one of the other areas that they pride themselves on is customer service. So, and I think it's interesting because, you know, when I said they're not wanting to like do all the decryption? Yeah. Cuz that's like painful techni technical slow work. But on the customer service side, they will create these customer service portals immediately stand one up, say, you know, hey it's, it's like an Amazon, you know, if you've ever had to return a package on Amazon for example, and you need to click through and like explain, you know, Hey, I didn't receive this package. A portal window pops up, you start talking to either a bot or a live agent on the backend. In this case they're hu what appeared to be very much humans who are explaining to you exactly what happened, what they're asking for, super pleasant, getting back within minutes of a response. And they know that in order for them to get paid, they need to have good customer service because otherwise they're not going to, you know, have a business. How, >>So what's the state of play look like from between nation states, criminals and how, how difficult or not so difficult is it for you to identify? Do you have clear signatures? My understanding in with Solar Winds it was a little harder, but maybe help us understand and help our audience understand what the state of play is right now. >>One of the interesting things that I think is occurring, and I highlighted this this morning, is this idea of convergence. And so I'll break it down for one example relates to the type of malware or tools that these attackers use. So traditionally, if we looked at a nation state actor like China or Russia, they were very, very specific and very strategic about the types of victims that they were going to go after when they had zero day. So, you know, new, new malware out there, new vulnerabilities that could be exploited only by them because the rest of the world didn't know about it. They might have one organization that they would target that at, at most, a handful and all very strategic for their objective. They wanted to keep that a secret as long as possible. Now what we're seeing actually is those same attackers going towards one, a much larger supply chain. >>So, so lorenzen is a great example of that. The Hafnia attacks towards Microsoft Exchange server last year. All great examples of that. But what they're also doing is instead of using zero days as much, or you know, because those are expensive to build, they take a lot of time, a lot of funding, a lot of patience and research. What they're doing is using commercially available tools. And so there's a tool that our team identified earlier this year called Brute Rael, C4 or BRC four for short. And that's a tool that we now know that nation state actors are using. But just two weeks ago we invested a ransomware attack where the ransomware actor was using that same piece of tooling. So to your point, yak can get difficult for defenders when you're looking through and saying, well wait, they're all using some of the same tools right now and some of the same approaches when it comes to nation states, that's great for them because they can blend into the noise and it makes it harder to identify as >>Quickly. And, and is that an example of living off the land or is that B BRC four sort of a homegrown hacker tool? Is it, is it a, is it a commercial >>Off the shelf? So it's a tool that was actually, so you can purchase it, I believe it's about 2,500 US dollars for a license. It was actually created by a former Red teamer from a couple well-known companies in the industry who then decided, well hey, I built this tool for work, I'm gonna sell this. Well great for Red teamers that are, you know, legitimately doing good work, but not great now because they're, they built a, a strong tool that has the ability to hide amongst a, a lot of protocols. It can actually hide within Slack and teams to where you can't even see the data is being exfiltrated. And so there's a lot of concern. And then now the reality that it gets into the wrong hands of nation state actors in ransomware actors, one of the really interesting things about that piece of malware is it has a setting where you can change wallpaper. And I don't know if you know offhand, you know what that means, but you know, if that comes to mind, what you would do with it. Well certainly a nation state actor is never gonna do something like that, right? But who likes to do that are ransomware actors who can go in and change the background wallpaper on a desktop that says you've been hacked by XYZ organization and let you know what's going on. So pretty interesting, obviously the developer doing some work there for different parts of the, you know, nefarious community. >>Tremendous amount of sophistication that's gone on the last couple of years alone. I was just reading that Unit 42 is now a founding member of the Cyber Threat Alliance includes now more than 35 organizations. So you guys are getting a very broad picture of today's threat landscape. How can customers actually achieve cyber resilience? Is it achievable and how do you help? >>So I, I think it is achievable. So let me kind of parse out the question, right. So the Cyber Threat Alliance, the J C D C, the Cyber Safety Review Board, which I'm a member of, right? I think one of the really cool things about Palo Alto Networks is just our partnerships. So those are just a handful. We've got partnerships with over 200 organizations. We work closely with the Ukrainian cert, for example, sharing information, incredible information about like what's going on in the war, sharing technical details. We do that with Interpol on a daily basis where, you know, we're sharing information. Just last week the Africa cyber surge operation was announced where millions of nodes were taken down that were part of these larger, you know, system of C2 channels that attackers are using to conduct exploits and attacks throughout the world. So super exciting in that regard and it's something that we're really passionate about at Palo Alto Networks in terms of resilience, a few things, you know, one is visibility, so really having a, an understanding of in a real, as much of real time as possible, right? What's happening. And then it goes into how you, how can we decrease operational impact. So that's everything from network segmentation to wanna add the terms and phrases I like to use a lot is the win is really increasing the time it takes for the attackers to get their work done and decreasing the amount of time it takes for the defenders to get their work done, right? >>Yeah. I I call it increasing the denominator, right? And the ROI equation benefit over or value, right? Equals equals or benefit equals value over cost if you can increase the cost to go go elsewhere, right? Absolutely. And that's the, that's the game. Yeah. You mentioned Ukraine before, what have we learned from Ukraine? I, I remember I was talking to Robert Gates years ago, 2016 I think, and I was asking him, yeah, but don't we have the best cyber technology? Can't we attack? He said, we got the most to lose too. Yeah. And so what have we learned from, from Ukraine? >>Well, I, I think that's part of the key point there, right? Is you know, a great offense essentially can also be for us, you know, deterrent. So in that aspect we have as an, as a company and or excuse me, as a country, as a company as well, but then as partners throughout all parts of the world have really focused on increasing the intelligence sharing and specifically, you know, I mentioned Ukrainian cert. There are so many different agencies and other sorts throughout the world that are doing everything they can to share information to help protect human life there. And so what we've really been concerned with, with is, you know, what cyber warfare elements are going to be used there, not only how does that impact Ukraine, but how does it potentially spread out to other parts of the world critical infrastructure. So you've seen that, you know, I mentioned CS rrb, but cisa, right? >>CISA has done a tremendous job of continuously getting out information and doing everything they can to make sure that we are collaborating at a commercial level. You know, we are sharing information and intelligence more than ever before. So partners like Mania and CrowdStrike, our Intel teams are working together on a daily basis to make sure that we're able to protect not only our clients, but certainly if we've got any information relevant that we can share that as well. And I think if there's any silver lining to an otherwise very awful situation, I think the fact that is has accelerated intelligence sharing is really positive. >>I was gonna ask you about this cause I think, you know, 10 or so years ago, there was a lot of talk about that, but the industry, you know, kind of kept things to themselves, you know, a a actually tried to monetize some of that private data. So that's changing is what I'm hearing from you >>More so than ever more, you know, I've, I mentioned I've been in the field for 20 years. You know, it, it's tough when you have a commercial business that relies on, you know, information to, in order to pay people's salaries, right? I think that has changed quite a lot. We see the benefit of just that continuous sharing. There are, you know, so many more walls broken down between these commercial competitors, but also the work on the public private partnership side has really increased some of those relationships. Made it easier. And you know, I have to give a whole lot of credit and mention sisa, like the fact that during log four J, like they had GitHub repositories, they were using Slack, they were using Twitter. So the government has really started pushing forward with a lot of the newer leadership that's in place to say, Hey, we're gonna use tools and technology that works to share and disseminate information as quickly as we can. Right? That's fantastic. That's helping everybody. >>We knew that every industry, no, nobody's spared of this. But did you notice in the last couple of years, any industries in particular that are more vulnerable? Like I think of healthcare with personal health information or financial services, any industries kind of jump out as being more susceptible than others? >>So I think those two are always gonna be at the forefront, right? Financial services and healthcare. But what's been really top of mind is critical infrastructure, just making sure right? That our water, our power, our fuel, so many other parts of right, the ecosystem that go into making sure that, you know, we're keeping, you know, houses heated during the winter, for example, that people have fresh water. Those are extremely critical. And so that is really a massive area of focus for the industry right now. >>Can I come back to public-private partnerships? My question is relates to regulations because the public policy tends to be behind tech, the technology industry as an understatement. So when you take something like GDPR is the obvious example, but there are many, many others, data sovereignty, you can't move the data. Are are, are, is there tension between your desire as our desire as an industry to share data and government's desire to keep data private and restrict that data sharing? How is that playing out? How do you resolve that? >>Well I think there have been great strides right in each of those areas. So in terms of regulation when it comes to breaches there, you know, has been a tendency in the past to do victim shaming, right? And for organizations to not want to come forward because they're concerned about the monetary funds, right? I think there's been tremendous acceleration. You're seeing that everywhere from the fbi, from cisa, to really working very closely with organizations to, to have a true impact. So one example would be a ransomware attack that occurred. This was for a client of ours within the United States and we had a very close relationship with the FBI at that local field office and made a phone call. This was 7:00 AM Eastern time. And this was an organization that had this breach gone public, would've made worldwide news. There would've been a very big impact because it would've taken a lot of their systems offline. >>Within the 30 minutes that local FBI office was on site said, we just saw this piece of malware last week, we have a decryptor for it from another organization who shared it with us. Here you go. And within 60 minutes, every system was back up and running. Our teams were able to respond and get that disseminated quickly. So efforts like that, I think the government has made a tremendous amount of headway into improving relationships. Is there always gonna be some tension between, you know, competing, you know, organizations? Sure. But I think that we're doing a whole lot to progress it, >>But governments will make exceptions in that case. Especially for something as critical as the example that you just gave and be able to, you know, do a reach around, if you will, on, on onerous regulations that, that ne aren't helpful in that situation, but certainly do a lot of good in terms of protecting privacy. >>Well, and I think there used to be exceptions made typically only for national security elements, right? And now you're seeing that expanding much more so, which I think is also positive. Right. >>Last question for you as we are wrapping up time here. What can organizations really do to stay ahead of the curve when it comes to, to threat actors? We've got internal external threats. What can they really do to just be ahead of that curve? Is that possible? >>Well, it is now, it's not an easy task so I'm not gonna, you know, trivialize it. But I think that one, having relationships with right organizations in advance always a good thing. That's a, everything from certainly a commercial relationships, but also your peers, right? There's all kinds of fantastic industry spec specific information sharing organizations. I think the biggest thing that impacts is having education across your executive team and testing regularly, right? Having a plan in place, testing it. And it's not just the security pieces of it, right? As security responders, we live these attacks every day, but it's making sure that your general counsel and your head of operations and your CEO knows what to do. Your board of directors, do they know what to do when they receive a phone call from Bloomberg, for example? Are they supposed supposed to answer? Do your employees know that those kind of communications in advance and training can be really critical and make or break a difference in an attack. >>That's a great point about the testing but also the communication that it really needs to be company wide. Everyone at every level needs to know how to react. Wendy, it's been so great having, >>Wait one last question. Sure. Do you have a favorite superhero growing up? >>Ooh, it's gotta be Wonder Woman. Yeah, >>Yeah, okay. Yeah, so cuz I'm always curious, there's not a lot of women in, in security in cyber. How'd you get into it? And many cyber pros like wanna save the world? >>Yeah, no, that's a great question. So I joined the Air Force, you know, I, I was a special agent doing computer crime investigations and that was a great job. And I learned about that from, we had an alumni day and all these alumni came in from the university and they were in flight suits and combat gear. And there was one woman who had long blonde flowing hair and a black suit and high heels and she was carrying a gun. What did she do? Because that's what I wanted do. >>Awesome. Love it. We >>Blonde >>Wonder Woman. >>Exactly. Wonder Woman. Wendy, it's been so great having you on the program. We, we will definitely be following unit 42 and all the great stuff that you guys are doing. Keep up the good >>Work. Thanks so much Lisa. Thank >>You. Day our pleasure. For our guest and Dave Valante, I'm Lisa Martin, live in Las Vegas at MGM Grand for Palo Alto Ignite, 22. You're watching the Cube, the leader in live enterprise and emerging tech coverage.
SUMMARY :
The Cube presents Ignite 22, brought to you by Palo Alto One of the things that they have is unit Have you had a, it's, have you had a little bit more of that this holiday season? A lot of smishing going on. Wendy Whitmore is here, the SVP One of the things that I saw that you said in your keynote this morning or I love that you already highlighted Smishing, And of course we only hear about the big ones. the data and then threaten to post it on these leak sites, you know, I mean, if you get compromised like that, you really So the, you know, infrastructure, the servers, the systems that they're using to conduct these attacks from immediately stand one up, say, you know, hey it's, it's like an Amazon, you know, if you've ever had to return a or not so difficult is it for you to identify? One of the interesting things that I think is occurring, and I highlighted this this morning, days as much, or you know, because those are expensive to build, And, and is that an example of living off the land or is that B BRC four sort of a homegrown for Red teamers that are, you know, legitimately doing good work, but not great So you guys are getting a very broad picture of today's threat landscape. at Palo Alto Networks in terms of resilience, a few things, you know, can increase the cost to go go elsewhere, right? And so what we've really been concerned with, with is, you know, And I think if there's any silver lining to an otherwise very awful situation, I was gonna ask you about this cause I think, you know, 10 or so years ago, there was a lot of talk about that, but the industry, And you know, I have to give a whole lot of credit and mention sisa, like the fact that during log four But did you notice in the last couple of years, making sure that, you know, we're keeping, you know, houses heated during the winter, is the obvious example, but there are many, many others, data sovereignty, you can't move the data. of regulation when it comes to breaches there, you know, has been a tendency in the past to Is there always gonna be some tension between, you know, competing, you know, Especially for something as critical as the example that you just And now you're seeing that expanding much more so, which I think is also positive. Last question for you as we are wrapping up time here. Well, it is now, it's not an easy task so I'm not gonna, you know, That's a great point about the testing but also the communication that it really needs to be company wide. Wait one last question. Yeah, How'd you get into it? So I joined the Air Force, you know, I, I was a special agent doing computer We Wendy, it's been so great having you on the program. For our guest and Dave Valante, I'm Lisa Martin, live in Las Vegas at MGM
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Valante | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Wendy | PERSON | 0.99+ |
2014 | DATE | 0.99+ |
FBI | ORGANIZATION | 0.99+ |
Lisa | PERSON | 0.99+ |
Interpol | ORGANIZATION | 0.99+ |
Palo Alto Networks | ORGANIZATION | 0.99+ |
Dave | PERSON | 0.99+ |
Cyber Threat Alliance | ORGANIZATION | 0.99+ |
Bloomberg | ORGANIZATION | 0.99+ |
two days | QUANTITY | 0.99+ |
Cyber Safety Review Board | ORGANIZATION | 0.99+ |
Wendi Whitmore | PERSON | 0.99+ |
Las Vegas | LOCATION | 0.99+ |
last year | DATE | 0.99+ |
Wendy Whitmore | PERSON | 0.99+ |
20 years | QUANTITY | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Palo Alto Networks | ORGANIZATION | 0.99+ |
last week | DATE | 0.99+ |
United States | LOCATION | 0.99+ |
two | QUANTITY | 0.99+ |
J C D C | ORGANIZATION | 0.99+ |
Palo Alto | ORGANIZATION | 0.99+ |
one woman | QUANTITY | 0.99+ |
CISA | ORGANIZATION | 0.99+ |
today | DATE | 0.99+ |
Netflix | ORGANIZATION | 0.99+ |
first day | QUANTITY | 0.99+ |
CrowdStrike | ORGANIZATION | 0.99+ |
Robert Gates | PERSON | 0.99+ |
a year ago | DATE | 0.99+ |
30 minutes | QUANTITY | 0.99+ |
XYZ | ORGANIZATION | 0.99+ |
Vegas | LOCATION | 0.99+ |
zero days | QUANTITY | 0.99+ |
over 200 organizations | QUANTITY | 0.99+ |
Unit 42 | ORGANIZATION | 0.99+ |
more than 35 organizations | QUANTITY | 0.99+ |
Mania | ORGANIZATION | 0.99+ |
GitHub | ORGANIZATION | 0.99+ |
Ignite | ORGANIZATION | 0.98+ |
this year | DATE | 0.98+ |
two weeks ago | DATE | 0.98+ |
one | QUANTITY | 0.98+ |
Microsoft | ORGANIZATION | 0.98+ |
one example | QUANTITY | 0.98+ |
each | QUANTITY | 0.98+ |
GDPR | TITLE | 0.98+ |
millions | QUANTITY | 0.98+ |
zero day | QUANTITY | 0.97+ |
2016 | DATE | 0.97+ |
MGM Grand | LOCATION | 0.97+ |
One | QUANTITY | 0.97+ |
Ukraine | LOCATION | 0.96+ |
one last question | QUANTITY | 0.96+ |
earlier this year | DATE | 0.95+ |
60 minutes | QUANTITY | 0.95+ |
Ukrainian | OTHER | 0.95+ |
unit 42 | OTHER | 0.95+ |
one organization | QUANTITY | 0.94+ |
fbi | ORGANIZATION | 0.93+ |
Intel | ORGANIZATION | 0.92+ |
Russia | ORGANIZATION | 0.92+ |
years ago | DATE | 0.92+ |
about 2,500 US dollars | QUANTITY | 0.92+ |
once every 11 seconds | QUANTITY | 0.9+ |
10 or so years ago | DATE | 0.9+ |
this morning | DATE | 0.89+ |
Day 2 Wrap Up | CrowdStrike Fal.Con 2022
(upbeat music) >> Okay, we're back to wrap up Fal.con 2022 CrowdStrike's customer event. You're watching theCUBE. My name is Dave Vellante. My co-host, Dave Nicholson, is on injured reserve today, so I'm solo. But I wanted to just give the audience a census to some of my quick takeaways. Really haven't given a ton of thought on this. We'll do review after we check out the videos and the transcripts, and do what we do at SiliconANGLE and theCUBE. I'd say the first thing is, look CrowdStrike continues to expand it's footprint. And, it's adding the identity module, through the preempt acquisition. Working very closely with managed service providers, MSPs, managed security service providers. Having an SMB play. So CrowdStrike has 20,000 customers. I think it could, it could 10X that, you know, over some period of time. As I've said earlier, it's on a path by mid-decade to be a 5 billion company, in terms of revenue. At the macro level, security is somewhat, I'd say it's less discretionary than some other investments. You know, you can, you can probably hold off buying a new storage device. You can maybe clean that up. You know, you might be able to hold off on some of your analytics, but at the end of the day, security is not completely non-discretionary. It's competing. The CISO is competing with other budgets. Okay? So it's, while it's less discretionary, it is still, you know, not an open checkbook for the CISO. Now, having said that, from CrowdStrike standpoint it has an excellent opportunity to consolidate tools. It's one of the biggest problems in the security business Go to Optiv and check out their security taxonomy. It'll make your eyes bleed. There's so many tools and companies that are really focused on one specialization. But really, what CrowdStrike can do with its 22 modules, to say, hey, we can give you ROI and consolidate those. And not only is it risk reduction, it's lowering the labor cost and labor intensity, so you can focus on other areas and free up the biggest problem that CISOs have. It's the lack of enough talent. So, really strong business value and value proposition. A lot of that is enabled by the architecture. We've talked about this. You can check out my breaking analysis that I dropped last weekend, on CrowdStrike. And, you know, can it become a generational company. But it's really built on a cloud-native architecture. George Kurtz and company, they shunned having an on-premise architecture. Much like Snowflake Frank Slootman has said, we're not doing a halfway house. We're going to put all our resources on a cloud-native architecture. The lightweight agent that allows them to add new modules and collect more data, and scale out. The purpose-built threat graph and and time series database, and asset graph that they've built. And very strong use of AI, to not only stop known malware, but stop unknown malware. Identify threats. Do that curation. And really, you know, support the SecOp teams. Product wise, I think the big three takeaways, and there were others, but the big three for me is EDR extending into XDR. You know, X is the extending for, in really, the core of endpoint detection and response, extending that further. Well, it seems to be a big buzzword these days. CrowdStrike, I think, is very focused on making a more complete, a holistic offering, beyond endpoint. And I think it's going to do very well in that space. They're not alone. There are others. It's a very competitive space. The second is identity. Through the acquisition of Preempt. CrowdStrike building that identity module. Partnering with leaders like Okta, to really provide that sort of, treating identity, if you will, as an endpoint. And then sort of Humio is now Falcon Log Scale. Bringing together, you know, the data and the observability piece, and the security piece, is kind of the three big product trends that I saw. I think the last point I'll make, before we wrap, is the ecosystem. The ecosystem here is good. It reminds me, I said, a number of times this week, of ServiceNow in 2013 I think the difference is, CrowdStrike has an SMB play it can go after many more customers, and actually have an even broader platform. And I think it can accelerate its ecosystem faster than ServiceNow was able to do that. I mean, it's got to be, sort of, an open and collaborative sort of ecosystem. You know, ServiceNow is kind of, more of, a one-way street. And I think the other piece of that ecosystem, that we see evolving, into IOT, into the operations technology and critical infrastructure. Which is so important, because critical infrastructure of nations is so vulnerable. We're seeing this in the Ukraine. Security is a key component now of any warfare. And going forward, it's always going to be a key component. Nation states are going to go after trust, or secure infrastructure, or critical infrastructure. Try to disable that and disrupt that. So securing those operation assets is going to be very critical. Not just the refrigerator and the coffee maker, but really going after those critical infrastructures. (chuckles) Getting asked to break. And the last thing I'll say, is the developer platform. We heard from ML that, the opportunity that's there, to build out a PaaS layer, super PaaS layer, if you will, so that developers can add value. I think if that happens, this ecosystem, which is breaking down, will explode. This is Dave Vellante, wrapping up at CrowdStrike, Fal.con 2022, Fal.con 2022. Go to SiliconAngle.com, for all the news. Check out theCUBE.net. You'll see these videos on demand and many others. Check out (indistinct).com for all the research. And look for where we'll be next. Of course, re:Invent is the big fall event, but there are many others in between. Thanks for watching. We're out. (music plays out)
SUMMARY :
is kind of the three big
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Nicholson | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Frank Slootman | PERSON | 0.99+ |
2013 | DATE | 0.99+ |
10X | QUANTITY | 0.99+ |
5 billion | QUANTITY | 0.99+ |
20,000 customers | QUANTITY | 0.99+ |
22 modules | QUANTITY | 0.99+ |
Ukraine | LOCATION | 0.99+ |
CrowdStrike | EVENT | 0.99+ |
George Kurtz | PERSON | 0.99+ |
second | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
Okta | ORGANIZATION | 0.98+ |
CrowdStrike | ORGANIZATION | 0.97+ |
this week | DATE | 0.96+ |
Fal.con 2022 | EVENT | 0.95+ |
SiliconANGLE | ORGANIZATION | 0.95+ |
first thing | QUANTITY | 0.94+ |
one | QUANTITY | 0.92+ |
CISO | ORGANIZATION | 0.92+ |
theCUBE.net | OTHER | 0.91+ |
indistinct).com | OTHER | 0.9+ |
theCUBE | ORGANIZATION | 0.9+ |
ServiceNow | TITLE | 0.89+ |
ML | ORGANIZATION | 0.87+ |
one specialization | QUANTITY | 0.87+ |
last weekend | DATE | 0.87+ |
Invent | EVENT | 0.87+ |
PaaS | TITLE | 0.86+ |
CrowdStrike Fal.Con 2022 | EVENT | 0.86+ |
Optiv | ORGANIZATION | 0.86+ |
Snowflake | ORGANIZATION | 0.85+ |
Humio | ORGANIZATION | 0.82+ |
three big product | QUANTITY | 0.81+ |
Day 2 | QUANTITY | 0.79+ |
one-way | QUANTITY | 0.78+ |
ServiceNow | ORGANIZATION | 0.71+ |
SecOp | ORGANIZATION | 0.66+ |
three | QUANTITY | 0.63+ |
SiliconAngle.com | OTHER | 0.61+ |
CrowdStrike | TITLE | 0.59+ |
Preempt | ORGANIZATION | 0.56+ |
Falcon Log Scale | OTHER | 0.48+ |
mid | QUANTITY | 0.44+ |
Greg Muscarella, SUSE | Kubecon + Cloudnativecon Europe 2022
>>The cube presents, Coon and cloud native con Europe 22, brought to you by the cloud native computing foundation. >>Welcome to Valencia Spain and con cloud native con 20 Europe, 2022. I'm your host, Keith Townson alongside a new host en Rico senior reti, senior editor. I'm sorry, senior it analyst at giong Enrique. Welcome to the program. >>Thank you very much. And thank you for having me. It's exciting. >>So thoughts, high level thoughts of CU con first time in person again in couple years? >>Well, this is amazing for several reasons. And one of the reasons is that yeah, I had the chance to meet, uh, with, uh, you know, people like you again. I mean, we, we met several times over the internet, over zoom codes. I, I started to eat these zoom codes. <laugh> because they're very impersonal in the end. And like last night we, we are together group of friends, industry folks. It's just amazing. And a part of that, I mean, the event is, uh, is a really cool, it's really cool. There are a lot from people interviews and, you know, real people doing real stuff, not just, uh, you know, again, in personal calls, you don't even know if they're telling the truth, but when you can, you know, look in their eyes, what they're doing, I, I think that's makes a difference. >>So speaking about real people, meeting people for the first time, new jobs, new roles, Greg Moscarella enterprise container management in general manager at SUSE, welcome to the show, welcome back clue belong. >>Thank you very much. It's awesome to be here. It's awesome to be back in person. And I completely agree with you. Like there's a certain fidelity to the conversation and a certain, uh, ability to get to know people a lot more. So it's absolutely fantastic to be here. >>So Greg, tell us about your new role and what SUSE has gone on at KU con. >>Sure. So I joined SA about three months ago to lead the rancher business unit, right? So our container management pieces and, you know, it's a, it's a fantastic time. Cause if you look at the transition from virtual machines to containers and to moving to micro services, right alongside that transition from on-prem to cloud, like this is a very exciting time to be in this industry and rancher's been setting the stage. And again, I'm go back to being here. Rancher's all about the community, right? So this is a very open, independent, uh, community driven product and project. And so this, this is kinda like being back to our people, right. And being able to reconnect here. And so, you know, doing it, digital is great, but, but being here is changes the game for us. So we, we feed off that community. We feed off the energy. So, uh, and again, going back to the space and what's happening in it, great time to be in this space. And you guys have seen the transitions you've seen, I mean, we've seen just massive adoption, uh, of containers and Kubernetes overall, and rancher has been been right there with some amazing companies doing really interesting things that I'd never thought of before. Uh, so I'm, I'm still learning on this, but, um, but it's been great so far. >>Yeah. And you know, when we talk about strategy about Kubernetes today, we are talking about very broad strategies. I mean, not just the data center or the cloud with, you know, maybe smaller organization adopting Kubernetes in the cloud, but actually large organization thinking guide and more and more the edge. So what's your opinion on this, you know, expansion of Kubernetes towards the edge. >>So I think you're, I think you're exactly right. And that's actually a lot of meetings I've been having here right now is these are some of these interesting use cases. So people who, uh, whether it be, you know, ones that are easy to understand in the telco space, right? Especially the adoption of 5g and you have all these base stations, new towers, and they have not only the core radio functions or network functions that they're trying to do there, but they have other applications that wanna run on that same environment, uh, spoke recently with some of our, our good friends at a major automotive manufacturer, doing things in their factories, right. That can't take the latency of being somewhere else. Right? So they have robots on the factory floor, the latency that they would experience if they tried to run things in the cloud meant that robot would've moved 10 centimeters. >>By the time, you know, the signal got back, it may not seem like a lot to you, but if, if, if you're an employee, you know, there, you know, uh, a big 2000 pound robot being 10 centimeters closer to you may not be what you, you really want. Um, there's, there's just a tremendous amount of activity happening out there on the retail side as well. So it's, it's amazing how people are deploying containers in retail outlets. You know, whether it be fast food and predicting, what, what, how many French fries you need to have going at this time of day with this sort of weather. Right. So you can make sure those queues are actually moving through. It's, it's, it's really exciting and interesting to look at all the different applications that are happening. So yes, on the edge for sure, in the public cloud, for sure. In the data center and we're finding is people want to common platform across those as well. Right? So for the management piece too, but also for security and for policies around these things. So, uh, it really is going everywhere. >>So talk to me, how do, how are we managing that as we think about pushing stuff out of the data center, out of the cloud cloud, closer to the edge security and life cycle management becomes like top of mind thought as, as challenges, how is rancher and sushi addressing >>That? Yeah. So I, I think you're, again, spot on. So it's, it starts off with the think of it as simple, but it's, it's not simple. It's the provisioning piece. How do we just get it installed and running right then to what you just asked the management piece of it, everything from your firmware to your operating system, to the, the cluster, uh, the Kubernetes cluster, that's running on that. And then the workloads on top of that. So with rancher, uh, and with the rest of SUSE, we're actually tacking all those parts of the problems from bare metal on up. Uh, and so we have lots of ways for deploying that operating system. We have operating systems that are, uh, optimized for the edge, very secure and ephemeral container images that you can build on top of. And then we have rancher itself, which is not only managing your Kubernetes cluster, but can actually start to manage the operating system components, uh, as well as the workload components. >>So all from your single interface, um, we mentioned policy and security. So we, yeah, we'll probably talk about it more, um, uh, in a little bit, but, but new vector, right? So we acquired a company called new vector, just open sourced, uh, that here in January, that ability to run that level of, of security software everywhere again, is really important. Right? So again, whether I'm running it on, whatever my favorite public cloud providers, uh, managed Kubernetes is, or out at the edge, you still have to have security, you know, in there. And, and you want some consistency across that. If you have to have a different platform for each of your environments, that's just upping the complexity and the opportunity for error. So we really like to eliminate that and simplify our operators and developers lives as much as possible. >>Yeah. From this point of view, are you implying that even you, you are matching, you know, self, uh, let's say managed clusters at the, at the very edge now with, with, you know, added security, because these are the two big problems lately, you know, so having something that is autonomous somehow easier to manage, especially if you are deploying hundreds of these that's micro clusters. And on the other hand, you need to know a policy based security that is strong enough to be sure again, if you have these huge robots moving too close to you, because somebody act the class that is managing them, that could be a huge problem. So are you, you know, approaching this kind of problems? I mean, is it, uh, the technology that you are acquired, you know, ready to, to do this? >>Yeah. I, I mean, it, it really is. I mean, there's still a lot of innovation happening. Don't, don't get me wrong. We're gonna see a lot of, a lot more, not just from, from SA and rancher, but from the community, right. There's a lot happening there, but we've come a long way and we've solved a lot of problems. Uh, if I think about, you know, how do you have this distributed environment? Uh, well, some of it comes down to not just, you know, all the different environments, but it's also the applications, you know, with microservices, you have very dynamic environment now just with your application space as well. So when we think about security, we really have to evolve from a fairly static policy where like, you might even be able to set an IP address in a port and some configuration on that. It's like, well, your workload's now dynamically moving. >>So not only do you have to have that security capability, like the ability to like, look at a process or look at a network connection and stop it, you have to have that, uh, manageability, right? You can't expect an operator or someone to like go in and manually configure a YAML file, right? Because things are changing too fast. It needs to be that combination of convenient, easy to manage with full function and ability to protect your, your, uh, your resources. And I think that's really one of the key things that new vector really brings is because we have so much intelligence about what's going on there. Like the configuration is pretty high level, and then it just runs, right? So it's used to this dynamic environment. It can actually protect your workloads wherever it's going from pod to pod. Uh, and it's that, that combination, again, that manageability with that high functionality, um, that, that is what's making it so popular. And what brings that security to those edge locations or cloud locations or your data center >>Mm-hmm <affirmative> so one of the challenges you're kind of, uh, touching on is this abstraction on upon abstraction. When I, I ran my data center, I could put, uh, say this IP address, can't talk to this IP address on this port. Then I got next generation firewalls where I could actually do, uh, some analysis. Where are you seeing the ball moving to when it comes to customers, thinking about all these layers of abstraction I IP address doesn't mean anything anymore in cloud native it's yes, I need one, but I'm not, I'm not protecting based on IP address. How are customers approaching security from the name space perspective? >>Well, so it's, you're absolutely right. In fact, even when you go to I P six, like, I don't even recognize IP addresses anymore. <laugh> >>Yeah. Doesn't mean anything like, oh, just a bunch of, yes, those are numbers, ER, >>And colons. Right. You know, it's like, I don't even know anymore. Right. So, um, yeah, so it's, it comes back to that, moving from a static, you know, it's the pets versus cattle thing. Right? So this static thing that I can sort of know and, and love and touch and kind of protect to this almost living, breathing thing, which is moving all around, it's a swarm of, you know, pods moving all over the place. And so, uh, it, it is, I mean, that's what Kubernetes has done for the workload side of it is like, how do you get away from, from that, that pet to a declarative approach to, you know, identifying your workload and the components of that workload and what it should be doing. And so if we go on the security side some more like, yeah, it's actually not even namespace namespace. >>Isn't good enough. We wanna get, if we wanna get to zero trust, it's like, just cuz you're running in my namespace doesn't mean I trust you. Right. So, and that's one of the really cool things about new vectors because of the, you know, we're looking at protocol level stuff within the network. So it's pod to pod, every single connection we can look at and it's at the protocol layer. So if you say you're on my database and I have a mye request going into it, I can confirm that that's actually a mye protocol being spoken and it's well formed. Right. And I know that this endpoint, you know, which is a, uh, container image or a pod name or some, or a label, even if it's in the same name, space is allowed to talk to and use this protocol to this other pod that's running in my same name space. >>Right. So I can either allow or deny. And if I can, I can look into the content that request and make sure it's well formed. So I'll give you an example is, um, do you guys remember the log four J challenges from not too long ago, right. Was, was a huge deal. So if I'm doing something that's IP and port based and name space based, so what are my protections? What are my options for something that's got log four J embedded in like I either run the risk of it running or I shut it down. Those are my options. Like those neither one of those are very good. So we can do, because again, we're at the protocol layers like, ah, I can identify any log for J protocol. I can look at whether it's well formed, you know, or if it's malicious, if it's malicious, I can block it. If it's well formed, I can let it go through. So I can actually look at those, those, um, those vulnerabilities. I don't have to take my service down. I can run and still be protected. And so that, that extra level, that ability to kind of peek into things and also go pod to pod, you know, not just name space level is one of the key differences. So I talk about the evolution or how we're evolving with, um, with the security. Like we've grown a lot, we've got a lot more coming. >>So let's talk about that a lot more coming what's in the pipeline for SUSE. >>Well, how, before I get to that, we just announced new vector five. So maybe I can catch us up on what was released last week. Uh, and then we can talk a little bit about going, going forward. So new vector five, introduce something called um, well, several things, but one of the things I can talk in more detail about is something called zero drift. So I've been talking about the network security, but we also have run time security, right? So any, any container that's running within your environment has processes that are running that container. What we can do is actually comes back to that manageability and configuration. We can look at the root level of trust of any process that's running. And as long as it has an inheritance, we can let that process run without any extra configuration. If it doesn't have a root level of trust, like it didn't spawn from whatever the, a knit, um, function was and that container we're not gonna let it run. Uh, so the, the configuration that you have to put in there is, is a lot simpler. Um, so that's something that's in, in new vector five, um, the web application firewall. So this layer seven security inspection has gotten a lot more granular now. So it's that pod Topo security, um, both for ingress egress and internal on the cluster. Right. >>So before we get to what's in the pipeline, one question around new vector, how is that consumed and deployed? >>How is new vector consumed, >>Deployed? And yeah, >>Yeah, yeah. So, uh, again with new vector five and, and also rancher 2 65, which just were released, there's actually some nice integration between them. So if I'm a rancher customer and I'm using 2 65, I can actually just deploy that new vector with a couple clicks of the button in our, uh, in our marketplace. And we're actually tied into our role-based access control. So an administrator who has that has the rights can just click they're now in a new vector interface and they can start setting those policies and deploying those things out very easily. Of course, if you aren't using, uh, rancher, you're using some other, uh, container management platform, new vector still works. Awesome. You can deploy it there still in a few clicks. Um, you're just gonna get into, you have to log into your new vector, uh, interface and, and use it from there. >>So that's how it's deployed. It's, it's very, it's very simple to use. Um, I think what's actually really exciting about that too, is we've opensourced it? Um, so it's available for anyone to go download and try, and I would encourage people to give it a go. Uh, and I think there's some compelling reasons to do that now. Right? So we have pause security policies, you know, depreciated and going away, um, pretty soon in, in Kubernetes. And so there's a few things you might look at to make sure you're still able to run a secure environment within Kubernetes. So I think it's a great time to look at what's coming next, uh, for your security within your Kubernetes. >>So, Paul, we appreciate you stopping by from ity of Spain. I'm Keith Townsend, along with en Rico Sinte. Thank you. And you're watching the, the leader in high tech coverage.
SUMMARY :
brought to you by the cloud native computing foundation. Welcome to the program. And thank you for having me. I had the chance to meet, uh, with, uh, you know, people like you again. So speaking about real people, meeting people for the first time, new jobs, So it's absolutely fantastic to be here. So Greg, tell us about your new role and what SUSE has gone So our container management pieces and, you know, it's a, it's a fantastic time. you know, maybe smaller organization adopting Kubernetes in the cloud, So people who, uh, whether it be, you know, By the time, you know, the signal got back, it may not seem like a lot to you, to what you just asked the management piece of it, everything from your firmware to your operating system, If you have to have a different platform for each of your environments, And on the other hand, you need to know a policy based security that is strong have to evolve from a fairly static policy where like, you might even be able to set an IP address in a port and some So not only do you have to have that security capability, like the ability to like, Where are you seeing the In fact, even when you go to I P six, like, it comes back to that, moving from a static, you know, it's the pets versus cattle thing. And I know that this endpoint, you know, and also go pod to pod, you know, not just name space level is one of the key differences. the configuration that you have to put in there is, is a lot simpler. Of course, if you aren't using, uh, rancher, you're using some other, So I think it's a great time to look at what's coming next, uh, for your security within your So, Paul, we appreciate you stopping by from ity of Spain.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Keith Townson | PERSON | 0.99+ |
SUSE | ORGANIZATION | 0.99+ |
Greg Muscarella | PERSON | 0.99+ |
Paul | PERSON | 0.99+ |
10 centimeters | QUANTITY | 0.99+ |
Keith Townsend | PERSON | 0.99+ |
January | DATE | 0.99+ |
Greg Moscarella | PERSON | 0.99+ |
last week | DATE | 0.99+ |
Spain | LOCATION | 0.99+ |
Greg | PERSON | 0.99+ |
2000 pound | QUANTITY | 0.99+ |
one question | QUANTITY | 0.98+ |
Kubernetes | TITLE | 0.98+ |
one | QUANTITY | 0.98+ |
both | QUANTITY | 0.98+ |
Valencia Spain | LOCATION | 0.97+ |
today | DATE | 0.97+ |
Kubecon | ORGANIZATION | 0.97+ |
first time | QUANTITY | 0.95+ |
single interface | QUANTITY | 0.95+ |
two big problems | QUANTITY | 0.95+ |
each | QUANTITY | 0.94+ |
Coon | ORGANIZATION | 0.94+ |
ingress | ORGANIZATION | 0.94+ |
zero | QUANTITY | 0.9+ |
three months ago | DATE | 0.9+ |
Cloudnativecon | ORGANIZATION | 0.88+ |
22 | EVENT | 0.86+ |
SUSE | TITLE | 0.86+ |
five | TITLE | 0.85+ |
I P six | OTHER | 0.84+ |
Europe | LOCATION | 0.81+ |
giong Enrique | PERSON | 0.81+ |
log four | OTHER | 0.8+ |
2 65 | COMMERCIAL_ITEM | 0.79+ |
2022 | DATE | 0.78+ |
vector five | TITLE | 0.77+ |
couple years | QUANTITY | 0.75+ |
rancher | ORGANIZATION | 0.73+ |
French | OTHER | 0.73+ |
cloud native computing | ORGANIZATION | 0.73+ |
Kubernetes | ORGANIZATION | 0.72+ |
last night | DATE | 0.71+ |
single connection | QUANTITY | 0.71+ |
one of the reasons | QUANTITY | 0.69+ |
Rico | ORGANIZATION | 0.68+ |
Rico Sinte | PERSON | 0.67+ |
SA | ORGANIZATION | 0.66+ |
about | DATE | 0.66+ |
layer seven | OTHER | 0.65+ |
vector | OTHER | 0.64+ |
5g | QUANTITY | 0.64+ |
65 | COMMERCIAL_ITEM | 0.62+ |
cloud native con | ORGANIZATION | 0.55+ |
telco | ORGANIZATION | 0.55+ |
2 | TITLE | 0.54+ |
SA | LOCATION | 0.53+ |
egress | ORGANIZATION | 0.52+ |
hundreds | QUANTITY | 0.51+ |
CU con | EVENT | 0.46+ |
KU con. | ORGANIZATION | 0.44+ |
vector | COMMERCIAL_ITEM | 0.39+ |
20 | EVENT | 0.31+ |
Ravi Maira, Synk | AWS Startup Showcase S2 E1 | Open Cloud Innovations
>>Hello everyone. And welcome to the cubes presentation of the AWS startup showcase open cloud innovations. This is season two episode one of our showcase ongoing series. We're covering very exciting startups from the AWS ecosystem. And we're going to be talking about the open source community. I'm your host, Lisa Martin. And today I'm excited to be joined by Robbie, Myra, the head of product and partner marketing at sneak. Robbie's here to talk with me about developer security for your digital transformation. Robbie, it's great to have you on the cube. >>Thanks Lisa. Nice to be here. >>So talk to me about what's going on in developer land. They're under a lot of pressure. A lot of them are building apps with open source, but what does sneak seeing from the developers lens >>From the developer's lens? There's a lot of pressure to build fast and that's probably the biggest challenge, right? We're in a world of digital transformation where everybody's trying to compete no matter what industry you're in, right on the technology and on the quality of your software or the capabilities of your software, which puts a lot of pressure on developers to build fast. That causes them to do a few things. One, it causes them to build, to develop in a way where they're doing constant iteration and so models that would have enabled a security check to come in at the end, aren't working anymore because they don't have time for those security checks. And it also causes them to do a good thing, which is to leverage other people's code when they can like open source. So they can just focus on, on their own functionality. And that's true, whether they're building new functionality or modernizing legacy applications by moving them to the cloud. >>So it's a high percentage of, of app code 80 to 90% is open source. Then that opens up. Talk to me about w where the vulnerabilities are and how you guys help customers and developers address that. >>Yeah, the vulnerabilities can be anywhere, but the key is that that point, right? If you're using open source in a typical application, 80 to 90 plus percent of the lines of code in that application are going to be open source code, their code. Somebody else wrote that you don't have a direct relationship with, and yet you own the risk that whatever they may have, whatever vulnerabilities may be in their code, you now own that risk. So what we're trying to do with sneakers, trying to do is enable developers to leverage open source, but do that securely. And then we also help them with the 10% that they rent as well, and, and do that all in one really easy environment for a developer that fits into their workflow and into their daily life. >>So security should shift left. I've had the chance to talk with a couple of, do you call them sneakers sneakers? Oh, you do a couple of sneakers recently. We've talked about security shifting lab. That's not a new concept, but I'd love to dig in more to how sneak and AWS do that. And I'm also curious if what you're doing helps. We've talked about the cybersecurity skills got for a long time. Now, just what you guys do, help address that >>It does because it's really leveraging a resource that, that is there, right? There's the number of developers worldwide is growing from, depending on who you believe for these numbers and their estimated numbers, right? But 25 million to 50 million over roughly a five-year period that's already started. So we're somewhere in the 30 now, right? Meanwhile, the security jobs, there's something like 9 million cyber security people in the world, and that's all cyber security roles. It's a much shorter, a smaller chunk that are application security folks. And there's three and a half million unfilled cybersecurity roles. So you can't get cyber security people and keep using the current model you're using. But just scale it linearly, you have to change things. And sneaks belief is the way you change things is you have the developers be part of your security solution, which means they need to have the ability to not only develop, but to develop securely. And that's our concept of developer security. We build tools and a platform that enables developers to be the first part of the security solution and enable security teams rather than individually auditing and fixing things to develop a process, govern the process, guide the development teams, but let the developers own that first step of security. And that's really how you solve that scale problem. >>When you're talking with customers, is this kind of a better together scenario, developers and security folks? Are you helping them align culturally because this is a change? >>Absolutely. I think one of the biggest misconceptions out there is that there's a tension between security and development. And I think that's because organizationally there might be right. Security is responsible for risk and developers responsible for speed of innovation and the faster you innovate, potentially there's more risk. So there might be some organizational tension, but at the human level, people understand each other, they understand the pressures that the other one's going through. They just don't have an easy way to work together. And if you can help them get that, then they, it really takes off it. The relationships form they'll build human to human programs like security champion programs and things to, to integrate the teams because they're both going after the same goal, both sides want to build awesome technology and grow in whatever market they're in. >>Right. And of course, with the need to do that at today's markets speed and scale is a great thing that you guys are doing to facilitate that collaboration. And of course the security let's kind of take a double-click now into the different integrations that sneek has with AWS services. I know there's quite a few, >>There's quite a few. The biggest one, probably the easiest one for the integrations is the native integration that we have with code pipeline. So it makes it easy for developers as they're finishing their builds and deploying to have an automatic security check that comes in, understands if there's things that need to be fixed before this really should be released, and then they can fix it and go forward. But we integrate across with our API across a lot of other services, ECR EKS code builder, so that wherever the developer is working, there's a way for us to integrate with them as they're building across their AWS development process. >>Okay. So giving them plenty of opportunity, let's dig into the platform. Talk to me about the platform, how it's really aimed at developers. You alluded to this a little bit, but I'd like to kind of take a double-click into the technology. >>Sure. That the platform, it, part of it is that idea of it we've wrapped it all as a developer tool. But the thing that makes sneak unique in this is not only we have the idea that we wanted to shift left in time, but we wanted to shift left in ownership. So the developers are primary user and we built a tool that is a developer tool that happens to do security. And we've extended that tool into a platform by enabling it to connect into the developers tools, sharing information, across different elements of what it securing. So for example, the open source that we're scanning for you and testing to find for vulnerabilities, we're also looking at the vulnerabilities in your code and where they may overlap or intersect. We can adjust priorities so that you might not need to fix something. Let's say you're using an open source, vulnerable, a package that has a vulnerability, but your code is never going to access that you don't need to fix it. >>So you can prioritize that one lower, right? Same thing with Kubernetes and containers. You may have a container vulnerability, but the way you're going to leverage the container that won't be used so we can adjust the priority to make it easy for the developer. And that's the other big thing that's different about a developer security platform than a typical security tool. A typical security tool is an audit tool it's designed to output. Here are all the things you have a problem with a developer security tool is a fixing tool. It's just defined as a, here are the problems you have developed with here's how you fix it and go back to building on that. That prioritization is a big part of that, because you can say, here's what you don't need to worry about. And then you can focus the rest of your energy on helping developers fix the problem either by giving them really good advice or automating it for them and saying, Hey, here's a button click that will generate a pull request. And your problem is this fixed. >>It must go a long way to improving developer productivity, one facilitating that speed and the agility with which they need to work, but also from a developer kind of crowd sourcing, crowd swell perspective. I imagine, talk to me about what some of the voices are, the developers that are in your community. What are some of the things that they're saying in terms of how much faster they're able to work, they're able to get those priorities established with automation so much faster? >>Well, that's the biggest thing. Is there a, the productivity gain happens because of the benefit of shift left, right? You're testing earlier. You're finding it at an earlier time when it's easier to fix, but that's because they're the ones doing it, right. If they're waiting to hand off to an auto report and then it comes back, even if somebody is, is giving them them audit faster, it's still after they've moved on. And the other way people try to solve it as well. They'll say, well, I'll take a security tool then to hand it to the developer and they can run it. But so developers are not security experts. So the tool needs to understand what they know and what they don't know, and, and working in an upload. And that's what developers generally say to us because sneak makes it easy to work, but also focuses on the fix and helps them guide them to that, to that answer. Then they're able to go much faster when we're evaluated by companies who are looking for a security solution. If the developers get involved in that evaluation, they'll choose sneak. >>So I'm curious a little bit about as, as the head of product marketing, I'm thinking customer advisory boards, things like that. What's the collaboration like between sneak and the developers to really tune and push the technology forward. I imagine it's quite collaborative, >>Quite collaborative and it's across a lot of, of spectrum. So we do have a customer advisory board and that's generally leaders, right? That's either security leaders or development leaders or operations leaders who are in that advisory board. And they're giving us input on things they need for program-wide governance or program wide adoption. We also have a developer community where we're talking directly to developers and that's where we get a lot of, Hey, here's how I could use this better as a developer. And that guides where we focus features that help developers work better, whether it's integrations with our IDs or whether it's the way we present information, help them prioritize. And then the third part is we have a lot of people using the tool because it has a free model, right? We're as a developer tool, we have a freemium model. There's a level of sneak that developers can use that they don't need to pay for. That's not a temporary trial, it's forever. If you want to use it at that level and we can observe what they're doing. So that observability gives us another insight into where folks get challenged run into, to struggles. And then we can look to address those in our roadmap as well. So, so all of that together really helps us drive the product forward. >>What is the perspective from the analyst view? You talked a little bit about the perspective from the customer. We'll get into a customer story in a bit, but I'd love to know what are the gardeners saying? >>Well, Gardner especially put us, we debuted in their magic quadrant for application security last year. And we did David as a visionary and sort of the highest part of the visionary quadrant you could get in before you crossed over into leader, which is kind of unheard of for a first time into the, into the quadrant. And the main reason for that is that they have built the way those, those magic quadrants are built is they have key capabilities and then they score companies against key capabilities and they weight those capabilities, you know, by order of importance. And Gardner has started to put some of this notion of developer security and cross cloud native application security into those key capabilities. And those tend to align really well with what sneakers. So they have a, for example, a software composition, which is sort of open source security analysis, where first, w w w where the top ranking in that, where the top ranking and container security, where the top ranking and developer enablement. So that's pulling us, they are so-so Gardner and the analyst community is seeing this same demand coming from their customers. And that's really aligning to where our vision is. >>And in terms of kind of propelling that vision forward, the voice of the customer, the voice of the analyst, aligning with what you guys are doing to kind of lead the vision going forward. I want to get into some of the intelligence before we kind of break into a customer example. Talk to me a little bit about snakes security intelligence, what the key capabilities are, and some customers that are leveraging it. Sure. >>The biggest thing is with all the developer tool wrapping that needs to be in this product than it is a developer tool. It's got a developers heart, but it has to have a security brain because it still is a security tool. There are some developer tools. We try to have little check the box capabilities of security and they'll crowdsource for vulnerabilities potentially. But if you're doing this, you need to make sure that all the vulnerabilities that could be found are in the database to be able to be found that the database is comprehensive, that it's timely. They get in very quickly that it's accurate. You don't waste time on false positives because that will turn developers off faster than anything. And that it's actionable. So when it does find something, it helps you go forward with it. And that's where sneaks really focused on. So we collect data from multiple public sources. >>We also have a fairly large proprietary research team that curates that information determines what needs to go in. Sometimes we'll adjust priorities. And we also get a lot of contributions from other sources like community contributions. Again, that big free user base of ours is giving us input academia. Open source groups are also in their social media trends. So if we see something trending on Twitter, then that'll not only get it into the database, but it'll drive prioritization. And that's a big part of what's in sneak Intel, which is the name we use for our vulnerability database. We also have a machine learning algorithm. That's constantly looking at all the code in public, in public applications and repositories. And we use that to train for our own proprietary code testing tool, but it also just gets a lot of it finds things there as well. So it brings a really good source of information that helps people make sure you're finding the vulnerabilities, you're prioritizing them correctly and fixing them. And so Amazon's one who is the, you know, one of the folks that using that tool where one of the primary sources of, of Amazon inspector for open source vulnerabilities, as well as a bunch of other security companies like rapid seven tenable and, and others. >>One of the things I was reading from, I'm always kind of looking at the differentiators and I'm sure you are as the head of product marketing and partner marketing, but it sounds like the database can, is, is a key differentiator finding vulnerabilities up to what is it? 46 days faster than competitors. >>Yeah. I mean, faster than especially public sources, which are the easier ones to, to know how you're doing against, but that's a big part of us. So when I talked about those categories, that's really what we measure ourselves against. How are we doing in terms of comprehensive? Do we have the vulnerabilities that we should have? So we have over four times the number of vulnerabilities as the next largest publicly available database, we find them faster, so timely. So that's at 46 days getting it in faster or faster than other public sources, they get into our solution and then accuracy. Again, we, it's not a stat we can test because you can't test it just from the database. You have to run the tools of our, of others in this space. And we don't have those, but making sure that you're not hitting a lot of false positives is a big part of it as well. >>Got it. Okay. And we only have a couple minutes left, but there's two more areas that I want to dig into with you just crack crack. The surface one is log four, shallow was reading. Snake says this. We were the perfect solution at the perfect time. Unpack that for me in the next minute or so. >>Yeah. And that's a bit, and it kind of wraps back to what we were talking about earlier. Everybody's using open source. If you're in the Java world, a lot of folks had logged for shell and we're using lock for shell for logging as a part of their, as a part of their applications. And so a lot of our customers, I think it was over 30%, 36% of our paying customers had the vulnerability. And you would only have the vulnerability of your Java. So it's a very large percentage of our Java using my customers had the vulnerability, but because they were using sneak, they were able, once we put it in the database, which we did the day, it was disclosed, they were able to find it and fix it very quickly. So 91% of our customers fixed that vulnerability in just two days, 98%, because this was a rolling thunder event, right. There was a vulnerability. And then there was a second vulnerability in the, in the fix. And then there was a vulnerability, even in the fix of that. So the second vulnerability that came out because everybody had been ready for it from the first time 98% picks within two days. Whereas the median number of days to generally fix a vulnerability is over two months. So really fast addressing the solution. >>So those are really impressive. And speaking of stats, I wanted to get into just really quickly a case study that really shows that lasting is one of your customer. One of your many customers, big developer community there about 3,500 developers. Give me some kind of the high level of business outcomes that at Lasagne is, is, is achieving thanks to sneaky. >>Yeah. I mean the biggest one is that almost 99% of their applications are deployed in containers. So being able to have the containers tested for vulnerabilities as they're being deployed before they're being deployed is huge for them to reduce the risk of a vulnerability. They, they had a 65% reduction in high severity container volumes a few months after using sneak across all those developers, which really reduces your, your risk profile of your, of your cloud native applications. They're obviously a big AWS user as well. So, so for them, that was the big thing. And again, it goes to that scale, right? They've got 3 3500 developers, more than 3,500 developers. If you try to go through the security team and have the security team fixing all those things, you'll just never catch up. >>Got it. Last question. Where can I get this available through the AWS market prays marketplace? You mentioned the freemium model, give folks kind of a direction on where to go. >>Yeah. So I would say if you are a, if you're someone in the security team, if you're a buyer, the AWS marketplace is a great place to go because you can probably leverage your existing spend commits with AWS. It's easy to purchase, easy billing, et cetera. If you're a developer, then there is this free version where you might go and just start using it and get comfort for it. And if you are a buyer, talk to your developers because there's a pretty good chance. Someone in your company, that's a developer is already using. Sneak will be comfortable with it. These solutions are only successful. If the developers actually use it, you can't shift left unless the developers pick it up and use it. So using the one that developers are already using is probably a good idea. >>Awesome. Robbie, this has been a great conversation, so much momentum at snake. You're the third sneaker I'd gotten to speak to you in the last month and I have, it's pretty exciting, but thanks for walking us through the technology, the capabilities, the differentiators, the voice of the customer, the voice of the analyst, we appreciate your insights and your time. And we look forward to next time we talk to you. >>Terrific. Lisa, I look forward to it as well, but there's a lot more Smith sneakers to go through before you get back to me again. I guess >>I look forward to adding to my repertoire of sneaker interviews, Ravi. Thanks so much. Thank you for Ravi Myra. I'm Lisa Martin. You're watching this cube interview as part of the AWS startup showcase. Stick around more great content coming up next.
SUMMARY :
Robbie, it's great to have you on the cube. So talk to me about what's going on in developer land. And it also causes them to do a good thing, which is to leverage other people's code when they can Talk to me about w where the vulnerabilities are and how you guys the lines of code in that application are going to be open source code, their code. I've had the chance to talk with a couple of, do you call them sneakers sneakers? And sneaks belief is the way you change things is you have the developers Security is responsible for risk and developers responsible for speed of innovation and the faster you And of course the security that we have with code pipeline. Talk to me about the platform, So the developers are primary user and we built a tool that is a developer tool that happens to And that's the other big thing that's that speed and the agility with which they need to work, but also from but also focuses on the fix and helps them guide them to that, to that answer. sneak and the developers to really tune and push the the way we present information, help them prioritize. You talked a little bit about the perspective from the customer. of the visionary quadrant you could get in before you crossed over into leader, which is kind of unheard of the voice of the analyst, aligning with what you guys are doing to kind of lead the vision the database to be able to be found that the database is comprehensive, that it's timely. of the primary sources of, of Amazon inspector for open source vulnerabilities, One of the things I was reading from, I'm always kind of looking at the differentiators and I'm sure you are as the as the next largest publicly available database, we find them faster, Unpack that for me in the next minute or so. Whereas the median number of days to generally fix a vulnerability is over two months. Give me some kind of the high level of business outcomes that at Lasagne is, And again, it goes to that scale, You mentioned the freemium model, give folks kind of a direction on where to go. the AWS marketplace is a great place to go because you can probably leverage your existing spend commits with AWS. You're the third sneaker I'd gotten to speak to you in the last month and I have, it's pretty exciting, but thanks for walking us through I guess I look forward to adding to my repertoire of sneaker interviews, Ravi.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Lisa Martin | PERSON | 0.99+ |
Robbie | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
25 million | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
10% | QUANTITY | 0.99+ |
80 | QUANTITY | 0.99+ |
65% | QUANTITY | 0.99+ |
36% | QUANTITY | 0.99+ |
David | PERSON | 0.99+ |
Ravi | PERSON | 0.99+ |
Lasagne | ORGANIZATION | 0.99+ |
46 days | QUANTITY | 0.99+ |
second vulnerability | QUANTITY | 0.99+ |
91% | QUANTITY | 0.99+ |
98% | QUANTITY | 0.99+ |
Myra | PERSON | 0.99+ |
30 | QUANTITY | 0.99+ |
50 million | QUANTITY | 0.99+ |
last year | DATE | 0.99+ |
two days | QUANTITY | 0.99+ |
Ravi Myra | PERSON | 0.99+ |
first part | QUANTITY | 0.99+ |
3 3500 developers | QUANTITY | 0.99+ |
Java | TITLE | 0.99+ |
over 30% | QUANTITY | 0.99+ |
more than 3,500 developers | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
Gardner | PERSON | 0.99+ |
One | QUANTITY | 0.99+ |
third sneaker | QUANTITY | 0.99+ |
first step | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
three and a half million | QUANTITY | 0.98+ |
two more areas | QUANTITY | 0.98+ |
Snake | PERSON | 0.98+ |
third part | QUANTITY | 0.98+ |
90% | QUANTITY | 0.98+ |
ORGANIZATION | 0.98+ | |
five-year | QUANTITY | 0.98+ |
over two months | QUANTITY | 0.98+ |
Ravi Maira | PERSON | 0.97+ |
one | QUANTITY | 0.97+ |
both sides | QUANTITY | 0.97+ |
about 3,500 developers | QUANTITY | 0.97+ |
first time | QUANTITY | 0.96+ |
last month | DATE | 0.96+ |
almost 99% | QUANTITY | 0.94+ |
90 plus percent | QUANTITY | 0.93+ |
first | QUANTITY | 0.93+ |
9 million cyber | QUANTITY | 0.91+ |
over four times | QUANTITY | 0.89+ |
Intel | ORGANIZATION | 0.86+ |
Kubernetes | TITLE | 0.83+ |
double | QUANTITY | 0.81+ |
couple | QUANTITY | 0.8+ |
Smith | ORGANIZATION | 0.75+ |
double-click | QUANTITY | 0.75+ |
episode one | QUANTITY | 0.71+ |
Synk | ORGANIZATION | 0.71+ |
season two | QUANTITY | 0.7+ |
Startup Showcase S2 E1 | EVENT | 0.68+ |
couple minutes | QUANTITY | 0.63+ |
Donald Fischer, Tidelift | AWS Startup Showcase S2 E1 | Open Cloud Innovations
>>Welcome everyone to the cubes presentation of the AWS startup showcase open cloud innovations. This is season two episode one of the ongoing series and we're covering exciting and innovative startups from the AWS ecosystem. Today. We're going to focus on the open source community. I'm your host, Dave Vellante. And right now we're going to talk about open source security and mitigating risk in light of a recent discovery of a zero day flaw in log for J a Java logging utility and a related white house executive order that points to the FTC pursuing companies that don't properly secure consumer data as a result of this vulnerability and with me to discuss this critical issue and how to more broadly address software supply chain risk is Don Fisher. Who's the CEO of tide lift. Thank you for coming on the program, Donald. >>Thanks for having me excited to be here. Yeah, pleasure. >>So look, there's a lot of buzz. You open the news, you go to your favorite news site and you see this, you know, a log for J this is an, a project otherwise known as logged for shell. It's this logging tool. My understanding is it's, it's both ubiquitous and very easy to exploit. Maybe you could explain that in a little bit more detail. And how do you think this vulnerability is going to affect things this year? >>Yeah, happy to, happy to dig in a little bit in orient around this. So, you know, just a little definitions to start with. So log for J is a very widely used course component that's been around for quite a while. It's actually an amazing piece of technology log for J is used in practically every serious enterprise Java application over the last 10 going on 20 years. So it's, you know, log for J itself is fantastic. The challenge that organization organizations have been facing relate to a specific security vulnerability that was discovered in log for J and that has been given this sort of brand's name as it happens these days. Folks may remember Heartbleed around the openness to sell vulnerability some years back. This one has been dubbed logged for shell. And the reason why it was given that name is that this is a form of security vulnerability that actually allows attackers. >>You know, if a system is found that hasn't been patched to remediate it, it allows hackers to get full control of a, of a system of a server that has the software running on it, or includes this log for J component. And that means that they can do anything. They can access, you know, private customer data on that system, or really do anything and so-called shell level access. So, you know, that's the sort of definitions of what it is, but the reason why it's important is in the, in the small, you know, this is a open door, right? It's a, if, if organizations haven't patched this, they need to respond to it. But one of the things that's kind of, you know, I think important to recognize here is that this log for J is just one of literally thousands of independently created open source components that flow into the applications that almost every organization built and all of them all software is going to have security vulnerabilities. And so I think that log for J is, has been a catalyst for organizations to say, okay, we've got to solve this specific problem, but we all also have to think ahead about how is this all gonna work. If our software supply chain originates with independent creators across thousands of projects across the internet, how are we going to put a better plan in place to think ahead to the next log for J log for shell style incident? And for sure there will be more >>Okay. So you see this incident as a catalyst to maybe more broadly thinking about how to secure the, the digital supply chain. >>Absolutely. Yeah, it's a, this is proving a point that, you know, a variety of folks have been making for a number of years. Hey, we depend, I mean, honestly these days more than 70% of most applications, most custom applications are comprised of this third party open source code. Project's very similar in origin and governance to log for J that's just reality. It's actually great. That's an amazing thing that the humans collaborating on the internet have caused to be possible that we have this rich comments of open source software to build with, but we also have to be practical about it and say, Hey, how are we going to work together to make sure that that software as much as possible is vetted to ensure that it meets commercial standards, enterprise standards ahead of time. And then when the inevitable issues arise like this incident around the log for J library, that we have a great plan in place to respond to it and to, you know, close the close the door on vulnerabilities when they, when they show up. >>I mean, you know, when you listen to the high level narrative, it's easy to point fingers at organizations, Hey, you're not doing enough now. Of course the U S government has definitely made attempts to emphasize this and, and shore up in, in, in, in, in push people to shore up the software supply chain, they've released an executive order last may, but, but specifically, I mean, it's just a complicated situation. So what steps should organizations really take to make sure that they don't fall prey to these future supply chain attacks, which, you know, are, as you pointed out are inevitable. >>Yeah. I mean, it's, it's a great point that you make that the us federal government has taken proactive steps starting last year, 2021 in the fallout of the solar winds breach, you know, about 12 months ago from the time that we're talking, talking here, the U S government actually was a bit ahead of the game, both in flagging the severity of this, you know, area of concern and also directing organizations on how to respond to it. So the, in May, 2021, the white house issued an executive order on cybersecurity and it S directed federal agencies to undertake a whole bunch of new measures to ensure the security of different aspects of their technology and software supply chain specifically called out open source software as an area where they put, you know, hard requirements around federal agencies when they're acquiring technology. And one of the things that the federal government that the white house cybersecurity executive order directed was that organizations need to start with creating a list of the third-party open source. >>That's flowing into their applications, just that even have a table of contents or an index to start working with. And that's, that's called a, a software bill of materials or S bomb is how some people pronounce that acronym. So th the federal government basically requires federal agencies to now create Nessbaum for their applications to demand a software bill of materials from vendors that are doing business with the government and the strategy there has been to expressly use the purchasing power of the us government to level up industry as a whole, and create the necessary incentives for organizations to, to take this seriously. >>You know, I, I feel like the solar winds hack that you mentioned, of course it was widely affected the government. So we kind of woke them up, but I feel like it was almost like a stuck set Stuxnet moment. Donald were very sophisticated. I mean, for the first time patches that were supposed to be helping us protect, now we have to be careful with them. And you mentioned the, the bill of its software, bill of materials. We have to really inspect that. And so let's get to what you guys do. How do you help organizations deal with this problem and secure their open source software supply chain? >>Yeah, absolutely happy to tell you about, about tide lift and, and how we're looking to help. So, you know, the company, I co-founded the company with a couple of colleagues, all of whom are long-term open source folks. You know, I've been working in around commercializing open source for the last 20 years that companies like red hat and, and a number of others as have my co-founders the opportunity that we saw is that, you know, while there have been vendors for some of the traditional systems level, open source components and stacks like Linux, you know, of course there's red hat and other vendors for Linux, or for Kubernetes, or for some of the databases, you know, there's standalone companies for these logs, for shell style projects, there just hasn't been a vendor for them. And part of it is there's a challenge to cover a really vast territory, a typical enterprise that we inspect has, you know, upwards of 10,000 log for shell log for J like components flowing into their application. >>So how do they get a hand around their hands around that challenge of managing that and ensuring it needs, you know, reasonable commercial standards. That's what tide lifts sets out to do. And we do it through a combination of two elements, both of which are fairly unique in the market. The first of those is a purpose-built software solution that we've created that keeps track of the third-party open source, flowing into your applications, inserts itself into your DevSecOps tool chain, your developer tooling, your application development process. And you can kind of think of it as next to the point in your release process, where you run your unit test to ensure the business logic in the code that your team is writing is accurate and sort of passes tests. We do a inspection to look at the state of the third-party open source packages like Apache log for J that are flowing into your, into your application. >>So there's a software element to it. That's a multi-tenant SAS service. We're excited to be partnered with, with AWS. And one of the reasons why we're here in this venue, talking about how we are making that available jointly with AWS to, to drink customers deploying on AWS platforms. Now, the other piece of the, of our solution is really, really unique. And that's the set of relationships that Tyler has built directly with these independent open source maintainers, the folks behind these open source packages that organizations rely on. And, you know, this is where we sort of have this idea. Somebody is making that software in the first place, right? And so would those folks be interested? Could we create a set of aligned incentives to encourage them, to make sure that that software meets a bunch of enterprise standards and areas around security, like, you know, relating to the log for J vulnerability, but also other complicated parts of open source consumption like licensing and open source license, accuracy, and compatibility, and also maintenance. >>Like if somebody looking after the software going forward. So just trying to basically invite open source creators, to partner with us, to level up their packages through those relationships, we get really, really clean, clear first party data from the folks who create, maintain the software. And we can flow that through the tools that I described so that end organizations can know that they're building with open source components that have been vetted to meet these standards, by the way, there's a really cool side effect of this business model, which is that we pay these open source maintainers to do this work with us. And so now we're creating a new income stream around what previously had been primarily a volunteer activity done for impact in this universe of open source software. We're helping these open source maintainers kind of GoPro on an aspect of what they do around open source. And that means they can spend more time applying more process and tools and methodology to making that open source software even better. And that's good for our customers. And it's good for everyone who relies on open source software, which is really everyone in society these days. That's interesting. I >>Was going to ask you what's their incentive other than doing the right thing. Can you give us an example of, of maybe a example of an open source maintainer that you're working with? >>Yeah. I mean, w we're working with hundreds of open source maintainers and a few of the key open source foundations in different areas across JavaScript, Java PHP, Ruby python.net, and, you know, like examples of categories of projects that we're working with, just to be clear, are things like, you know, web frameworks or parser libraries or logging libraries, like a, you know, log for J and all the other languages, right? Or, you know, time and date manipulation libraries. I mean, they, these are sort of the, you know, kind of core building blocks of applications and individually, they, you know, they may seem like, you know, maybe a minor, a minor thing, but when you multiply them across how many applications these get used in and log for J is a really, really clarifying case for folks to understand this, you know, what can seemingly a small part of your overall application estate can have disproportionate impact on, on your operations? As we saw with many organizations that spent, you know, a weekend or a week, or a large part of the holidays, scrambling to patch and remediate this, a single vulnerability in one of those thousands of packages in that case log. >>Okay, got it. So you have this two, two headed, two vectors that I'm going to call it, your ecosystem, your relationship with these open source maintainers is kind of a, that just didn't happen overnight, and it develop those relationships. And now you get first party data. You monetize that with a software service that is purpose built as the monitor of the probe that actually tracks that third, third party activity. So >>Exactly right. Got it. >>Okay. So a lot of companies, Donald, I mean, this is, like I said before, it's a complicated situation. You know, a lot of people don't have the skillsets to deal with this. And so many companies just kind of stick their head in the sand and, you know, hope for the best, but that's not a great strategy. What are the implications for organizations if they don't really put the tools and processes into place to manage their open source, digital supply chain. >>Yeah. Ignoring the problem is not a viable strategy anymore, you know, and it's just become increasingly clear as these big headline incidents that happened like Heartbleed and solar winds. And now this logged for shell vulnerability. So you can, you can bet on that. Continuing into the future and organizations I think are, are realizing the ones that haven't gotten ahead of this problem are realizing this is a critical issue that they need to address, but they have help, right. You know, the federal government, another action beyond that cybersecurity executive order that was directed at federal agencies early last year, just in the last week or so, the FTC of the U S federal trade commission has made a much more direct warning to private companies and industry saying that, you know, issues like this log for J vulnerability risk exposing private, you know, consumer data. That is one of the express mandates of the FTC is to avoid that the FTC has said that this is, you know, bears on both the federal trade commission act, as well as the Gramm-Leach-Bliley act, which relates to consumer data privacy. >>And the FTC just came right out and said it, they said they cited the $700 million settlements that Equifax was subject to for their data breach that also related to open source component, by the way, that that had not been patched by, by Equifax. And they said the FTC intents to use its full legal authority to pursue companies that failed to take reasonable steps, to protect consumer data from exposure as a result of log for J or similar known vulnerabilities in the future. So the FTC is saying, you know, this is a critical issue for consumer privacy and consumer data. We are going to enforce against companies that do not take reasonable precautions. What are reasonable precautions? I think it's kind of a mosaic of solutions, but I'm glad to say tide lift is contributing a really different and novel solution to the mix that we hope will help organizations contend with this and avoid that kind of enforcement action from FTC or other regulators. >>Well, and the good news is that you can tap a tooling like tide lift in the cloud as a service and you know, much easier today than it was 10 or 15 years ago to, to resolve, or at least begin to demonstrate that you're taking action against this problem. >>Absolutely. There's new challenges. Now I'm moving into a world where we build on a foundation of independently created open source. We need new solutions and new ideas, and that's a, you know, that's part of what we're, we're, we're showing up with from the tide lift angle, but there's many other elements that are going to be necessary to provide the full solution around securing the open source supply chain going forward. >>Well, Donald Fisher of tide lift, thanks so much for coming to the cube and best of luck to your organization. Thanks for the good work that you guys do. >>Thanks, Dave. Really appreciate your partnership on this, getting the word out and yeah, thanks so much for today. >>Very welcome. And you are watching the AWS startup showcase open cloud innovations. Keep it right there for more action on the cube, your leader in enterprise tech coverage.
SUMMARY :
order that points to the FTC pursuing companies that don't properly secure consumer Thanks for having me excited to be here. You open the news, you go to your favorite news site and you see this, So it's, you know, log for J itself is fantastic. But one of the things that's kind of, you know, I think important to recognize here is that this the, the digital supply chain. Yeah, it's a, this is proving a point that, you know, a variety of folks have been making for I mean, you know, when you listen to the high level narrative, it's easy to point fingers at organizations, Hey, you're not doing enough now. the solar winds breach, you know, about 12 months ago from the time that we're talking, So th the federal government basically requires federal agencies And so let's get to what you guys do. a typical enterprise that we inspect has, you know, And you can kind of think of it as next to the point in And, you know, this is where we sort of have this idea. open source creators, to partner with us, to level up their packages through Was going to ask you what's their incentive other than doing the right thing. folks to understand this, you know, what can seemingly a small part of your overall application And now you get first party data. Got it. you know, hope for the best, but that's not a great strategy. of the FTC is to avoid that the FTC has said that this is, So the FTC is saying, you know, this is a critical issue for Well, and the good news is that you can tap a tooling like you know, that's part of what we're, we're, we're showing up with from the tide lift angle, Thanks for the good work that you guys do. And you are watching the AWS startup showcase open cloud innovations.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Donald Fisher | PERSON | 0.99+ |
Equifax | ORGANIZATION | 0.99+ |
May, 2021 | DATE | 0.99+ |
Dave | PERSON | 0.99+ |
Don Fisher | PERSON | 0.99+ |
Donald | PERSON | 0.99+ |
$700 million | QUANTITY | 0.99+ |
U S federal trade commission | ORGANIZATION | 0.99+ |
two elements | QUANTITY | 0.99+ |
JavaScript | TITLE | 0.99+ |
two | QUANTITY | 0.99+ |
FTC | ORGANIZATION | 0.99+ |
both | QUANTITY | 0.99+ |
Today | DATE | 0.99+ |
Tyler | PERSON | 0.99+ |
first | QUANTITY | 0.99+ |
Java | TITLE | 0.99+ |
last week | DATE | 0.99+ |
Donald Fischer | PERSON | 0.99+ |
more than 70% | QUANTITY | 0.99+ |
Linux | TITLE | 0.98+ |
10 | DATE | 0.98+ |
two vectors | QUANTITY | 0.98+ |
one | QUANTITY | 0.98+ |
tide lift | ORGANIZATION | 0.98+ |
hundreds | QUANTITY | 0.98+ |
last year | DATE | 0.98+ |
Gramm-Leach-Bliley act | TITLE | 0.98+ |
10,000 log | QUANTITY | 0.97+ |
today | DATE | 0.97+ |
white house | ORGANIZATION | 0.97+ |
zero day | QUANTITY | 0.97+ |
Nessbaum | PERSON | 0.97+ |
U S government | ORGANIZATION | 0.96+ |
early last year | DATE | 0.96+ |
thousands | QUANTITY | 0.96+ |
Java PHP | TITLE | 0.96+ |
Ruby python.net | TITLE | 0.95+ |
this year | DATE | 0.95+ |
first time | QUANTITY | 0.95+ |
federal trade commission act | TITLE | 0.95+ |
about 12 months ago | DATE | 0.95+ |
20 years | QUANTITY | 0.94+ |
Stuxnet | PERSON | 0.93+ |
a week | QUANTITY | 0.93+ |
15 years ago | DATE | 0.93+ |
single vulnerability | QUANTITY | 0.93+ |
thousands of projects | QUANTITY | 0.92+ |
2021 | DATE | 0.92+ |
GoPro | ORGANIZATION | 0.92+ |
J | TITLE | 0.92+ |
Heartbleed | EVENT | 0.91+ |
DevSecOps | TITLE | 0.84+ |
FTC | TITLE | 0.83+ |
Tidelift | ORGANIZATION | 0.78+ |
Apache | ORGANIZATION | 0.78+ |
SAS | ORGANIZATION | 0.77+ |
last 20 years | DATE | 0.77+ |
a weekend | QUANTITY | 0.73+ |
some years back | DATE | 0.73+ |
season two | QUANTITY | 0.72+ |
episode | QUANTITY | 0.71+ |
Startup Showcase S2 E1 | EVENT | 0.7+ |
hat | TITLE | 0.69+ |
federal government | ORGANIZATION | 0.69+ |
Peter Cho | KubeCon + CloudNativeCon NA 2021
(soft techno music) >> Good evening. Welcome back to the Kube. Live in Los Angeles. We are at KubeCon Cloud Native Con 2021. Lisa Martin with Dave Nicholson, rounding out our day. We're going to introduce you to a new company, a new company that's new to us. I should say, log DNA. Peter Choi joins us the VP of product. Peter, welcome to the program. >> Thanks for having me. >> (Lisa) Talk to us about log DNA. Who are you guys? What do you do? >> So, you know, log DNA is a log medicine platform. Traditionally, we've been focused on, you know, offering log analysis, log management capabilities to dev ops teams. So your classic kind of troubleshooting, debugging, getting into your systems. More recently, maybe in like the last year or so we've been focused on a lot of control functionality around log medicine. So what I mean by that is a lot of people typically think of kind of the analysis or the dashboards, but with the pandemic, we noticed that you see this kind of surge of data because all of the services are being used, but you also see a downward pressure on cost, right? Because all of a sudden you don't want to be spending two X on those digital experiences. So we've been focused really on kind of tamping down kind of controls on the volume of log data coming in and making sure that they have a higher kind of signal and noise ratio. And then, you know, I'll talk about it a little bit, but we've really been honing in on how can we take those capabilities and kind of form them more in a pipeline. So log management, dev ops, you know, focusing on log data, but moving forward really focused on that flow of data. >> (Dave) So, when you talk about the flow of data and logs that are being read, make this a little more real, bring it up, bring it up just to level in terms of data, from what? >> Yeah. >> What kind of logs? What things are generating logs? What's the relevant information that's being. Kept track of? >> Yeah, I mean, so from our perspective, we're actually agnostic to data source. So we have an assist log integration. We have kind of basic API's. We have, you know, agents for any sort of operating system. Funny enough people actually use those agents to install, log DNA on robots, right? And so we have a customer they're, you know, one of the largest E-commerce platforms on, in the, in the world and they have a warehouse robots division and they installed the agent on every single one of those robots. They're, you know, they're running like arm 64 processors and they will send the log data directly to us. Right? So to us, it's no different. A robot is no different from a server is no different from an application is no different from a router. We take in all that data. Traditionally though, to answer your question, I guess, in the simplest way, mostly applications, servers, firewalls, all the traditional stuff you'd expect kind of going into a log platform. >> So you mentioned a big name customer. I've got a guess as to who that is. I won't, I won't say, but talk to us about the observability pipeline. What is that? What are the benefits in it for customers? >> (Peter) Sure. So, like if we zoom out again, you know, you think about logs traditionally. I think a lot of folks say, okay, we'll ingest the logs. We'll analyze them. What we noticed is that there's a lot of value in the step before that. So I think in the earlier days it was really novel to say, Hey, we're going to get logs and we're going to put it into a system. We're going to analyze it. We're going to centralize. Right. And that had its merits. But I think over time it got a little chaotic. And so you saw a lot of the vendors over the last three years consolidating and doing more of a single pane of glass, all the pillars of observability and whatnot. But then the downside of that is you're seeing a lot of the teams that are using that then saying being constrained by single vendor for all the ways that you can access that data. So we decided that the control point being on the analysis side on, on the very far right side was constricting. So we said, okay, let's move the control point up into a pipeline where the logs are coming to a single point of ingress. And then what we'll do is we will offer views, but also allow you to stream into other systems. So we'll allow you to stream into like a SIM or a data warehouse or something, something like that. Right? So, and you know, we're still trying to like nail down the messaging. I'm sure our marketing person's going to roast me after this. But the simplest way to think of observability pipeline is it's the step before the analysis part, that kind of ingest processes and routes the data. >> (Dave) Yeah. This is the Kube, by the way, neither one of us is a weather reporter. (laughing) So, so the technical stuff is good with us. >> Yes. It is. What are, and talk to us about some of the key features and capabilities and maybe anything that's newly announced are going to be announced. >> Yeah. For sure. So what we recently announced early access on is our streaming capabilities. So it's something that we built in conjunction with IBM and with a couple of, you know, large major institutions that we were working with on the IBM cloud. And basically we realized as we were ingesting a log data, some of those consumers wanted to access subsets of that data and other systems such as Q radar or, you know, a security product. So we ended up taking, we filtered down a subset of that data and we stream it out into those systems. And so we're taking those capabilities and then bringing it into our direct product, you know, whatever you access via logging.com. That is what's essentially going to be the seed for the kind of observability pipeline moving forward. So when you start thinking about it, all of this stuff that I mentioned, where we say, we're focusing on control, like allowing you to exclude logs, allowing you to transform logs, you take those processing capabilities, you take the streaming capabilities, you put them together and all of a sudden that's the pipeline, right? So that's the biggest focus for us now. And then we also have supporting features such as, you know, control API's. We have index rate alerting so that you can get notified if you see aberrations in the amount of flow of data. We have things like variable retention. So when a certain subset of logs come in, if you want it store it for seven days or 30 days, you can go ahead and do that because we know that a large block of logs is going to have many different use cases and many different associated values, right? >> So let's pretend for a moment that a user, somebody who has spent their money on log DNA is putting together a Yelp review and they've given you five stars. >> Yup. >> What do they say about log DNA? Why did they give you that five star rating? >> Yeah. Absolutely. I think, you know, the most common one and it's funny it's Yelp because we actually religiously mine, our G2 crowd reviews. (all laughing) And so the thing that we hear most often is, it's ease of use, right? A lot of these tools. I mean, I'm sure, you know, you're talking to founders and product leaders every day with developers. Like the, the bar, the baseline is so low, you know, a lot of, a lot of organizations where like, we'll give them the, you know, their coders, they'll figure it out. We'll just give them docs and they'll figure it out. But we, we went a little bit extra in terms of like, how can we smooth that experience so that when you go to your computer and you type in QTPL, blah, blah, blah, two lines, and all of a sudden all your logs are shipping from your cluster to log DNA. So that's the constant theme for us in all of our views is, Hey, I showed up, I signed up and within 30 minutes I had everything going that I needed to get. >> (Lisa) So fast time to value. >> Yes. >> Which is critical these days. >> Absolutely. >> Talk to me. So here we are at, at KubeCon, the CNCF community is huge. I think I, the number I saw yesterday was 138,000 contributors. Lots of activity, because we're in person, which is great. We can have those hallway networking conversations that we haven't been able to have in a year and a half. What are some of the things that you guys have heard at the booth in terms of being able to engage with the community again? >> You know, the thing that we've heard most often is just like having a finger on the pulse. It's so hard to do that because you know, when we're all at our computers, we just go from zoom to zoom. And so it, it like, unless it punches you in the face, you're not aware of it. Right. But when you come here, you look around, you go, you can start to identify trends, you hear the conversations in the hallway, you see the sessions. It's just that, that sense of, it's almost like a Phantom limb that, that sense of community and being kind of connected. I think that's the thing that we've heard most often that people are excited. And, you know, I think a lot of us are just kind of treating this like a dry run. Like we're kind of easing our way back in. And so it, you know, it felt good to be back. >> Well, they've done a great job here, right? I mean, you have to show your proof of vaccination. They're doing temperature checks, or you can show your clear health pass. So they're making it. We were talking to the executive director of CNCF earlier today and you're making it, it's not rocket science. We have enough data to know that this can be done carefully and safely. >> (David) Don't forget the wristbands. >> That's right. And, and did you see the wristbands? >> (Peter) Oh yeah. >> Yeah, yeah that's great. >> Yep, it is great. >> I was, I was on the fence by the way. I was like, I was a green or yellow, depending on the person. >> (both) Yeah. >> Yeah. But giving, giving everybody the opportunity to socialize again and to have those, those conversations that you just can't have by zoom, because you have somebody you've seen someone and it jogs your memory and also the control of do I want to shake someone's hand or do I not. They've done a great job. And I think hopefully this is a good test in the water for others, other organizations to learn. This can be done safely because of the community. You can't replicate that on video. >> (Peter) Absolutely. And I'll tell you this one for us, this is our, this is our event. This is the event for us every single year. We, we it's the only event we care about at the end of the day. So. >> What are some of the things that you've seen in the last year, in terms of where, we were talking a lot about the, the adoption of Kubernetes, kind of, where is it in its maturation state, but we've seen so much acceleration and digital transformation in the last 18 months for every industry businesses rapidly pivoting multiple times to try to, to survive one and then figure out a new way to thrive in this, this new I'll call it the new. Now I'll borrow that from a friend at Citrix, the new now, not the new normal, the new now, what are some of the things that you've seen in the last year and a half from, from your customer base in terms of what have they been coming to you saying help? >> (Peter) You know, I think going back to the earlier point about time to value, that's the thing that a lot. So a lot of our customers are, you know, very big Kubernetes, you know, they're, they're big consumers of Kubernetes. I would say, you know, for me, when I do the, we do our, our QBRs with our top customers, I would say 80% of them are huge Kubernetes shops. Right. And the biggest bottleneck for them actually is onboarding new engineers because a lot of the, and you know, we have a customer, we have better mortgage. We have, IBM, we have Rappi is a customer of ours. They're like Latin American version of Instacart. They double their engineering base and you, you know, like 18 in 18 months. And so that's, you know, I think it was maybe from 1500 to 3000 developers or so, so their thing is like, we need to get people on board as soon as possible. We need to get them in these tools, getting access to, to, to their longs, to whatever they need. And so that's been the biggest thing that we've heard over and over again is A, how can we hire? And then B when we hire them, how do we onboard them as quickly as possible, so that they're ramped up and they're adding value. >> How do you help with that onboarding, making it faster, seamless so that they can get value faster? >> So for us, you know, we really lean in on our, our customer success teams. So they do, you know, they do trainings, they do best practices. Basically. We kind of think of ourselves given how much Kubernetes contradiction we have, we think of ourselves as cross pollinators. So a lot of the times we'll go into those decks and we'll try to learn just as much as we're trying to try to teach. And then we'll go and repeat that process through every single set of our customers. So a lot of the patterns that we'll see are, well, you know, what kinds of tools are you using for orchestration? What kind of tools are you using for deployment? How are you thinking about X, Y, and Z? And then, you know, even our own SRE teams will kind of get into the mix and, you know, provide tips and feedback. >> (Lisa) Customer centricity is key. We've heard that a lot today. We hear that from a lot of companies. It's one thing to hear it. It's another thing to see it. And it sounds like the Yelp review that you would have given, or, or what you're hearing through G2 crowd. I mean, that voice of the customer is valid. That's, that's the only validation. I think that really matters because analysts are paid. >> Yeah. >> But hearing that validation through the voice of the customer consistently lets you know, we're going in the right direction here. >> Absolutely. >> I think it's, it's interesting that ease of use comes up. You wonder if those are only anonymous reviews, you don't necessarily associate open source community with cutting edge, you know, we're the people on the pirate ship. >> (Peter) Yeah. And so when, when, when people start to finally admit, you know, some ease of use would be nice. I think that's an indication of maturity at a certain point. It's saying, okay, not everyone is going to come in and sit behind a keyboard and program things in machine language. Every time we want to do some simple tasks, let's automate, let's get some ease of use into this. >> And I'll tell you in the early days it drove me and our, our CEO talker. It drove us nuts that people would say easiest to be like, that's so shallow. It doesn't mean anything. Well, you know, all of that. However, but to your point, if we don't meet the use case, if we don't have the power behind it, the ease of use is abstracting away. It's like an iceberg, right. It's abstracting away a lot. So we can't even have the ease of use conversation unless we're able to meet the use case. So, so what we've been doing is digging into that more, be like, okay, ease of use, but what were you trying to do? What, what is it that we enabled? Because ease of use, if it's a very shallow set of use cases is not as valid as ease of use for petabytes of data for an organization like IBM. Right? >> That's a great, I'm glad that you dug into that because ease of use is one of those things that you'll see it in marketing materials, but to your point, you want to know what does this actually mean? What are we delivering? >> Right. >> And now, you know what you're delivering with Peter, thank you for sharing with us about logged in and what you guys are doing, how you're helping your community of customers and hearing the voice of the customer through G2 and others. Good work. >> Thank you. And by the way, I'll be remiss if I, if I don't say this, if you're interested in learning more about some of the stuff that we're working on, just go to logging in dot com. We've got, I think we've got a banner for the early access programs that I mentioned earlier. So, you know, at the end of the day, to your point about customer centricity, everything we prioritize is based on our customers, what they need, what they tell us about. And so, you know, whatever engagement that we get from the people at the show and prospects, like that's how we drive a roadmap. >> (Lisa) Yup. That's why we're all here. Log dna.com. Peter, thank you for joining Dave and me today. We appreciate it. >> Thanks for having me. >> Our pleasure for Dave Nicholson. I'm Lisa Martin signing off from Los Angeles today. The Kubes coverage of KubeCon clouding of con 21 continues tomorrow. We'll see then. (soft techno music)
SUMMARY :
you to a new company, What do you do? And then, you know, I'll What kind of logs? We have, you know, So you mentioned a big name customer. So, and you know, we're So, so the technical some of the key features and so that you can get notified they've given you five stars. experience so that when you go to that you guys have heard It's so hard to do that because you know, I mean, you have to show did you see the wristbands? depending on the person. that you just can't have I'll tell you this one for us, coming to you saying help? lot of the, and you know, So for us, you know, review that you would have customer consistently lets you know, cutting edge, you know, you know, some ease of use would be nice. Well, you know, all of that. And now, you know what And so, you know, Peter, thank you for The Kubes coverage of KubeCon
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Nicholson | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Peter | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Peter Choi | PERSON | 0.99+ |
seven days | QUANTITY | 0.99+ |
Dave Nicholson | PERSON | 0.99+ |
Citrix | ORGANIZATION | 0.99+ |
five star | QUANTITY | 0.99+ |
30 days | QUANTITY | 0.99+ |
five stars | QUANTITY | 0.99+ |
Los Angeles | LOCATION | 0.99+ |
David | PERSON | 0.99+ |
18 | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
138,000 contributors | QUANTITY | 0.99+ |
Peter Cho | PERSON | 0.99+ |
CNCF | ORGANIZATION | 0.99+ |
80% | QUANTITY | 0.99+ |
yesterday | DATE | 0.99+ |
last year | DATE | 0.99+ |
Lisa | PERSON | 0.99+ |
KubeCon | EVENT | 0.99+ |
18 months | QUANTITY | 0.99+ |
tomorrow | DATE | 0.99+ |
last year | DATE | 0.99+ |
1500 | QUANTITY | 0.98+ |
both | QUANTITY | 0.98+ |
two lines | QUANTITY | 0.98+ |
CloudNativeCon | EVENT | 0.98+ |
two X | QUANTITY | 0.98+ |
Kubernetes | ORGANIZATION | 0.98+ |
a year and a half | QUANTITY | 0.97+ |
one | QUANTITY | 0.96+ |
Latin American | OTHER | 0.96+ |
Yelp | ORGANIZATION | 0.95+ |
pandemic | EVENT | 0.95+ |
3000 developers | QUANTITY | 0.95+ |
single vendor | QUANTITY | 0.94+ |
G2 | ORGANIZATION | 0.94+ |
last 18 months | DATE | 0.93+ |
Kube | ORGANIZATION | 0.92+ |
con 21 | EVENT | 0.91+ |
Kubernetes | TITLE | 0.91+ |
single point | QUANTITY | 0.91+ |
single pane | QUANTITY | 0.91+ |
last year and | DATE | 0.88+ |
single | QUANTITY | 0.87+ |
earlier today | DATE | 0.86+ |
last three years | DATE | 0.86+ |
30 minutes | QUANTITY | 0.86+ |
KubeCon Cloud Native Con 2021 | EVENT | 0.84+ |
logging.com | OTHER | 0.82+ |
one thing | QUANTITY | 0.77+ |
single set | QUANTITY | 0.72+ |
NA 2021 | EVENT | 0.7+ |
Log dna.com | OTHER | 0.69+ |
every single year | QUANTITY | 0.68+ |
Rappi | PERSON | 0.68+ |
double | QUANTITY | 0.66+ |
arm 64 | OTHER | 0.59+ |
half | DATE | 0.55+ |
QTPL | TITLE | 0.54+ |
SRE | ORGANIZATION | 0.53+ |
Instacart | TITLE | 0.51+ |
Kubes | PERSON | 0.37+ |
Dipti Borkar, Ahana, and Derrick Harcey, Securonix | CUBE Conversation, July 2021
(upbeat music) >> Welcome to theCUBE Conversation. I'm John Furrier, host of theCUBE here in Palo Alto, California, in our studios. We've got a great conversation around open data link analytics on AWS, two great companies, Ahana and Securonix. Dipti Borkar, Co-founder and Chief Product Officer at Ahana's here. Great to see you, and Derrick Harcey, Chief Architect at Securonix. Thanks for coming on, really appreciate you guys spending the time. >> Yeah, thanks so much, John. Thank you for having us and Derrick, hello again. (laughing) >> Hello, Dipti. >> We had a great conversation around our startup showcase, which you guys were featured last month this year, 2021. The conversation continues and a lot of people are interested in this idea of open systems, open source. Obviously open data lakes is really driving a lot of value, especially with machine learning and whatnot. So this is a key, key point. So can you guys just take a step back before we get under the hood and set the table on Securonix and Ahana? What's the big play here? What is the value proposition? >> Why sure, I'll give a quick update. Securonix has been in the security business. First, a user and entity, behavioral analytics, and then the next generation SIEM platform for 10 years now. And we really need to take advantage of some cutting edge technologies in the open source community and drive adoption and momentum that we can not only bring in data from our customers, that they can find security threats, but also store in a way that they can use for other purposes within their organization. That's where the open data lake is very critical. >> Yeah and to add on to that, John, what we've seen, you know, traditionally we've had data warehouses, right? We've had operational systems move all of their data into the warehouse and those, you know, while these systems are really good, built for good use cases, the amount of data is exploding, the types of data is exploding, different types, semi-structured, structured and so when, as companies like Securonix in the security space, as well as other verticals, look for getting more insights out of their data, there's a new approach that's emerging where you have a data lake, which AWS has revolutionized with S3 and commoditized and there's analytics that's built on top of it. And so we're seeing a lot of good advantages that come out of this new approach. >> Well, it's interesting EC2 and S3 are having their 15th birthday, as they say in Amazon's interesting teenage years, but while I got you guys here, I want to just ask you, can you define the SIEM thing because the SIEM market is exploding, it just changed a little bit. Obviously it's data, event management, but again, as data becomes more proliferating, and it's not stopping anytime soon, as cloud native applications emerge, why is this important? What is this SIEM category? What's it about? >> Yeah, thanks. I'll take that. So obviously SIEM traditionally has been around for about a couple of decades and it really started with first log collection and management and rule-based threat detection. Now what we call next generation SIEM is really the modernization of a security platform that includes streaming threat detection and behavioral analysis and data analytics. We literally look for thousands of different threat detection techniques, and we chained together sequences of events and we stream everything in real time and it's very important to find threats as quickly as possible. But the momentum that we see in the industry as we see massive sizes of customers, we have made a transition from on-premise to the cloud and we literally are processing tens of petabytes of data for our customers. And it's critical that we can adjust data quickly, find threats quickly and allow customers to have the tools to respond to those security incidents quickly and really get the handle on their security posture. >> Derrick, if I ask you what's different about this next gen SIEM, what would you say and what's the big a-ha? What's the moment there? What's the key thing? >> The real key is taking the off the boundaries of scale. We want to be able to ingest massive quantities of data. We want to be able to do instant threat detection, and we want to be able to search on the entire forensic data set across all of the history of our customer base. In the past, we had to make sacrifices, either on the amount of data we ingest or the amount of time that we stored that data. And the really the next generation SIEM platform is offering advanced capabilities on top of that data set because those boundaries are no longer barriers for us. >> Dipti, any comment before I jump into the question for you? >> Yeah, you know, absolutely. It is about scale and like I mentioned earlier, the amount of data is only increasing and it's also the types of information. So the systems that were built to process this information in the past are, you know, support maybe terabytes of data, right? And that's where new technologies open source engines like Presto come in, which were built to handle internet scale. Presto was kind of created at Facebook to handle these petabytes that Derrick is talking about that every industry is now seeing where we're are moving from gigs to terabytes to petabytes. And that's where the analytic stack is moving. >> That's a great segue. I want to ask you while I got you here 'cause this is again, the definitions, 'cause people love to hear the experts weigh in. What is open data lake analytics? How would you define that? And then talk about where Presto fits in. >> Yeah, that's a great question. So the way I define open data lake analytics is you have a data lake on the core, which is, let's say S3, it's the most popular one, but on top of it, there are open aspects, it is open format. Open formats play a very important role because you can have different types of processing. It could be SQL processing, it could be machine learning, it could be other types of workloads, all work on these open formats versus a proprietary format where it's locked and it's open interfaces. Open interfaces that are like SQL, JDBC, ODBC is widely accessible to a range of tools. And so it's everywhere. Open source is a very important part of it. As companies like Securonix pick these technologies for their mission critical systems, they want to know that this is going to be available and open for them for a long period of time. And that's why open source becomes important. And then finally, I would say open cloud because at the end of the day, you know, while AWS is where a lot of the innovations happening, a lot of the market is, there are other clouds and open cloud is something that these engines were built for, right? So that's how I define open data lake analytics. It's analytics with query engines built on top of these open formats, open source, open interfaces and open cloud. Now Presto comes in where you want to find the needle in the haystack, right? And so when you have these deep questions about where did the threat come from or who was it, right? You have to ask these questions of your data. And Presto is an open source distributed SQL engine that allows data platform teams to run queries on their data lakes in a high-performance ways, in memory and on these petabytes of data. So that's where Presto fits in. It's one of the defacto query engines for SQL analysis on the data lake. So hopefully that answers the question, gives more context. >> Yeah, I mean, the joke about data lakes has been you don't want to be a data swamp, right? That's what people don't want. >> That's right. >> But at the same time, the needle in the haystack, it's like big data is like a needle in a haystack of needles. So there's a constant struggle to getting that data, the right data at the right time. And what I learned in the last presentation, you guys both presented, your teams presented at the conference was the managed service approach. Could you guys talk about why that approach works well together with you guys? Because I think when people get to the cloud, they replatform, then they start refactoring and data becomes a real big part of that. Why is the managed service the best approach to solving these problems? >> Yeah and interestingly, both Securonix and Ahana have a managed service approach so maybe Derrick can go first and I can go after. >> Yeah, yeah. I'll be happy to go first. You know, we really have found making the transition over the last decade from off premise to the cloud for the majority of our customers that running a large open data lake requires a lot of different skillsets and there's hundreds of technologies in the open source community to choose from and to be able to choose the right blend of skillsets and technologies to produce a comprehensive service is something that customers can do, many customers did do, and it takes a lot of resources and effort. So what we really want to be able to do is take and package up our security service, our next generation SIEM platform to our customers where they don't need to become experts in every aspect of it. Now, an underlying component of that for us is how we store data in an open standards way and how we access that data in an open standards way. So just like we want our customers to get immediate value from the security services that we provide, we also want to be able take advantage of a search service that is offered to us and supported by a vendor like Ahana where we can very quickly take advantage of that value within our core underlying platform. So we really want to be able to make a frictionless effort to allow our customers achieve value as quick as possible. >> That's great stuff. And on the Ahana side, open data lakes, really the ease of use there, it sounds easy to me, but we know it's not easy just to put data in a data lake. At the end of the day, a lot of customers want simplicity 'cause they don't have the staffing. This comes up a lot. How do you leverage their open source participation and/or getting stood up quickly so they can get some value? Because that seems to be the number one thing people want right now. Dipti, how does that work? How do people get value quickly? >> Yeah, absolutely. When you talk about these open source press engines like Presto and others, right? They came out of these large internet companies that have a lot of distributed systems, engineers, PhDs, very kind of advanced level teams. And they can manage these distributed systems building onto them, add features at large scale, but not every company can and these engines are extremely powerful. So when you combine the power of Presto with the cloud and a managed service, that's where value for everyone comes in. And that's what I did with Ahana is looked at Presto, which is a great engine, but converted it into a great user experience so that whether it's a three person platform team or a five person platform team, they still get the same benefit of Presto that a Facebook gets, but at much, much a less operational complexity cost, as well as the ability to depend on a vendor who can then drive the innovation and make it even better. And so that's where managed services really com in. There's thousands of credit parameters that need to be tuned. With Ahana, you get it out of the box. So you have the best practices that are followed at these larger companies. Our team comes from Facebook, HuBERT and others, and you get that out of the box, with a few clicks you can get up and running. And so you see value immediately, in 30 minutes you're up and running and you can create your data lake versus with Hadoop and these prior systems, it would take months to receive real value from some of these systems. >> Yeah, we saw the Hadoop scar tissue is all great and all good now, but it takes too much resource, standing up clusters, managing it, you can't hire enough people. I got to ask you while you're on that topic, do you guys ship templates? How do you solve the problem of out of the box? You mentioned some out of the box capability. Do you guys think of as recipes, templates? What's your thoughts around what you're providing customers to get up and running? >> Yeah so in the case of Securonix, right, let's say they want to create a Presto cluster. They go into our SAS console. You essentially put in the number of nodes that you want. Number of workers you want. There's a lot of additional value that we built in like caching capabilities if you want more performance, built in cataloging that's again, another single click. And there isn't really as much of a template. Everybody gets the best tuned Presto for their workloads. Now there are certain workloads where you might have interactive in some cases, or you might have transformation batch ETL, and what we're doing next is actually giving you the knobs so that it comes pre tuned for the type of workload that you want to run versus you figuring it out. And so that's what I mean by out of the box, where you don't have to worry about these configuration parameters. You get the performance. And maybe Derrick can you talk a little bit about the benefits of the managed service and the usage as well. >> Yeah, absolutely. So, I'll answer the same question and then I'll tie back to what Dipti asked. Really, you know, our customers, we want it to be very easy for them to ingest security event logs. And there's really hundreds of types of a security event logs that we support natively out of the box, but the key for us is a standard that we call the open event format. And that is a normalized schema. We take any data source in it's normalized format, be a collector device a customer uses on-premise, they send the data up to our cloud, we do streaming analysis and data analytics to determine where the threats are. And once we do that, then we send the data off to a long-term storage format in a standards-based Parquet file. And that Parquet file is natively read by the Ahana service. So we simply deploy an Ahana cluster that uses the Presto engine that natively supports our open standard file format. And we have a normalized schema that our application can immediately start to see value from. So we handle the collection and streaming ingest, and we simply leverage the engine in Ahana to give us the appropriate scale. We can size up and down and control the cost to give the users the experience that they're paying for. >> I really love this topic because one, not only is it cutting edge, but it's very relevant for modern applications. You mentioned next gen SIEMs, SIEM, security information event management, not SIM as memory card, which I think of all the time because I always want to add more, but this brings up the idea of streaming data real-time, but as more services go to the cloud, Derrick, if you don't mind sharing more on this. Share the journey that you guys gone through, because I think a lot of people are looking at the cloud and saying, and I've been in a lot of these conversations about repatriation versus cloud. People aren't going that way. They're going more innovation with his net new revenue models emerging from the value that they're getting out of understanding events that are happening within the network and the apps, even when they're being stood up and torn down. So there's a lot of cloud native action going on where just controlling and understanding is way beyond the, just put stuff into an event log. It's a whole nother animal. >> Well, there's a couple of paradigm shifts that we've seen major patterns for in the last five or six years. Like I said, we started with the safe streaming ingest platform on premise. We use some different open source technologies. What we've done when we moved to the cloud is we've adopted cloud native services as part of our underlying platform to modernize and make our service cloud native. But what we're seeing as many customers either want to focus on on-premise deployments and especially financial institutions and government institute things, because they are very risk averse. Now we're seeing even those customers are realizing that it's very difficult to maintain the hundreds or thousands of servers that it requires on premise and have the large skilled staff required to keep it running. So what we're seeing now is a lot of those customers deployed some packaged products like our own, and even our own customers are doing a mass migration to the cloud because everything is handled for them as a service. And we have a team of experts that we maintain to support all of our global customers, rather than every one of our global customers having their own teams that we then support on the back end. So it's a much more efficient model. And then the other major approach that many of our customers also went down the path of is, is building their own security data lake. And many customers were somewhat successful in building their own security data lake but in order to keep up with the innovation, if you look at the analyst groups, the Gartner Magic Quadrant on the SIEM space, the feature set that is provided by a packaged product is a very large feature set. And even if somebody was put together all of the open source technologies to meet 20% of those features, just maintaining that over time is very expensive and very difficult. So we want to provide a service that has all of the best in class features, but also leverages the ability to innovate on the backend without the customer knowing. So we can do a technology shift to Ahana and Presto from our previous technology set. The customer doesn't know the difference, but they see the value add within the service that we're offering. >> So if I get this right, Derrick, Presto's enabling you guys to do threat detection at a level that you're super happy with as well as giving you the option for give self-service. Is that right for the, is that a kind of a- >> Well, let me clarify our definition. So we do streaming threat detection. So we do a machine learning based behavioral analysis and threat detection on rule-based correlation as well. So we do threat detection during the streaming process, but as part of the process of managing cybersecurity, the customer has a team of security analysts that do threat hunting. And the threat hunting is where Ahana comes in. So a human gets involved and starts searches for the forensic logs to determine what happened over time that might be suspicious and they start to investigate through a series of queries to give them the information that's relevant. And once they find information that's relevant, then they package it up into an algorithm that will do a analysis on an ongoing basis as part of the stream processing. So it's really part of the life cycle of hunting a real time threat detection. >> It's kind of like old adage hunters and farmers, you're farming through the streaming and hunting with the detection. I got to ask you, what would it be the alternative if you go back, I mean, I know cloud's so great because you have cutting edge applications and technologies. Without Presto, where would you be? I mean, what would be life like without these capabilities? What would have to happen? >> Well, the issue is not that we had the same feature set before we moved to Presto, but the challenge was on scale. The cost profile to continue to grow from 100 terabytes to one petabyte, to tens of petabytes, not only was it expensive, but it just, the scaling factors were not linear. So not only did we have a problem with the costs, but we also had a problem with the performance tailing off and keeping the service running. A large Hadoop cluster, for example, our first incarnation of this use, the hive service, in order to query data in a MapReduce cluster. So it's a completely different technology that uses a distributed Hadoop compute cluster to do the query. It does work, but then we start to see resource contention with that, and all the other things in the Hadoop platform. The Presto engine has the beauty of it, not only was it designed for scale, but it's feature built just for a query engine and that's the providing the right tool for the job, as opposed to a general purpose tool. >> Derrick, you've got a very busy job as chief architect. What are you excited about going forward when you look at the cloud technologies? What are you looking at? What are you watching? What are you getting excited about or what worries you? >> Well, that's a good question. What we're really doing, I'm leading up a group called the Securonix Innovation Labs, and we're looking at next generation technologies. We go through and analyze both open source technologies, technologies that are proprietary as well as building own technologies. And that's where we came across Ahana as part of a comprehensive analysis of different search engines, because we wanted to go through another round of search engine modernization, and we worked together in a partnership, and we're going to market together as part of our modernization efforts that we're continuously going through. So I'm looking forward to iterative continuous improvement over time. And this next journey, what we're seeing because of the growth in cybersecurity, really requires new and innovative technologies to work together holistically. >> Dipti, you got a great company that you co-founded. I got to ask you as the co-founder and chief product officer, you both the lead entrepreneur also, got the keys to the kingdom with the products. You got to balance that 20 miles stare out in the future while driving product excellence. You've got open source as a tailwind. What's on your mind as you go forward with your venture? >> Yeah. Great question. It's been super exciting to have found the Ahana in this space, cloud data and open source. That's where the action is happening these days, but there's two parts to it. One is making our customers successful and continuously delivering capabilities, features, continuing on our ease of use theme and a foundation to get customers like Securonix and others to get most value out of their data and as fast as possible, right? So that's a continuum. In terms of the longer term innovation, the way I see the space, there is a lot more innovation to be done and Presto itself can be made even better and there's a next gen Presto that we're working on. And given that Presto is a part of the foundation, the Linux Foundation, a lot of this innovation is happening together collaboratively with Facebook, with Uber who are members of the foundation with us. Securonix, we look forward to making a part of that foundation. And that innovation together can then benefit the entire community as well as the customer base. This includes better performance with more capabilities built in, caching and many other different types of database innovations, as well as scaling, auto scaling and keeping up with this ease of use theme that we're building on. So very exciting to work together with all these companies, as well as Securonix who's been a fantastic partner. We work together, build features together, and I look at delivering those features and functionalities to be used by these analysts, data scientists and threat hunters as Derrick called them. >> Great success, great partnership. And I love the open innovation, open co-creation you guys are doing together and open data lakes, great concept, open data analytics as well. This is the future. Insights coming from the open and sharing and actually having some standards. I love this topic, so Dipti, thank you very much, and Derrick, thanks for coming on and sharing on this Cube Conversation. Thanks for coming on. >> Thank you so much, John. >> Thanks for having us. >> Thanks. Take care. Bye-bye. >> Okay, it's theCube Conversation here in Palo Alto, California. I'm John furrier, your host of theCube. Thanks for watching. (upbeat music)
SUMMARY :
guys spending the time. and Derrick, hello again. and set the table on Securonix and Ahana? and momentum that we can into the warehouse and those, you know, because the SIEM market is exploding, and really get the handle either on the amount of data we ingest and it's also the types of information. hear the experts weigh in. So hopefully that answers the Yeah, I mean, the joke Why is the managed Yeah and interestingly, a search service that is offered to us And on the Ahana side, open data lakes, and you get that out of the box, I got to ask you while and the usage as well. and control the cost from the value that they're getting and have the large skilled staff as well as giving you the for the forensic logs to and hunting with the detection. and that's the providing when you look at the cloud technologies? because of the growth in cybersecurity, got the keys to the and a foundation to get And I love the open here in Palo Alto, California.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Securonix | ORGANIZATION | 0.99+ |
John | PERSON | 0.99+ |
Derrick Harcey | PERSON | 0.99+ |
Derrick | PERSON | 0.99+ |
ORGANIZATION | 0.99+ | |
Ahana | ORGANIZATION | 0.99+ |
Ahana | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
20% | QUANTITY | 0.99+ |
July 2021 | DATE | 0.99+ |
Uber | ORGANIZATION | 0.99+ |
Dipti | PERSON | 0.99+ |
100 terabytes | QUANTITY | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
10 years | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
hundreds | QUANTITY | 0.99+ |
Linux Foundation | ORGANIZATION | 0.99+ |
two parts | QUANTITY | 0.99+ |
thousands | QUANTITY | 0.99+ |
Securonix Innovation Labs | ORGANIZATION | 0.99+ |
tens of petabytes | QUANTITY | 0.99+ |
30 minutes | QUANTITY | 0.99+ |
one petabyte | QUANTITY | 0.99+ |
Dipti Borkar | PERSON | 0.99+ |
20 miles | QUANTITY | 0.99+ |
Palo Alto, California | LOCATION | 0.99+ |
five person | QUANTITY | 0.99+ |
First | QUANTITY | 0.99+ |
SQL | TITLE | 0.99+ |
last month | DATE | 0.99+ |
both | QUANTITY | 0.99+ |
One | QUANTITY | 0.98+ |
15th birthday | QUANTITY | 0.97+ |
two great companies | QUANTITY | 0.96+ |
HuBERT | ORGANIZATION | 0.96+ |
Hadoop | TITLE | 0.96+ |
S3 | TITLE | 0.96+ |
hundreds of technologies | QUANTITY | 0.96+ |
three person | QUANTITY | 0.95+ |
Parquet | TITLE | 0.94+ |
first incarnation | QUANTITY | 0.94+ |
first | QUANTITY | 0.94+ |
Presto | ORGANIZATION | 0.93+ |
Gartner | ORGANIZATION | 0.93+ |
last decade | DATE | 0.92+ |
terabytes of data | QUANTITY | 0.92+ |
first log | QUANTITY | 0.91+ |
single click | QUANTITY | 0.9+ |
Presto | PERSON | 0.9+ |
theCUBE | ORGANIZATION | 0.88+ |
Ariel Assaraf, Coralogix | AWS Startup Showcase: The Next Big Thing in AI, Security, & Life Sciences
(upbeat music) >> Hello and welcome today's session for the AWS Startup Showcase, the next big thing in AI, Security and Life Sciences featuring Coralogix for the AI track. I'm your host, John Furrier with theCUBE. We're here we're joined by Ariel Assaraf, CEO of Coralogix. Ariel, great to see you calling in from remotely, videoing in from Tel Aviv. Thanks for coming on theCUBE. >> Thank you very much, John. Great to be here. >> So you guys are features a hot next thing, start next big thing startup. And one of the things that you guys do we've been covering for many years is, you're into the log analytics, from a data perspective, you guys decouple the analytics from the storage. This is a unique thing. Tell us about it. What's the story? >> Yeah. So what we've seen in the market is that probably because of the great job that a lot of the earlier generation products have done, more and more companies see the value in log data, what used to be like a couple rows, that you add, whenever you have something very important to say, became a standard to document all communication between different components, infrastructure, network, monitoring, and the application layer, of course. And what happens is that data grows extremely fast, all data grows fast, but log data grows even faster. What we always say is that for sure data grows faster than revenue. So as fast as a company grows, its data is going to outpace that. And so we found ourselves thinking, how can we help companies be able to still get the full coverage they want without cherry picking data or deciding exactly what they want to monitor and what they're taking risk with. But still give them the real time analysis that they need to make sure that they get the full insight suite for the entire data, wherever it comes from. And that's why we decided to decouple the analytics layer from storage. So instead of ingesting the data, then indexing and storing it, and then analyzing the stored data, we analyze everything, and then we only store it matters. So we go from the insights backwards. That allowed us to reduce the amount of data, reduce the digital exhaust that it creates, and also provide better insights. So the idea is that as this world of data scales, the need for real time streaming analytics is going to increase. >> So what's interesting is we've seen this decoupling with storage and compute be a great success formula and cloud scale, for instance, that's a known best practice. You're taking a little bit different. I love how you're coming backwards from it, you're working backwards from the insights, almost doing some intelligence on the front end of the data, probably sees a lot of storage costs. But I want to get specifically back to this real time. How do you do that? And how did you come up with this? What's the vision? How did you guys come up with the idea? What was the magic light bulb that went off for Coralogix? >> Yes, the Coralogix story is very interesting. Actually, it was no light bulb, it was a road of pain for years and years, we started by just you know, doing the same, maybe faster, a couple more features. And it didn't work out too well. The first few years, the company were not very successful. And we've grown tremendously in the past three years, almost 100X, since we've launched this, and it came from a pain. So once we started scaling, we saw that the side effects of accessing the storage for analytics, the latency it creates, the the dependency on schema, the price that it poses on our customers became unbearable. And then we started thinking, so okay, how do we get the same level of insights, because there's this perception in the world of storage. And now it started to happen in analytics, also, that talks about tiers. So you want to get a great experience, you pay a lot, you want to get a less than great experience, you pay less, it's a lower tier. And we decided that we're looking for a way to give the same level of real time analytics and the same level of insights. Only without the issue of dependencies, decoupling all the storage schema issues and latency. And we built our real time pipeline, we call it Streama. Streama is a Coralogix real time analysis platform that analyzes everything in real time, also the stateful thing. So stateless analytics in real time is something that's been done in the past and it always worked well. The issue is, how do you give a stateful insight on data that you analyze in real time without storing and I'll explain how can you tell that a certain issue happened that did not happen in the past three months if you did not store the past three months? Or how can you tell that behavior is abnormal if you did not store what's normal, you did not store to state. So we created what we call the state store that holds the state of the system, the state of data, were a snapshot on that state for the entire history. And then instead of our state being the storage, so you know, you asked me, how is this compared to last week? Instead of me going to the storage and compare last week, I go to the state store, and you know, like a record bag, I just scroll fast, I find out one piece of state. And I say, okay, this is how it looked like last week, compared to this week, it changed in ABC. And once we started doing that we on boarded more and more services to that model. And our customers came in and say, hey, you're doing everything in real time. We don't need more than that. Yeah, like a very small portion of data, we actually need to store and frequently search, how about you guys fit into our use cases, and not just sell on quota? And we decided to basically allow our customers to choose what is the use case that they have, and route the data through different use cases. And then each log records, each log record stops at the relevant stops in our data pipeline based on the use case. So just like you wouldn't walk into the supermarket, you fill in a bag, you go out, they weigh it and they say, you know, it's two kilograms, you pay this amount, because different products have different costs and different meaning to you. That same way, exactly, We analyze the data in real time. So we know the importance of data, and we allow you to route it based on your use case and pay a different amount per use case. >> So this is really interesting. So essentially, you guys, essentially capture insights and store those, you call them states, and then not have to go through the data. So it's like you're eliminating the old problem of, you know, going back to the index and recovering the data to get the insights, did we have that? So anyway, it's a round trip query, if you will, you guys are start saving all that data mining cost and time. >> We call it node zero side effects, that round trip that you that you described is exactly it, no side effects to an analysis that is done in real time. I don't need to get the latency from the storage, a bit of latency from the database that holds the model, a bit of latency from the cache, everything stays in memory, everything stays in stream. >> And so basically, it's like the definition of insanity, doing the same thing over and over again and expecting a different result. Here, that's kind of what that is, the old model of insight is go query the database and get something back, you're actually doing the real time filtering on the front end, capturing the insights, if you will, storing those and replicating that as use case. Is that right? >> Exactly. But then, you know, there's still the issue of customer saying, yeah, but I need that data. Someday, I need to really frequently search, I don't know, you know, the unknown unknowns, or some of the day I need for compliance, and I need an immutable record that stays in my compliance bucket forever. So we allowed customers, we have this some that screen, we call the TCO optimizer, that allows them to define those use cases. And they can always access the data by creating their remote storage from Coralogix, or carrying the hot data that is stored with Coralogix. So it's all about use cases. And it's all about how you consume the data because it doesn't make sense for me to pay the same amount or give the same amount of attention to a record that is completely useless. It's just there for the record or for a compliance audit, that may or may not happen in the future. And, you know, do the same with the most critical exception in my application log that has immediate business impact. >> What's really good too, is you can actually set some policy up if you want a certain use cases, okay, store that data. So it's not to say you don't want to store it, but you might want to store it on certain use cases. So I can see that. So I got to ask the question. So how does this differ from the competition? How do you guys compete? Take us through a use case of a customer? How do you guys go to the customer and you just say, hey, we got so much scar tissue from this, we learned the hard way, take it from us? How does it go? Take us through an example. >> So an interesting example of actually a company that is not the your typical early adopter, let's call it this way. A very advanced in technology and smart company, but a huge one, one of the largest telecommunications company in India. And they were actually cherry picking about 100 gigs of data per day, and sending it to one of the legacy providers which has a great solution that does give value. But they weren't even thinking about sending their entire data set because of cost because of scale, because of, you know, just a clutter. Whenever you search, you have to sift through millions of records that many of them are not that important. And we help them actually ask analyze their data and work with them to understand these guys had over a terabyte of data that had incredible insights, it was like a goldmine of insights. But now you just needed to prioritize it by their use case, and they went from 100 gig with the other legacy solution to a terabyte, at almost the same cost, with more advanced insights within one week, which isn't in that scale of an organization is something that is is out of the ordinary, took them four months to implement the other product. But now, when you go from the insights backwards, you understand your data before you have to store it, you understand the data before you have to analyze it, or before you have to manually sift through it. So if you ask about the difference, it's all about the architecture. We analyze and only then index instead of indexing and then analyzing. It sounds simple. But of course, when you look at this stateful analytics, it's a lot more, a lot more complex. >> Take me through your growth story, because first of all, I'll get back to the secret sauce in the same way. I want to get back to how you guys got here. (indistinct) you had this problem? You kind of broke through, you hit the magic formula, talking about the growth? Where's the growth coming from? And what's the real impact? What's the situation relative to the company's growth? >> Yeah, so we had a first rough three years that I kind of mentioned, and then I was not the CEO at the beginning, I'm one of the co founders. I'm more of the technical guy, was the product manager. And I became CEO after the company was kind of on the verge of closing at the end of 2017. And the CTO left the CEO left, the VP of R&D became the CTO, I became the CEO, we were five people with $200,000 in the bank that you know, you know that that's not a long runway. And we kind of changed attitudes. So we kind of, so we first we launched this product, and then we understood that we need to go bottoms up, you can go to enterprises and try to sell something that is out of the ordinary, or that changes how they're used to working or just, you know, sell something, (indistinct) five people will do under $1,000 in the bank. So we started going from bottoms up, and the earlier adopters. And it's still until today, you know, the the more advanced companies, the more advanced teams. This is our Gartner friend Coralogix, the preferred solution for Advanced, DevOps and Platform Teams. So they started adopting Coralogix, and then it grew to the larger organization, and they were actually pushing, there are champions within their organizations. And ever since. So until the beginning of 2018, we raised about $2 million and had sales or marginal. Today, we have over 1500, pink accounts, and we raised almost $100 million more. >> Wow, what a great pivot. That was great example of kind of getting the right wave here, cloud wave. You said in terms of customers, you had the DevOps kind of (indistinct) initially. And now you said expanded out to a lot more traditional enterprise, you can take me through the customer profile. >> Yeah, so I'd say it's still the core would be cloud native and (indistinct) companies. These are typical ones, we have very tight integration with AWS, all the services, all the integrations required, we know how to read and write back to the different services and analysis platforms in AWS. Also for Asia and GCP, but mostly AWS. And then we do have quite a few big enterprise accounts, actually, five of the largest 50 companies in the world use Coralogix today. And it grew from those DevOps and platform evangelists into the level of IT, execs and even (indistinct). So today, we have our security product that already sells to some of the biggest companies in the world, it's a different profile. And the idea for us is that, you know, once you solve that issue of too much data, too expensive, not proactive enough, too couple with the storage, you can actually expand that from observability logging metrics, now into tracing and then into security and maybe even to other fields, where the cost and the productivity are an issue for many companies. >> So let me ask you this question, then Ariel, if you don't mind. So if a customer has a need for Coralogix, is it because the data fall? Or they just got data kind of sprawled all over the place? Or is it that storage costs are going up on S3 or what's some of the signaling that you would see, that would be like, telling you, okay, okay, what's the opportunity to come in and either clean house or fix the mess or whatnot, Take us through what you see. What do you see is the trend? >> Yeah. So like the tip customer (indistinct) Coralogix will be someone using one of the legacy solution and growing very fast. That's the easiest way for us to know. >> What grows fast? The storage, the storage is growing fast? >> The company is growing fast. >> Okay. And you remember, the data grows faster than revenue. And we know that. So if I see a company that grew from, you know, 50 people to 500, in three years, specifically, if it's cloud native or internet company, I know that their data grew not 10X, but 100X. So I know that that company that might started with a legacy solution at like, you know, $1,000 a month, and they're happy with it. And you know, for $1,000 a month, if you don't have a lot of data, those legacy solutions, you know, they'll do the trick. But now I know that they're going to get asked to pay 50, 60, $70,000 a month. And this is exactly where we kick in. Because now, when it doesn't fit the economic model, when it doesn't fit the unit economics, and he started damaging the margins of those companies. Because remember, those internet and cloud companies, it's not costs are not the classic costs that you'll see in an enterprise, they're actually damaging your unit economics and the valuation of the business, the bigger deal. So now, when I see that type of organization, we come in and say, hey, better coverage, more advanced analytics, easier integration within your organization, we support all the common open source syntaxes, and dashboards, you can plug it into your entire environment, and the costs are going to be a quarter of whatever you're paying today. So once they see that they see, you know, the Dev friendliness of the product, the ease of scale, the stability of the product, it makes a lot more sense for them to engage in a PLC, because at the end of the day, if you don't prove value, you know, you can come with 90% discount, it doesn't do anything, not to prove the value to them. So it's a great door opener. But from then on, you know, it's a PLC like any other. >> Cloud is all about the PLC or pilot, as they say. So take me through the product, today, and what's next for the product, take us through the vision of the product and the product strategy. >> Yeah, so today, the product allows you to send any log data, metric data or security information, analyze it a million ways, we have one of the most extensive alerting mechanism to market, automatic anomaly detection, data flustering. And all the real law, you know, the real time pipeline, things that help companies make their data smarter, and more readable, parsing, enriching, getting external sources to enrich the data, and so on, so forth. Where we're stepping in now is actually to make the final step of decoupling the analytics from storage, what we call the datalist data platform in which no data will sit or reside within the Coralogix cloud, everything will be analyzed in real time, stored in a storage of choice of our customers, then we'll allow our customers to remotely query that incredible performance. So that'll bring our customers away, to have the first ever true SaaS experience for observability. Think about no quota plans, no retention, you send whatever you want, you pay only for what you send, you retain it, how long you want to retain it, and you get all the real time insights much, much faster than any other product that keeps it on a hot storage. So that'll be our next step to really make sure that, you know, we're kind of not reselling cloud storage, because a lot of the times when you are dependent on storage, and you know, we're a cloud company, like I mentioned, you got to keep your unit economics. So what do you do? You sell storage to the customer, you add your markup, and then you you charge for it. And this is exactly where we don't want to be. We want to sell the intelligence and the insights and the real time analysis that we know how to do and let the customers enjoy the, you know, the wealth of opportunities and choices their cloud providers offer for storage. >> That's great vision in a way, the hyper scalars early days showed that decoupling compute from storage, which I mentioned earlier, was a huge category creation. Here, you're doing it for data. We call hyper data scale, or like, maybe there's got to be a name for this. What do you see, about five years from now? Take us through the trajectory of the next five years, because certainly observability is not going away. I mean, it's data management, monitoring, real time, asynchronous, synchronous, linear, all the stuffs happening, what's the what's the five year vision? >> Now add security and observability, which is something we started preaching for, because no one can say I have observability to my environment when people you know, come in and out and steal data. That's no observability. But the thing is that because data grows exponentially, because it grows faster than revenue what we believe is that in five years, there's not going to be a choice, everyone are going to have to analyze the data in real time. Extract the insights and then decide whether to store it on a you know long term archive or not, or not store it at all. You still want to get the full coverage and insights. But you know, when you think about observability, unlike many other things, the more data you have many times, the less observability you get. So you think of log data unlike statistics, if my system was only in recording everything was only generating 10 records a day, I have full, incredible observability I know everything that I've done. what happens is that you pay more, you get less observability, and more uncertainty. So I think that you know, with time, we'll start seeing more and more real time streaming analytics, and a lot less storage based and index based solutions. >> You know, Ariel, I've always been saying to Dave Vellante on theCUBE, many times that there needs to be insights as to be the norm, not the exception, where, and then ultimately, it would be a database of insights. I mean, at the end of the day, the insights become more plentiful. You have the ability to actually store those insights, and refresh them and challenge them and model update them, verify them, either sunset them or add to them or you know, saying that's like, when you start getting more data into your organization, AI and machine learning prove that pattern recognition works. So why not grab those insights? >> And use them as your baseline to know what's important, and not have to start by putting everything in a bucket. >> So we're going to have new categories like insight, first, software (indistinct) >> Go from insights backwards, that'll be my tagline, if I have to, but I'm a terrible marketing (indistinct). >> Yeah, well, I mean, everyone's like cloud, first data, data is data driven, insight driven, what you're basically doing is you're moving into the world of insights driven analytics, really, as a way to kind of bring that forward. So congratulations. Great story. I love the pivot love how you guys entrepreneurially put it all together and had the problem your own problem and brought it out and to the to the rest of the world. And certainly DevOps in the cloud scale wave is just getting bigger and bigger and taking over the enterprise. So great stuff. Real quick while you're here. Give a quick plug for the company. What you guys are up to, stats, vitals, hiring, what's new, give the commercial. >> Yeah, so like mentioned over 1500 being customers growing incredibly in the past 24 months, hiring, almost doubling the company in the next few months. offices in Israel, East Center, West US, and UK and Mumbai. Looking for talented engineers to join the journey and build the next generation of data lists data platforms. >> Ariel Assaraf, CEO of Coralogix. Great to have you on theCUBE and thank you for participating in the AI track for our next big thing in the Startup Showcase. Thanks for coming on. >> Thank you very much John, really enjoyed it. >> Okay, I'm John Furrier with theCUBE. Thank you for watching the AWS Startup Showcase presented by theCUBE. (calm music)
SUMMARY :
Ariel, great to see you Thank you very much, John. And one of the things that you guys do So instead of ingesting the data, And how did you come up with this? and we allow you to route and recovering the data database that holds the model, capturing the insights, if you will, that may or may not happen in the future. So it's not to say you that is not the your sauce in the same way. and the earlier adopters. And now you said expanded out to And the idea for us is that, the opportunity to come in So like the tip customer and the costs are going to be a quarter and the product strategy. and let the customers enjoy the, you know, of the next five years, the more data you have many times, You have the ability to and not have to start by Go from insights backwards, I love the pivot love how you guys and build the next generation and thank you for Thank you very much the AWS Startup Showcase
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
Ariel Assaraf | PERSON | 0.99+ |
$200,000 | QUANTITY | 0.99+ |
Israel | LOCATION | 0.99+ |
India | LOCATION | 0.99+ |
90% | QUANTITY | 0.99+ |
John | PERSON | 0.99+ |
last week | DATE | 0.99+ |
$1,000 | QUANTITY | 0.99+ |
Tel Aviv | LOCATION | 0.99+ |
10X | QUANTITY | 0.99+ |
John Furrier | PERSON | 0.99+ |
two kilograms | QUANTITY | 0.99+ |
100 gig | QUANTITY | 0.99+ |
Mumbai | LOCATION | 0.99+ |
UK | LOCATION | 0.99+ |
50 | QUANTITY | 0.99+ |
Ariel | PERSON | 0.99+ |
50 people | QUANTITY | 0.99+ |
Coralogix | ORGANIZATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
five | QUANTITY | 0.99+ |
this week | DATE | 0.99+ |
three years | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
five people | QUANTITY | 0.99+ |
100X | QUANTITY | 0.99+ |
Today | DATE | 0.99+ |
five year | QUANTITY | 0.99+ |
each log | QUANTITY | 0.99+ |
about $2 million | QUANTITY | 0.99+ |
four months | QUANTITY | 0.99+ |
five years | QUANTITY | 0.99+ |
one piece | QUANTITY | 0.99+ |
millions of records | QUANTITY | 0.99+ |
60 | QUANTITY | 0.99+ |
50 companies | QUANTITY | 0.99+ |
almost $100 million | QUANTITY | 0.99+ |
one week | QUANTITY | 0.99+ |
Gartner | ORGANIZATION | 0.99+ |
500 | QUANTITY | 0.98+ |
Asia | LOCATION | 0.98+ |
Coralogix | PERSON | 0.98+ |
West US | LOCATION | 0.98+ |
over 1500 | QUANTITY | 0.98+ |
East Center | LOCATION | 0.97+ |
under $1,000 | QUANTITY | 0.97+ |
first | QUANTITY | 0.96+ |
each log records | QUANTITY | 0.96+ |
10 records a day | QUANTITY | 0.96+ |
one | QUANTITY | 0.96+ |
end of 2017 | DATE | 0.96+ |
about 100 gigs | QUANTITY | 0.96+ |
Streama | TITLE | 0.95+ |
$1,000 a month | QUANTITY | 0.95+ |
R&D | ORGANIZATION | 0.95+ |
beginning | DATE | 0.95+ |
first few years | QUANTITY | 0.93+ |
past three months | DATE | 0.93+ |
$70,000 a month | QUANTITY | 0.9+ |
Coralogix | TITLE | 0.9+ |
GCP | ORGANIZATION | 0.88+ |
TCO | ORGANIZATION | 0.88+ |
AWS Startup Showcase | EVENT | 0.87+ |
Breaking Analysis: Why Apple Could be the Key to Intel's Future
>> From theCUBE studios in Palo Alto, in Boston bringing you data-driven insights from theCUBE and ETR. This is Breaking Analysis with Dave Vellante >> The latest Arm Neoverse announcement further cements our opinion that it's architecture business model and ecosystem execution are defining a new era of computing and leaving Intel in it's dust. We believe the company and its partners have at least a two year lead on Intel and are currently in a far better position to capitalize on a major waves that are driving the technology industry and its innovation. To compete our view is that Intel needs a new strategy. Now, Pat Gelsinger is bringing that but they also need financial support from the US and the EU governments. Pat Gelsinger was just noted as asking or requesting from the EU government $9 billion, sorry, 8 billion euros in financial support. And very importantly, Intel needs a volume for its new Foundry business. And that is where Apple could be a key. Hello, everyone. And welcome to this week's weekly bond Cube insights powered by ETR. In this breaking analysis will explain why Apple could be the key to saving Intel and America's semiconductor industry leadership. We'll also further explore our scenario of the evolution of computing and what will happen to Intel if it can't catch up. Here's a hint it's not pretty. Let's start by looking at some of the key assumptions that we've made that are informing our scenarios. We've pointed out many times that we believe Arm wafer volumes are approaching 10 times those of x86 wafers. This means that manufacturers of Arm chips have a significant cost advantage over Intel. We've covered that extensively, but we repeat it because when we see news reports and analysis and print it's not a factor that anybody's highlighting. And this is probably the most important issue that Intel faces. And it's why we feel that Apple could be Intel's savior. We'll come back to that. We've projected that the chip shortage will last no less than three years, perhaps even longer. As we reported in a recent breaking analysis. Well, Moore's law is waning. The result of Moore's law, I.e the doubling of processor performance every 18 to 24 months is actually accelerating. We've observed and continue to project a quadrupling of performance every two years, breaking historical norms. Arm is attacking the enterprise and the data center. We see hyperscalers as the tip of their entry spear. AWS's graviton chip is the best example. Amazon and other cloud vendors that have engineering and software capabilities are making Arm-based chips capable of running general purpose applications. This is a huge threat to x86. And if Intel doesn't quickly we believe Arm will gain a 50% share of an enterprise semiconductor spend by 2030. We see the definition of Cloud expanding. Cloud is no longer a remote set of services, in the cloud, rather it's expanding to the edge where the edge could be a data center, a data closet, or a true edge device or system. And Arm is by far in our view in the best position to support the new workloads and computing models that are emerging as a result. Finally geopolitical forces are at play here. We believe the U S government will do, or at least should do everything possible to ensure that Intel and the U S chip industry regain its leadership position in the semiconductor business. If they don't the U S and Intel could fade to irrelevance. Let's look at this last point and make some comments on that. Here's a map of the South China sea in a way off in the Pacific we've superimposed a little pie chart. And we asked ourselves if you had a hundred points of strategic value to allocate, how much would you put in the semiconductor manufacturing bucket and how much would go to design? And our conclusion was 50, 50. Now it used to be because of Intel's dominance with x86 and its volume that the United States was number one in both strategic areas. But today that orange slice of the pie is dominated by TSMC. Thanks to Arm volumes. Now we've reported extensively on this and we don't want to dwell on it for too long but on all accounts cost, technology, volume. TSMC is the clear leader here. China's president Xi has a stated goal of unifying Taiwan by China's Centennial in 2049, will this tiny Island nation which dominates a critical part of the strategic semiconductor pie, go the way of Hong Kong and be subsumed into China. Well, military experts say it was very hard for China to take Taiwan by force, without heavy losses and some serious international repercussions. The US's military presence in the Philippines and Okinawa and Guam combined with support from Japan and South Korea would make it even more difficult. And certainly the Taiwanese people you would think would prefer their independence. But Taiwanese leadership, it ebbs and flows between those hardliners who really want to separate and want independence and those that are more sympathetic to China. Could China for example, use cyber warfare to over time control the narrative in Taiwan. Remember if you control the narrative you can control the meme. If you can crawl the meme you control the idea. If you control the idea, you control the belief system. And if you control the belief system you control the population without firing a shot. So is it possible that over the next 25 years China could weaponize propaganda and social media to reach its objectives with Taiwan? Maybe it's a long shot but if you're a senior strategist in the U S government would you want to leave that to chance? We don't think so. Let's park that for now and double click on one of our key findings. And that is the pace of semiconductor performance gains. As we first reported a few weeks ago. Well, Moore's law is moderating the outlook for cheap dense and efficient processing power has never been better. This slideshows two simple log lines. One is the traditional Moore's law curve. That's the one at the bottom. And the other is the current pace of system performance improvement that we're seeing measured in trillions of operations per second. Now, if you calculate the historical annual rate of processor performance improvement that we saw with x86, the math comes out to around 40% improvement per year. Now that rate is slowing. It's now down to around 30% annually. So we're not quite doubling every 24 months anymore with x86 and that's why people say Moore's law is dead. But if you look at the (indistinct) effects of packaging CPU's, GPU's, NPUs accelerators, DSPs and all the alternative processing power you can find in SOC system on chip and eventually system on package it's growing at more than a hundred percent per annum. And this means that the processing power is now quadrupling every 24 months. That's impressive. And the reason we're here is Arm. Arm has redefined the core process of model for a new era of computing. Arm made an announcement last week which really recycle some old content from last September, but it also put forth new proof points on adoption and performance. Arm laid out three components and its announcement. The first was Neoverse version one which is all about extending vector performance. This is critical for high performance computing HPC which at one point you thought that was a niche but it is the AI platform. AI workloads are not a niche. Second Arm announced the Neoverse and two platform based on the recently introduced Arm V9. We talked about that a lot in one of our earlier Breaking Analysis. This is going to performance boost of around 40%. Now the third was, it was called CMN-700 Arm maybe needs to work on some of its names, but Arm said this is the industry's most advanced mesh interconnect. This is the glue for the V1 and the N2 platforms. The importance is it allows for more efficient use and sharing of memory resources across components of the system package. We talked about this extensively in previous episodes the importance of that capability. Now let's share with you this wheel diagram underscores the completeness of the Arm platform. Arms approach is to enable flexibility across an open ecosystem, allowing for value add at many levels. Arm has built the architecture in design and allows an open ecosystem to provide the value added software. Now, very importantly, Arm has created the standards and specifications by which they can with certainty, certify that the Foundry can make the chips to a high quality standard, and importantly that all the applications are going to run properly. In other words, if you design an application, it will work across the ecosystem and maintain backwards compatibility with previous generations, like Intel has done for years but Arm as we'll see next is positioning not only for existing workloads but also the emerging high growth applications. To (indistinct) here's the Arm total available market as we see it, we think the end market spending value of just the chips going into these areas is $600 billion today. And it's going to grow to 1 trillion by 2030. In other words, we're allocating the value of the end market spend in these sectors to the marked up value of the Silicon as a percentage of the total spend. It's enormous. So the big areas are Hyperscale Clouds which we think is around 20% of this TAM and the HPC and AI workloads, which account for about 35% and the Edge will ultimately be the largest of all probably capturing 45%. And these are rough estimates and they'll ebb and flow and there's obviously some overlap but the bottom line is the market is huge and growing very rapidly. And you see that little red highlighted area that's enterprise IT. Traditional IT and that's the x86 market in context. So it's relatively small. What's happening is we're seeing a number of traditional IT vendors, packaging x86 boxes throwing them over the fence and saying, we're going after the Edge. And what they're doing is saying, okay the edge is this aggregation point for all these end point devices. We think the real opportunity at the Edge is for AI inferencing. That, that is where most of the activity and most of the spending is going to be. And we think Arm is going to dominate that market. And this brings up another challenge for Intel. So we've made the point a zillion times that PC volumes peaked in 2011. And we saw that as problematic for Intel for the cost reasons that we've beat into your head. And lo and behold PC volumes, they actually grew last year thanks to COVID and we'll continue to grow it seems for a year or so. Here's some ETR data that underscores that fact. This chart shows the net score. Remember that's spending momentum it's the breakdown for Dell's laptop business. The green means spending is accelerating and the red is decelerating. And the blue line is net score that spending momentum. And the trend is up and to the right now, as we've said this is great news for Dell and HP and Lenovo and Apple for its laptops, all the laptops sellers but it's not necessarily great news for Intel. Why? I mean, it's okay. But what it does is it shifts Intel's product mix toward lower margin, PC chips and it squeezes Intel's gross margins. So the CFO has to explain that margin contraction to wall street. Imagine that the business that got Intel to its monopoly status is growing faster than the high margin server business. And that's pulling margins down. So as we said, Intel is fighting a war on multiple fronts. It's battling AMD in the core x86 business both PCs and servers. It's watching Arm mop up in mobile. It's trying to figure out how to reinvent itself and change its culture to allow more flexibility into its designs. And it's spinning up a Foundry business to compete with TSMC. So it's got to fund all this while at the same time propping up at stock with buybacks Intel last summer announced that it was accelerating it's $10 billion stock buyback program, $10 billion. Buy stock back, or build a Foundry which do you think is more important for the future of Intel and the us semiconductor industry? So Intel, it's got to protect its past while building his future and placating wall street all at the same time. And here's where it gets even more dicey. Intel's got to protect its high-end x86 business. It is the cash cow and funds their operation. Who's Intel's biggest customer Dell, HP, Facebook, Google Amazon? Well, let's just say Amazon is a big customer. Can we agree on that? And we know AWS is biggest revenue generator is EC2. And EC2 was powered by microprocessors made from Intel and others. We found this slide in the Arm Neoverse deck and it caught our attention. The data comes from a data platform called lifter insights. The charts show, the rapid growth of AWS is graviton chips which are they're custom designed chips based on Arm of course. The blue is that graviton and the black vendor A presumably is Intel and the gray is assumed to be AMD. The eye popper is the 2020 pie chart. The instant deployments, nearly 50% are graviton. So if you're Pat Gelsinger, you better be all over AWS. You don't want to lose this customer and you're going to do everything in your power to keep them. But the trend is not your friend in this account. Now the story gets even gnarlier and here's the killer chart. It shows the ISV ecosystem platforms that run on graviton too, because AWS has such good engineering and controls its own stack. It can build Arm-based chips that run software designed to run on general purpose x86 systems. Yes, it's true. The ISV, they got to do some work, but large ISV they have a huge incentives because they want to ride the AWS wave. Certainly the user doesn't know or care but AWS cares because it's driving costs and energy consumption down and performance up. Lower cost, higher performance. Sounds like something Amazon wants to consistently deliver, right? And the ISV portfolio that runs on our base graviton and it's just going to continue to grow. And by the way, it's not just Amazon. It's Alibaba, it's Oracle, it's Marvell. It's 10 cents. The list keeps growing Arm, trotted out a number of names. And I would expect over time it's going to be Facebook and Google and Microsoft. If they're not, are you there? Now the last piece of the Arm architecture story that we want to share is the progress that they're making and compare that to x86. This chart shows how Arm is innovating and let's start with the first line under platform capabilities. Number of cores supported per die or, or system. Now die is what ends up as a chip on a small piece of Silicon. Think of the die as circuit diagram of the chip if you will, and these circuits they're fabricated on wafers using photo lithography. The wafers then cut up into many pieces each one, having a chip. Each of these pieces is the chip. And two chips make up a system. The key here is that Arm is quadrupling the number of cores instead of increasing thread counts. It's giving you cores. Cores are better than threads because threads are shared and cores are independent and much easier to virtualize. This is particularly important in situations where you want to be as efficient as possible sharing massive resources like the Cloud. Now, as you can see in the right hand side of the chart under the orange Arm is dramatically increasing the amount of capabilities compared to previous generations. And one of the other highlights to us is that last line that CCIX and CXL support again Arm maybe needs to name these better. These refer to Arms and memory sharing capabilities within and between processors. This allows CPU's GPU's NPS, et cetera to share resources very often efficiently especially compared to the way x86 works where everything is currently controlled by the x86 processor. CCIX and CXL support on the other hand will allow designers to program the system and share memory wherever they want within the system directly and not have to go through the overhead of a central processor, which owns the memory. So for example, if there's a CPU, GPU, NPU the CPU can say to the GPU, give me your results at a specified location and signal me when you're done. So when the GPU is finished calculating and sending the results, the GPU just signals the operation is complete. Versus having to ping the CPU constantly, which is overhead intensive. Now composability in that chart means the system it's a fixed. Rather you can programmatically change the characteristics of the system on the fly. For example, if the NPU is idle you can allocate more resources to other parts of the system. Now, Intel is doing this too in the future but we think Arm is way ahead. At least by two years this is also huge for Nvidia, which today relies on x86. A major problem for Nvidia has been coherent memory management because the utilization of its GPU is appallingly low and it can't be easily optimized. Last week, Nvidia announced it's intent to provide an AI capability for the data center without x86 I.e using Arm-based processors. So Nvidia another big Intel customer is also moving to Arm. And if it's successful acquiring Arm which is still a long shot this trend is only going to accelerate. But the bottom line is if Intel can't move fast enough to stem the momentum of Arm we believe Arm will capture 50% of the enterprise semiconductor spending by 2030. So how does Intel continue to lead? Well, it's not going to be easy. Remember we said, Intel, can't go it alone. And we posited that the company would have to initiate a joint venture structure. We propose a triumvirate of Intel, IBM with its power of 10 and memory aggregation and memory architecture And Samsung with its volume manufacturing expertise on the premise that it coveted in on US soil presence. Now upon further review we're not sure the Samsung is willing to give up and contribute its IP to this venture. It's put a lot of money and a lot of emphasis on infrastructure in South Korea. And furthermore, we're not convinced that Arvind Krishna who we believe ultimately made the call to Jettisons. Jettison IBM's micro electronics business wants to put his efforts back into manufacturing semi-conductors. So we have this conundrum. Intel is fighting AMD, which is already at seven nanometer. Intel has a fall behind in process manufacturing which is strategically important to the United States it's military and the nation's competitiveness. Intel's behind the curve on cost and architecture and is losing key customers in the most important market segments. And it's way behind on volume. The critical piece of the pie that nobody ever talks about. Intel must become more price and performance competitive with x86 and bring in new composable designs that maintain x86 competitive. And give the ability to allow customers and designers to add and customize GPU's, NPUs, accelerators et cetera. All while launching a successful Foundry business. So we think there's another possibility to this thought exercise. Apple is currently reliant on TSMC and is pushing them hard toward five nanometer, in fact sucking up a lot of that volume and TSMC is maybe not servicing some other customers as well as it's servicing Apple because it's a bit destructive, it is distracted and you have this chip shortage. So Apple because of its size gets the lion's share of the attention but Apple needs a trusted onshore supplier. Sure TSMC is adding manufacturing capacity in the US and Arizona. But back to our precarious scenario in the South China sea. Will the U S government and Apple sit back and hope for the best or will they hope for the best and plan for the worst? Let's face it. If China gains control of TSMC, it could block access to the latest and greatest process technology. Apple just announced that it's investing billions of dollars in semiconductor technology across the US. The US government is pressuring big tech. What about an Apple Intel joint venture? Apple brings the volume, it's Cloud, it's Cloud, sorry. It's money it's design leadership, all that to the table. And they could partner with Intel. It gives Intel the Foundry business and a guaranteed volume stream. And maybe the U S government gives Apple a little bit of breathing room and the whole big up big breakup, big tech narrative. And even though it's not necessarily specifically targeting Apple but maybe the US government needs to think twice before it attacks big tech and thinks about the long-term strategic ramifications. Wouldn't that be ironic? Apple dumps Intel in favor of Arm for the M1 and then incubates, and essentially saves Intel with a pipeline of Foundry business. Now back to IBM in this scenario, we've put a question mark on the slide because maybe IBM just gets in the way and why not? A nice clean partnership between Intel and Apple? Who knows? Maybe Gelsinger can even negotiate this without giving up any equity to Apple, but Apple could be a key ingredient to a cocktail of a new strategy under Pat Gelsinger leadership. Gobs of cash from the US and EU governments and volume from Apple. Wow, still a long shot, but one worth pursuing because as we've written, Intel is too strategic to fail. Okay, well, what do you think? You can DM me @dvellante or email me at david.vellante@siliconangle.com or comment on my LinkedIn post. Remember, these episodes are all available as podcasts so please subscribe wherever you listen. I publish weekly on wikibon.com and siliconangle.com. And don't forget to check out etr.plus for all the survey analysis. And I want to thank my colleague, David Floyer for his collaboration on this and other related episodes. This is Dave Vellante for theCUBE insights powered by ETR. Thanks for watching, be well, and we'll see you next time. (upbeat music)
SUMMARY :
This is Breaking Analysis and most of the spending is going to be.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
David Floyer | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
HP | ORGANIZATION | 0.99+ |
Apple | ORGANIZATION | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Samsung | ORGANIZATION | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
TSMC | ORGANIZATION | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
2011 | DATE | 0.99+ |
Lenovo | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.99+ | |
Pat Gelsinger | PERSON | 0.99+ |
$10 billion | QUANTITY | 0.99+ |
Nvidia | ORGANIZATION | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
ORGANIZATION | 0.99+ | |
50% | QUANTITY | 0.99+ |
Alibaba | ORGANIZATION | 0.99+ |
$600 | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
45% | QUANTITY | 0.99+ |
two chips | QUANTITY | 0.99+ |
10 times | QUANTITY | 0.99+ |
10 cents | QUANTITY | 0.99+ |
South Korea | LOCATION | 0.99+ |
US | LOCATION | 0.99+ |
Last week | DATE | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
Arizona | LOCATION | 0.99+ |
U S | ORGANIZATION | 0.99+ |
Boston | LOCATION | 0.99+ |
david.vellante@siliconangle.com | OTHER | 0.99+ |
1 trillion | QUANTITY | 0.99+ |
2030 | DATE | 0.99+ |
Marvell | ORGANIZATION | 0.99+ |
China | ORGANIZATION | 0.99+ |
Arvind Krishna | PERSON | 0.99+ |
two years | QUANTITY | 0.99+ |
Moore | PERSON | 0.99+ |
$9 billion | QUANTITY | 0.99+ |
10 | QUANTITY | 0.99+ |
EU | ORGANIZATION | 0.99+ |
last year | DATE | 0.99+ |
last week | DATE | 0.99+ |
twice | QUANTITY | 0.99+ |
first line | QUANTITY | 0.99+ |
Okinawa | LOCATION | 0.99+ |
last September | DATE | 0.99+ |
Hong Kong | LOCATION | 0.99+ |
Andre Dufour, AWS | AWS re:Invent 2020
>>From around the globe with digital coverage of AWS reinvent 2020, sponsored by Intel and AWS welcome everyone to the cube live and our coverage of AWS reinvent 2020. I'm your host Rebecca Knight. Today we are joined by Andre due for, he is the general manager of Amazon location service. Thank you so much for coming on the show. Andre. >>Thanks so much, Rebecca. It's a pleasure. >>So Amazon, AWS is announcing a Amazon location service in preview. Tell us a little bit more about what it does. What was the impetus for it? >>Of course. Well, Amazon location service is a new geospatial service that makes it easy for customers on AWS to integrate location information into their applications. And when I say location information, I mean a couple of specific things, mops points of interest places, and geocodes from trusted global high quality data partners. And one of the things that's really cool about Amazon location is we enable customers to access this high-quality data in a way that's incredibly cost effective. It's up to 10 times cheaper than some of the alternatives. And so what that means for customers is they can bring to life use cases that previously would have been inconceivable because they just weren't cost effective. Additionally, Amazon location takes privacy very seriously. And so, you know, customers have told us many times that they're, they're, they're very concerned about their location information, leaving their control. Whereas with Amazon location, we keep customer's location data in their AWS account unless they decide otherwise. And finally, what we've seen with customers who are using Amazon location is they're able to move from experimentation with location ideas, to scale production, much more quickly than they otherwise could have because it's a native AWS service. So we're so excited to be announcing this >>Well, you just mentioned cost privacy scale production, three things that are definitely on customers' minds right now. Tell us a little bit more about these use cases. How are customers using it? >>Yeah, that's a great question. I think it's often easiest to understand the capabilities through the lens of a use case. Now it turns out location in, in more and more customer conversations is pervasive across a bunch of different use cases, but I'll touch on maybe just for today. So one thing that we're seeing customers commonly using location for is location-based customer engagement. And so what that means is including a location component, when you are reaching out to your customers with timely offers. So for example, when they're in close proximity to one of your retail locations, sending them an offer tends to increase their satisfaction and their conversion an additional use case that Springs to mind immediately in many of the conversations is using maps for striking visualizations of data, either showing a route between two points or dropping location pins on a map in order to enhance the visual understanding of subject matter. >>Additionally, customers tend to use Amazon location for asset tracking. They want to know where their things are in the world and be able to reason over that both in real time in order to make decisions or retrospectively in order to optimize or to audit. And additionally, um, customers also use us in end to end delivery use cases, be it last mile delivery for, uh, goods that were ordered online or, uh, food delivery, which of course is, uh, increasingly prevalent these days. And so, yeah, you know, one of the customer examples that I think is especially compelling here because it touches on a couple of these is a company called Singleton solutions and their product is called mobile log. Uh, it's effectively last mile as a service in the cloud. And what it lets customers do is manage the logistics of a delivery business. And so what mobile log and Singleton have been able to do is retire a lot of the custom code that they had built because nothing was really available to meet their location needs. They were able to consolidate their location infrastructure from multiple clouds onto just AWS, which simplifies their solution. They were able to move more quickly as they innovate on behalf of their customers. And they managed to reduce their costs while doing this by up to 60%. So I think it's a pretty cool example of what location can do for customers. >>What are some other industries and apps and applications that would benefit most from this affordable location data? >>Yeah, well, it's, uh, it tends to spend many different industries. So we're seeing a lot of uses as you can imagine in transportation and logistics and, and certainly that's, uh, an industry that's growing very quickly, um, government and public sector attempt to have a need to, uh, visualize a lot of information, uh, on, on maps. Um, we are seeing retail and folks interested in customer engagement. Um, it really is springing up everywhere and often B uh, the conversations kind of have a location component in disguise. For example, we were talking to a telecom service provider who is telling us, well, you know, I can save billions of dollars if I increase the efficiency of my truck rolls. Well, that's the location use case, right? If people are talking about, uh, actually one, one customer, uh, or a person who has used us in beta is post NL, and they're telling us, you know, if they can increase just the, um, loading factor of their trucks by 1%, uh, in, uh, over time, this is big dollar savings for them. And not, that's all about location and about optimizing, uh, the, the routing and dispatch of their vehicles. And so really it's springing up everywhere, but it doesn't always sound like a map or a geocode it's, uh, more of these business level considerations around optimization around moving faster and around serving customers more quickly. >>You mentioned a couple of, of industries and logistics areas where this is being used. What are, which customers are currently using Amazon location service? >>Well, so there are a couple that I, uh, I mentioned, so of course we're only just launching today. We've had a beta program, uh, and we have a couple of references that we can talk about publicly. So Singleton is the very first that we touched on, and this is a company that's operative in the delivery and, uh, dispatch logistics space. And so they they've been using us to, to advantage and, and have realized some pretty significant cost savings. Uh, the other company that's been, uh, experimenting with Amazon location, uh, again in sort of a similar space, but with a different geography is posted on owl. And so they're the number one, uh, e-commerce and delivery, uh, her postal logistics company in the Netherlands. And what, what they're actually using us for is to, uh, do asset tracking on their delivery roller cages in order to, uh, understand where they are in the world and make better decisions as to where they should be in relation to the demand. >>Andre, I want you to close this out here. And as you said, you launched today, you've been in beta, what is in store for 2021 with Amazon location service? What can, what can we expect? What can customers expect? >>Yeah, so we're, we're in preview today and it's an open preview, so people can, can just go to the console and directly use it. You don't need to sign up. And what we have to look forward to in the first part of 2021 is general availability of the service. And you can imagine that we'll be rolling that out over everyone regions, because there's significant demand for this all over the world. And then it's a fairly typical, uh, AWS motion where what we're going to do is listen, because 90% of our roadmap is compelled by customer requests. And so we'll be very attentive to how people are using the service, where they see additional opportunities for us to serve them better. And we will move with vigor on those. >>Great. And for customers who want to find out more, what, what should they do? >>Well, the easiest thing to do is to go to aws.amazon.com/location, and then, uh, check, check us out there and get started with the service today. >>Great, well, Andre do for, thank you so much for coming on the Cuba really interesting conversation. >>Thank you so much. It's been a privilege. >>I'm Rebecca Knight stay tuned for more of the cubes coverage of AWS reinvent 2020.
SUMMARY :
From around the globe with digital coverage of AWS Thanks so much, Rebecca. Tell us a little bit more about what it does. And so what that means for customers is they can bring to life use cases that previously would have been inconceivable Well, you just mentioned cost privacy scale production, three things that are definitely on customers' minds And so what that means is including a location component, when you are reaching out to your customers And so what mobile log and Singleton And so really it's springing up everywhere, You mentioned a couple of, of industries and logistics areas where this is being used. Uh, the other company that's been, uh, experimenting with Amazon location, uh, And as you said, you launched today, you've been in beta, And then it's a fairly typical, uh, AWS motion where what we're going to do is listen, And for customers who want to find out more, what, what should they do? Well, the easiest thing to do is to go to aws.amazon.com/location, Thank you so much.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Rebecca | PERSON | 0.99+ |
Rebecca Knight | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Andre Dufour | PERSON | 0.99+ |
Andre | PERSON | 0.99+ |
Singleton | ORGANIZATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
90% | QUANTITY | 0.99+ |
two points | QUANTITY | 0.99+ |
aws.amazon.com/location | OTHER | 0.99+ |
2021 | DATE | 0.99+ |
Netherlands | LOCATION | 0.99+ |
1% | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
first | QUANTITY | 0.99+ |
Cuba | LOCATION | 0.99+ |
Today | DATE | 0.98+ |
up to 60% | QUANTITY | 0.98+ |
Intel | ORGANIZATION | 0.97+ |
billions of dollars | QUANTITY | 0.97+ |
one thing | QUANTITY | 0.96+ |
up to 10 times | QUANTITY | 0.95+ |
both | QUANTITY | 0.94+ |
mobile log | ORGANIZATION | 0.92+ |
first part | QUANTITY | 0.91+ |
Invent 2020 | TITLE | 0.82+ |
three things | QUANTITY | 0.82+ |
couple | QUANTITY | 0.79+ |
one customer | QUANTITY | 0.75+ |
reinvent 2020 | TITLE | 0.7+ |
2020 | DATE | 0.65+ |
Andre Dufour, AWS | AWS re:Invent 2020
>>From around the globe with digital coverage of AWS reinvent 2020, sponsored by Intel and AWS welcome everyone to the cube live and our coverage of AWS reinvent 2020. I'm your host Rebecca Knight. Today we are joined by Andre due for, he is the general manager of Amazon location service. Thank you so much for coming on the show. Andre. >>Thanks so much, Rebecca. It's a pleasure. >>So Amazon, AWS is announcing a Amazon location service in preview. Tell us a little bit more about what it does. What was the impetus for it? >>Of course. Well, Amazon location service is a new geospatial service that makes it easy for customers on AWS to integrate location information into their applications. And when I say location information, I mean a couple of specific things, mops points of interest places, and geocodes from trusted global high quality data partners. And one of the things that's really cool about Amazon location is we enable customers to access this high-quality data in a way that's incredibly cost effective. It's up to 10 times cheaper than some of the alternatives. And so what that means for customers is they can bring to life use cases that previously would have been inconceivable because they just weren't cost effective. Additionally, Amazon location takes privacy very seriously. And so, you know, customers have told us many times that they're, they're, they're very concerned about their location information, leaving their control. Whereas with Amazon location, we keep customer's location data in their AWS account unless they decide otherwise. And finally, what we've seen with customers who are using Amazon location is they're able to move from experimentation with location ideas, to scale production, much more quickly than they otherwise could have because it's a native AWS service. So we're so excited to be announcing this >>Well, you just mentioned cost privacy scale production, three things that are definitely on customers' minds right now. Tell us a little bit more about these use cases. How are customers using it? >>Yeah, that's a great question. I think it's often easiest to understand the capabilities through the lens of a use case. Now it turns out location in, in more and more customer conversations is pervasive across a bunch of different use cases, but I'll touch on maybe just for today. So one thing that we're seeing customers commonly using location for is location-based customer engagement. And so what that means is including a location component, when you are reaching out to your customers with timely offers. So for example, when they're in close proximity to one of your retail locations, sending them an offer tends to increase their satisfaction and their conversion an additional use case that Springs to mind immediately in many of the conversations is using maps for striking visualizations of data, either showing a route between two points or dropping location pins on a map in order to enhance the visual understanding of subject matter. >>Additionally, customers tend to use Amazon location for asset tracking. They want to know where their things are in the world and be able to reason over that both in real time in order to make decisions or retrospectively in order to optimize or to audit. And additionally, um, customers also use us in end to end delivery use cases, be it last mile delivery for, uh, goods that were ordered online or, uh, food delivery, which of course is, uh, increasingly prevalent these days. And so, yeah, you know, one of the customer examples that I think is especially compelling here because it touches on a couple of these is a company called Singleton solutions and their product is called mobile log. Uh, it's effectively last mile as a service in the cloud. And what it lets customers do is manage the logistics of a delivery business. And so what mobile log and Singleton have been able to do is retire a lot of the custom code that they had built because nothing was really available to meet their location needs. They were able to consolidate their location infrastructure from multiple clouds onto just AWS, which simplifies their solution. They were able to move more quickly as they innovate on behalf of their customers. And they managed to reduce their costs while doing this by up to 60%. So I think it's a pretty cool example of what location can do for customers. >>What are some other industries and apps and applications that would benefit most from this affordable location data? >>Yeah, well, it's, uh, it tends to spend many different industries. So we're seeing a lot of uses as you can imagine in transportation and logistics and, and certainly that's, uh, an industry that's growing very quickly, um, government and public sector attempt to have a need to, uh, visualize a lot of information, uh, on, on maps. Um, we are seeing retail and folks interested in customer engagement. Um, it really is springing up everywhere and often B uh, the conversations kind of have a location component in disguise. For example, we were talking to a telecom service provider who is telling us, well, you know, I can save billions of dollars if I increase the efficiency of my truck rolls. Well, that's the location use case, right? If people are talking about, uh, actually one, one customer, uh, or a person who has used us in beta is post NL, and they're telling us, you know, if they can increase just the, um, loading factor of their trucks by 1%, uh, in, uh, over time, this is big dollar savings for them. And not, that's all about location and about optimizing, uh, the, the routing and dispatch of their vehicles. And so really it's springing up everywhere, but it doesn't always sound like a map or a geocode it's, uh, more of these business level considerations around optimization around moving faster and around serving customers more quickly. >>You mentioned a couple of, of industries and logistics areas where this is being used. What are, which customers are currently using Amazon location service? >>Well, so there are a couple that I, uh, I mentioned, so of course we're only just launching today. We've had a beta program, uh, and we have a couple of references that we can talk about publicly. So Singleton is the very first that we touched on, and this is a company that's operative in the delivery and, uh, dispatch logistics space. And so they they've been using us to, to advantage and, and have realized some pretty significant cost savings. Uh, the other company that's been, uh, experimenting with Amazon location, uh, again in sort of a similar space, but with a different geography is posted on owl. And so they're the number one, uh, e-commerce and delivery, uh, her postal logistics company in the Netherlands. And what, what they're actually using us for is to, uh, do asset tracking on their delivery roller cages in order to, uh, understand where they are in the world and make better decisions as to where they should be in relation to the demand. >>Andre, I want you to close this out here. And as you said, you launched today, you've been in beta, what is in store for 2021 with Amazon location service? What can, what can we expect? What can customers expect? >>Yeah, so we're, we're in preview today and it's an open preview, so people can, can just go to the console and directly use it. You don't need to sign up. And what we have to look forward to in the first part of 2021 is general availability of the service. And you can imagine that we'll be rolling that out over everyone regions, because there's significant demand for this all over the world. And then it's a fairly typical, uh, AWS motion where what we're going to do is listen, because 90% of our roadmap is compelled by customer requests. And so we'll be very attentive to how people are using the service, where they see additional opportunities for us to serve them better. And we will move with vigor on those. >>Great. And for customers who want to find out more, what, what should they do? >>Well, the easiest thing to do is to go to aws.amazon.com/location, and then, uh, check, check us out there and get started with the service today. >>Great, well, Andre do for, thank you so much for coming on the Cuba really interesting conversation. >>Thank you so much. It's been a privilege. >>I'm Rebecca Knight stay tuned for more of the cubes coverage of AWS reinvent 2020.
SUMMARY :
From around the globe with digital coverage of AWS Thanks so much, Rebecca. Tell us a little bit more about what it does. And so what that means for customers is they can bring to life use cases that previously would have been inconceivable Well, you just mentioned cost privacy scale production, three things that are definitely on customers' minds And so what that means is including a location component, when you are reaching out to your customers And so what mobile log and Singleton And so really it's springing up everywhere, You mentioned a couple of, of industries and logistics areas where this is being used. Uh, the other company that's been, uh, experimenting with Amazon location, uh, And as you said, you launched today, you've been in beta, And then it's a fairly typical, uh, AWS motion where what we're going to do is listen, And for customers who want to find out more, what, what should they do? Well, the easiest thing to do is to go to aws.amazon.com/location, Thank you so much.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Rebecca | PERSON | 0.99+ |
Rebecca Knight | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Andre Dufour | PERSON | 0.99+ |
Andre | PERSON | 0.99+ |
Singleton | ORGANIZATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
90% | QUANTITY | 0.99+ |
two points | QUANTITY | 0.99+ |
aws.amazon.com/location | OTHER | 0.99+ |
2021 | DATE | 0.99+ |
Netherlands | LOCATION | 0.99+ |
1% | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
first | QUANTITY | 0.99+ |
Cuba | LOCATION | 0.99+ |
Today | DATE | 0.98+ |
up to 60% | QUANTITY | 0.98+ |
Intel | ORGANIZATION | 0.97+ |
billions of dollars | QUANTITY | 0.97+ |
one thing | QUANTITY | 0.96+ |
up to 10 times | QUANTITY | 0.95+ |
both | QUANTITY | 0.94+ |
mobile log | ORGANIZATION | 0.92+ |
first part | QUANTITY | 0.91+ |
Invent 2020 | TITLE | 0.82+ |
three things | QUANTITY | 0.82+ |
couple | QUANTITY | 0.79+ |
one customer | QUANTITY | 0.75+ |
reinvent 2020 | TITLE | 0.7+ |
2020 | DATE | 0.65+ |
Tomer Levy, Logz.io | AWS re:Invent 2020
>> Narrator: From around the globe it's theCUBE with digital coverage of AWS reinvent 2020. Sponsored by Intel, AWS and our community partners. >> All right, you're continuing coverage of AWS reinvent 2020 virtual event. We get the pleasure of covering this show like no other AWS reinvent. We are pulling in from the other side of the world Tomer Levy, CEO of Logz.io. First time Cuber so we're going to ease them into it but it's going to be a great conversation. I'm Keith Townsend at CTO advisor. Tomer, welcome to the show. >> Keith, thank you for having me. I'm super excited to be here. >> You know what? We love having founders here on theCUBE. We have a long history of having deep conversations with builders and we're probably the show for builders. AWS reinvent is virtual. However, I think the spirit of re-invent is highlighted in companies like this. We've seen a lot of observability companies sprout up around the industry. AWS is a big, big magnet for these types of solutions. What's the assets Logz.io and how are you guys differentiating yourselves in this crowded space? >> Yeah, absolutely Keith you see observability is so fundamental to building applications on AWS that as companies develop more applications, they have to have solid observability. And we have a mission and our mission is to enable develop engineers and any engineer out there to use open source to run their observability. So when we were developers we wanted to use open source but we had to compromise on a proprietary solution. We decided to build the company so engineers can use the observability tools they're already using for logging, for metrics, for tracing, Whatever they're already using we want to enable them to use that at scale on AWS. So it's easy to use, it's super smart and the data is coordinated. And I think fundamentally it's what we're doing very differently in the market. There is no other company in the market today that takes the best open sources and bring them together as one super strong platform and we're proud to be that company. >> Well, when you say there's no other company doing open source the way that you guys are doing it, that really intrigues me especially as we look at this from the angle of Cooper Netties, the CEO of the leading virtualization company called Kubernetes, the doubts home of the internet. How do you see the intersection of opensource observability in kubernetes especially in the public cloud? >> Yeah, for sure. People say that kubernetes is almost the operating system of the future and why do people use kubernetes? They use it to make sure they can run multiple microservices. They can take their application which used to be a monolith and put it in a distributed way. So it becomes so much harder to monitor or to troubleshoot even to secure applications. So the way we built Logz.io was really designed for companies that are moving into the cloud, companies moving into kubernetes, into microservices and by having logs and metrics and traces all work together through the best open sources. I think we can help customers really get the visibility and just accelerate the software delivery. Just provide better service to their customers. >> So Levy, walk me through that journey. What is it like for a developer to come from their traditional open source roots and enter the cloud where they're melding public cloud services in AWS alongside their tools that they're using in observability. How do you help ease that transition? >> Yeah, absolutely Keith because one of the main drivers for companies adopting tools like Logz.io is actually the migration to AWS. So imagine now migration to a new ground, what do you have to think about first? Do I have the glasses? Can I see what's going on? Like when I see what's going on, I feel more confident. So if I'm now using, let's call it elk or using the open-source Grafana or using tools like Jaeger, which are all open sources too that we offer as part of our platform. So when I use these tools I'm using them to get visibility into my own application, my own infrastructure. So Logz.io faster transition to Logz.io is super easy. This is the whole notion of having an open source compatible platform. So I want to move to Loz.io, everything that worked with my open source currently still works with Logz.io but now when you move to the cloud Logz.io on AWS, we have a very strong relationship so all the services are automatically monitored. You have pre-configured dashboard, everything is interconnected so just when I jump into the AWS platform I immediately get visibility of my existing apps and of the AWS infrastructure. And that eventually helped me become confident, grow and deliver faster on AWS. >> So again this is a conference full of builders but you used the term devOps. We're starting to see a bleeding of DevOps and builders or operations and builders come together. One of the big trans and DevOps and observability is AI and machine learning. What are some of the features of AI and Machine Learning you guys are bringing to bear to this market? >> Yeah, listen I'm a big believer in AI. You know, the amount of data that companies like Logz.io have to ingest and our customers have to process. It's just something a human being cannot possibly understand. It's like billions and millions of lines of data. So this is where we bring machines to help humans. I'll give you one example, right? If you're a DevOps engineer and you see an issue in your logs, what do you do? You usually copy that and putting it into Google and you'll end up on stack overflow, maybe on GitHub, maybe on another website. What we have done is we've scraped the web and we have learned from any user on our platform. So we actually know which log line is important and which one is not. So when companies send a log line, our AI automatically scans it and says, "Hey, here are the billion log lines. No one cares about but here is one that you should really look at right now because either you know half a million people that were searching for it. There are 7,000 alerts on this and it just happened to you. Keith look, maybe you should jump in and look at that". This is where AI makes us just better operate or better DevOp people and not kind of try to replace us. >> So I'm a technical founder, you're a technical founder, theCUBE loves supporting founders. One of the advantages of being the CEO of your company is that you get to decide the culture and the mission of your company. Talk to me about the people side of your organization and how you're making a change for the better. >> Yeah, absolutely. You know, it is a privilege and to the privilege to start and come with a mission that you want to change something in the world and we were just two developers, a staff, my co-founder and myself having to use a product we didn't want to use and you know still really wanted to use an open source product. So we said let's build the company around that and this is kind of set the mission for the company as the company evolved, so is our mission. It evolves from logging to monitoring, to tracing and we also added a cloud SIEM solution all based on open source. So we're going to DevOps engineers and any engineers and we tag any engineer we tell them, "Hey, you can use the best open source tools in the cloud is one platform without compromising". And that's something that really is very differentiated today and I'm very humbled and excited to be part of this journey and I think the team at Logz.io is as well. >> You know I'm always intrigued about this journey to the cloud. Security is one of these things that intrigues me especially as we look at something as mature in the way open source. We often associate open source with public cloud, cloud native but open source is as old as technology itself. So there is a lot of practices that we bring from legacy, traditional infrastructures into the public cloud. So talk to me about that transition of security and security models? How does observability help to either take our existing tools and migrate them to the public cloud or adopt all new cloud native tools in the public cloud? >> Yeah, for sure. I think security is probably together with observability. One of the top priority that when you think about CTOs and VP of Engineering and CSOs, they're concerned about. So we've taken the observability path and bringing better glasses to our users and then on the security side there's a whole market called the SIEM market where companies look at detecting threats, investigating them and most of these tools were that companies use our legacy, incumbents and for design on their own premises world. And are not really a fit for the dynamic world of kubernetes and the cloud. And this is when we decided a couple of years ago to launch a product in that space and today this product is extremely successful. We have customers protecting their AWS environments across the board. So basically with one product for observability, you can with a single checkbox enable security and then you can detect threats. You can look at kind of the common pitfalls of AWS environment and how you can avoid them. And eventually when you see a threat, you can use our tool to investigate and find the root cause in a tool which was designed on AWS for AWS. And it's really designed for the kind of the native cloud environment rather than the on-premise as well. >> Now, is there an integration between the AI ML law of management and the threat management solutions from our observability perspective? >> Yeah, for sure. This is the beauty, it's all one data platform. So customers ship their data, loads, metrics and traces into one place and then we start to look at how can we provide more value on the data, right? How can we look at the logs from an operational perspective and tell you, "Hey, your production might be going down because of a production risks or maybe we can provide you threat intelligence". We can enrich the data and tell you, "Hey, we think you're undergoing an attack right now". So this is all done by users and it is all enraged by AI that provides more visibility, more enrichment of the data and just advice on where to look. >> So Tomer levy, CEO, founder of Logz.io. You're now a few belong. Thank you for joining the show. I hope you have a very successful AWS reinvent. Speaking of AWS reinvent, theCUBE's nonstop coverage of AWS reinvent continues. Watch some of the world's greatest builders, innovators get challenged on their vision and for us to understand and appreciate the work that's been done in this dynamic community. Continue to watch this coverage and more. Talk to you next interview on the CUBE's coverage, of AWS reinvent 2020. (soft music)
SUMMARY :
the globe it's theCUBE We are pulling in from the I'm super excited to be here. around the industry. differently in the market. doing open source the way So the way we built Logz.io and enter the cloud where is actually the migration to AWS. One of the big trans and You know, the amount of data One of the advantages of in the world and we were in the way open source. One of the top priority that more enrichment of the data on the CUBE's coverage,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Keith Townsend | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Keith | PERSON | 0.99+ |
Tomer Levy | PERSON | 0.99+ |
7,000 alerts | QUANTITY | 0.99+ |
Tomer | PERSON | 0.99+ |
billions | QUANTITY | 0.99+ |
Logz.io | TITLE | 0.99+ |
Intel | ORGANIZATION | 0.99+ |
Loz.io | TITLE | 0.99+ |
two developers | QUANTITY | 0.99+ |
Logz.io | ORGANIZATION | 0.99+ |
One | QUANTITY | 0.99+ |
half a million people | QUANTITY | 0.98+ |
GitHub | ORGANIZATION | 0.98+ |
one platform | QUANTITY | 0.98+ |
Levy | PERSON | 0.98+ |
billion log lines | QUANTITY | 0.98+ |
Tomer levy | PERSON | 0.98+ |
today | DATE | 0.97+ |
one | QUANTITY | 0.97+ |
one example | QUANTITY | 0.97+ |
first | QUANTITY | 0.97+ |
ORGANIZATION | 0.97+ | |
theCUBE | ORGANIZATION | 0.96+ |
CTO | ORGANIZATION | 0.94+ |
Cuber | ORGANIZATION | 0.93+ |
Jaeger | TITLE | 0.93+ |
millions of lines | QUANTITY | 0.89+ |
First time | QUANTITY | 0.87+ |
Kubernetes | ORGANIZATION | 0.85+ |
one data platform | QUANTITY | 0.82+ |
Cooper Netties | PERSON | 0.81+ |
single checkbox | QUANTITY | 0.81+ |
reinvent 2020 | EVENT | 0.77+ |
one place | QUANTITY | 0.77+ |
one super strong platform | QUANTITY | 0.75+ |
a couple of years ago | DATE | 0.74+ |
DevOps | ORGANIZATION | 0.72+ |
Grafana | TITLE | 0.72+ |
CEO | PERSON | 0.69+ |
CUBE | ORGANIZATION | 0.63+ |
reinvent 2020 | TITLE | 0.62+ |
Invent 2020 | TITLE | 0.5+ |
UNLIST TILL 4/2 - Autonomous Log Monitoring
>> Sue: Hi everybody, thank you for joining us today for the virtual Vertica BDC 2020. Today's breakout session is entitled "Autonomous Monitoring Using Machine Learning". My name is Sue LeClaire, director of marketing at Vertica, and I'll be your host for this session. Joining me is Larry Lancaster, founder and CTO at Zebrium. Before we begin, I encourage you to submit questions or comments during the virtual session. You don't have to wait, just type your question or comment in the question box below the slide and click submit. There will be a Q&A session at the end of the presentation and we'll answer as many questions as we're able to during that time. Any questions that we don't address, we'll do our best to answer them offline. Alternatively, you can also go and visit Vertica forums to post your questions after the session. Our engineering team is planning to join the forums to keep the conversation going. Also, just a reminder that you can maximize your screen by clicking the double arrow button in the lower right corner of the slides. And yes, this virtual session is being recorded and will be available for you to view on demand later this week. We'll send you a notification as soon as it's ready. So, let's get started. Larry, over to you. >> Larry: Hey, thanks so much. So hi, my name's Larry Lancaster and I'm here to talk to you today about something that I think who's time has come and that's autonomous monitoring. So, with that, let's get into it. So, machine data is my life. I know that's a sad life, but it's true. So I've spent most of my career kind of taking telemetry data from products, either in the field, we used to call it in the field or nowadays, that's been deployed, and bringing that data back, like log file stats, and then building stuff on top of it. So, tools to run the business or services to sell back to users and customers. And so, after doing that a few times, it kind of got to the point where I was really sort of sick of building the same kind of thing from scratch every time, so I figured, why not go start a company and do it so that we don't have to do it manually ever again. So, it's interesting to note, I've put a little sentence here saying, "companies where I got to use Vertica" So I've been actually kind of working with Vertica for a long time now, pretty much since they came out of alpha. And I've really been enjoying their technology ever since. So, our vision is basically that I want a system that will characterize incidents before I notice. So an incident is, you know, we used to call it a support case or a ticket in IT, or a support case in support. Nowadays, you may have a DevOps team, or a set of SREs who are monitoring a production sort of deployment. And so they'll call it an incident. So I'm looking for something that will notice and characterize an incident before I notice and have to go digging into log files and stats to figure out what happened. And so that's a pretty heady goal. And so I'm going to talk a little bit today about how we do that. So, if we look at logs in particular. Logs today, if you look at log monitoring. So monitoring is kind of that whole umbrella term that we use to talk about how we monitor systems in the field that we've shipped, or how we monitor production deployments in a more modern stack. And so basically there are log monitoring tools. But they have a number of drawbacks. For one thing, they're kind of slow in the sense that if something breaks and I need to go to a log file, actually chances are really good that if you have a new issue, if it's an unknown unknown problem, you're going to end up in a log file. So the problem then becomes basically you're searching around looking for what's the root cause of the incident, right? And so that's kind of time-consuming. So, they're also fragile and this is largely because log data is completely unstructured, right? So there's no formal grammar for a log file. So you have this situation where, if I write a parser today, and that parser is going to do something, it's going to execute some automation, it's going to open or update a ticket, it's going to maybe restart a service, or whatever it is that I want to happen. What'll happen is later upstream, someone who's writing the code that produces that log message, they might do something really useful for me, or for users. And they might go fix a spelling mistake in that log message. And then the next thing you know, all the automation breaks. So it's a very fragile source for automation. And finally, because of that, people will set alerts on, "Oh, well tell me how many thousands of errors are happening every hour." Or some horrible metric like that. And then that becomes the only visibility you have in the data. So because of all this, it's a very human-driven, slow, fragile process. So basically, we've set out to kind of up-level that a bit. So I touched on this already, right? The truth is if you do have an incident, you're going to end up in log files to do root cause. It's almost always the case. And so you have to wonder, if that's the case, why do most people use metrics only for monitoring? And the reason is related to the problems I just described. They're already structured, right? So for logs, you've got this mess of stuff, so you only want to dig in there when you absolutely have to. But ironically, it's where a lot of the information that you need actually is. So we have a model today, and this model used to work pretty well. And that model is called "index and search". And it basically means you treat log files like they're text documents. And so you index them and when there's some issue you have to drill into, then you go searching, right? So let's look at that model. So 20 years ago, we had sort of a shrink-wrap software delivery model. You had an incident. With that incident, maybe you had one customer and you had a monolithic application and a handful of log files. So it's perfectly natural, in fact, usually you could just v-item the log file, and search that way. Or if there's a lot of them, you could index them and search them that way. And that all worked very well because the developer or the support engineer had to be an expert in those few things, in those few log files, and understand what they meant. But today, everything has changed completely. So we live in a software as a service world. What that means is, for a given incident, first of all you're going to be affecting thousands of users. You're going to have, potentially, 100 services that are deployed in your environment. You're going to have 1,000 log streams to sift through. And yet, you're still kind of stuck in the situation where to go find out what's the matter, you're going to have to search through the log files. So this is kind of the unacceptable sort of position we're in today. So for us, the future will not be index and search. And that's simply because it cannot scale. And the reason I say that it can't scale is because it all kind of is bottlenecked by a person and their eyeball. So, you continue to drive up the amount of data that has to be sifted through, the complexity of the stack that has to be understood, and you still, at the end of the day, for MTTR purposes, you still have the same bottleneck, which is the eyeball. So this model, I believe, is fundamentally broken. And that's why, I believe in five years you're going to be in a situation where most monitoring of unknown unknown problems is going to be done autonomously. And those issues will be characterized autonomously because there's no other way it can happen. So now I'm going to talk a little bit about autonomous monitoring itself. So, autonomous monitoring basically means, if you can imagine in a monitoring platform and you watch the monitoring platform, maybe you watch the alerts coming from it or more importantly, you kind of watch the dashboards and try to see if something looks weird. So autonomous monitoring is the notion that the platform should do the watching for you and only let you know when something is going wrong and should kind of give you a window into what happened. So if you look at this example I have on screen, just to take it really slow and absorb the concept of autonomous monitoring. So here in this example, we've stopped the database. And as a result, down below you can see there were a bunch of fallout. This is an Atlassian Stack, so you can imagine you've got a Postgres database. And then you've got sort of Bitbucket, and Confluence, and Jira, and these various other components that need the database operating in order to function. So what this is doing is it's calling out, "Hey, the root cause is the database stopped and here's the symptoms." Now, you might be wondering, so what. I mean I could go write a script to do this sort of thing. Here's what's interesting about this very particular example, and I'll show a couple more examples that are a little more involved. But here's the interesting thing. So, in the software that came up with this incident and opened this incident and put this root cause and symptoms in there, there's no code that knows anything about timestamp formats, severities, Atlassian, Postgres, databases, Bitbucket, Confluence, there's no regexes that talk about starting, stopped, RDBMS, swallowed exception, and so on and so forth. So you might wonder how it's possible then, that something which is completely ignorant of the stack, could come up with this description, which is exactly what a human would have had to do, to figure out what happened. And I'm going to get into how we do that. But that's what autonomous monitoring is about. It's about getting into a set of telemetry from a stack with no prior information, and understanding when something breaks. And I could give you the punchline right now, which is there are fundamental ways that software behaves when it's breaking. And by looking at hundreds of data sets that people have generously allowed us to use containing incidents, we've been able to characterize that and now generalize it to apply it to any new data set and stack. So here's an interesting one right here. So there's a fella, David Gill, he's just a genius in the monitoring space. He's been working with us for the last couple of months. So he said, "You know what I'm going to do, is I'm going to run some chaos experiments." So for those of you who don't know what chaos engineering is, here's the idea. So basically, let's say I'm running a Kubernetes cluster and what I'll do is I'll use sort of a chaos injection test, something like litmus. And basically it will inject issues, it'll break things in my application randomly to see if my monitoring picks it up. And so this is what chaos engineering is built around. It's built around sort of generating lots of random problems and seeing how the stack responds. So in this particular case, David went in and he deleted, basically one of the tests that was presented through litmus did a delete of a pod delete. And so that's going to basically take out some containers that are part of the service layer. And so then you'll see all kinds of things break. And so what you're seeing here, which is interesting, this is why I like to use this example. Because it's actually kind of eye-opening. So the chaos tool itself generates logs. And of course, through Kubernetes, all the log files locations that are on the host, and the container logs are known. And those are all pulled back to us automatically. So one of the log files we have is actually the chaos tool that's doing the breaking, right? And so what the tool said here, when it went to determine what the root cause was, was it noticed that there was this process that had these messages happen, initializing deletion lists, selection a pod to kill, blah blah blah. It's saying that the root cause is the chaos test. And it's absolutely right, that is the root cause. But usually chaos tests don't get picked up themselves. You're supposed to be just kind of picking up the symptoms. But this is what happens when you're able to kind of tease out root cause from symptoms autonomously, is you end up getting a much more meaningful answer, right? So here's another example. So essentially, we collect the log files, but we also have a Prometheus scraper. So if you export Prometheus metrics, we'll scrape those and we'll collect those as well. And so we'll use those for our autonomous monitoring as well. So what you're seeing here is an issue where, I believe this is where we ran something out of disk space. So it opened an incident, but what's also interesting here is, you see that it pulled that metric to say that the spike in this metric was a symptom of this running out of space. So again, there's nothing that knows anything about file system usage, memory, CPU, any of that stuff. There's no actual hard-coded logic anywhere to explain any of this. And so the concept of autonomous monitoring is looking at a stack the way a human being would. If you can imagine how you would walk in and monitor something, how you would think about it. You'd go looking around for rare things. Things that are not normal. And you would look for indicators of breakage, and you would see, do those seem to be correlated in some dimension? That is how the system works. So as I mentioned a moment ago, metrics really do kind of complete the picture for us. We end up in a situation where we have a one-stop shop for incident root cause. So, how does that work? Well, we ingest and we structure the log files. So if we're getting the logs, we'll ingest them and we'll structure them, and I'm going to show a little bit what that structure looks like and how that goes into the database in a moment. And then of course we ingest and structure the Prometheus metrics. But here, structure really should have an asterisk next to it, because metrics are mostly structured already. They have names. If you have your own scraper, as opposed to going into the time series Prometheus database and pulling metrics from there, you can keep a lot more information about metadata about those metrics from the exporter's perspective. So we keep all of that too. Then we do our anomaly detection on both of those sets of data. And then we cross-correlate metrics and log anomalies. And then we create incidents. So this is at a high level, kind of what's happening without any sort of stack-specific logic built in. So we had some exciting recent validation. So Mayadata's a pretty big player in the Kubernetes space. Essentially, they do Kubernetes as a managed service. They have tens of thousands of customers that they manage their Kubernetes clusters for them. And then they're also involved, both in the OpenEBS project, as well as in the Litmius project I mentioned a moment ago. That's their tool for chaos engineering. So they're a pretty big player in the Kubernetes space. So essentially, they said, "Oh okay, let's see if this is real." So what they did was they set up our collectors, which took three minutes in Kubernetes. And then they went and they, using Litmus, they reproduced eight incidents that their actual, real-world customers had hit. And they were trying to remember the ones that were the hardest to figure out the root cause at the time. And we picked up and put a root cause indicator that was correct in 100% of these incidents with no training configuration or metadata required. So this is kind of what autonomous monitoring is all about. So now I'm going to talk a little bit about how it works. So, like I said, there's no information included or required about, so if you imagine a log file for example. Now, commonly, over to the left-hand side of every line, there will be some sort of a prefix. And what I mean by that is you'll see like a timestamp, or a severity, and maybe there's a PID, and maybe there's function name, and maybe there's some other stuff there. So basically that's kind of, it's common data elements for a large portion of the lines in a given log file. But you know, of course, the contents change. So basically today, like if you look at a typical log manager, they'll talk about connectors. And what connectors means is, for an application it'll generate a certain prefix format in a log. And that means what's the format of the timestamp, and what else is in the prefix. And this lets the tool pick it up. And so if you have an app that doesn't have a connector, you're out of luck. Well, what we do is we learn those prefixes dynamically with machine learning. You do not have to have a connector, right? And what that means is that if you come in with your own application, the system will just work for it from day one. You don't have to have connectors, you don't have to describe the prefix format. That's so yesterday, right? So really what we want to be doing is up-leveling what the system is doing to the point where it's kind of working like a human would. You look at a log line, you know what's a timestamp. You know what's a PID. You know what's a function name. You know where the prefix ends and where the variable parts begin. You know what's a parameter over there in the variable parts. And sometimes you may need to see a couple examples to know what was a variable, but you'll figure it out as quickly as possible, and that's exactly how the system goes about it. As a result, we kind of embrace free-text logs, right? So if you look at a typical stack, most of the logs generated in a typical stack are usually free-text. Even structured logging typically will have a message attribute, which then inside of it has the free-text message. For us, that's not a bad thing. That's okay. The purpose of a log is to inform people. And so there's no need to go rewrite the whole logging stack just because you want a machine to handle it. They'll figure it out for themselves, right? So, you give us the logs and we'll figure out the grammar, not only for the prefix but also for the variable message part. So I already went into this, but there's more that's usually required for configuring a log manager with alerts. You have to give it keywords. You have to give it application behaviors. You have to tell it some prior knowledge. And of course the problem with all of that is that the most important events that you'll ever see in a log file are the rarest. Those are the ones that are one out of a billion. And so you may not know what's going to be the right keyword in advance to pick up the next breakage, right? So we don't want that information from you. We'll figure that out for ourselves. As the data comes in, essentially we parse it and we categorize it, as I've mentioned. And when I say categorize, what I mean is, if you look at a certain given log file, you'll notice that some of the lines are kind of the same thing. So this one will say "X happened five times" and then maybe a few lines below it'll say "X happened six times" but that's basically the same event type. It's just a different instance of that event type. And it has a different value for one of the parameters, right? So when I say categorization, what I mean is figuring out those unique types and I'll show an example of that next. Anomaly detection, we do on top of that. So anomaly detection on metrics in a very sort of time series by time series manner with lots of tunables is a well-understood problem. So we also do this on the event types occurrences. So you can think of each event type occurring in time as sort of a point process. And then you can develop statistics and distributions on that, and you can do anomaly detection on those. Once we have all of that, we have extracted features, essentially, from metrics and from logs. We do pattern recognition on the correlations across different channels of information, so different event types, different log types, different hoses, different containers, and then of course across to the metrics. Based on all of this cross-correlation, we end up with a root cause identification. So that's essentially, at a high level, how it works. What's interesting, from the perspective of this call particularly, is that incident detection needs relationally structured data. It really does. You need to have all the instances of a certain event type that you've ever seen easily accessible. You need to have the values for a given sort of parameter easily, quickly available so you can figure out what's the distribution of this over time, how often does this event type happen. You can run analytical queries against that information so that you can quickly, in real-time, do anomaly detection against new data. So here's an example of that this looks like. And this kind of part of the work that we've done. At the top you see some examples of log lines, right? So that's kind of a snippet, it's three lines out of a log file. And you see one in the middle there that's kind of highlighted with colors, right? I mean, it's a little messy, but it's not atypical of the log file that you'll see pretty much anywhere. So there, you've got a timestamp, and a severity, and a function name. And then you've got some other information. And then finally, you have the variable part. And that's going to have sort of this checkpoint for memory scrubbers, probably something that's written in English, just so that the person who's reading the log file can understand. And then there's some parameters that are put in, right? So now, if you look at how we structure that, the way it looks is there's going to be three tables that correspond to the three event types that we see above. And so we're going to look at the one that corresponds to the one in the middle. So if we look at that table, there you'll see a table with columns, one for severity, for function name, for time zone, and so on. And date, and PID. And then you see over to the right with the colored columns there's the parameters that were pulled out from the variable part of that message. And so they're put in, they're typed and they're in integer columns. So this is the way structuring needs to work with logs to be able to do efficient and effective anomaly detection. And as far as I know, we're the first people to do this inline. All right, so let's talk now about Vertica and why we take those tables and put them in Vertica. So Vertica really is an MPP column store, but it's more than that, because nowadays when you say "column store", people sort of think, like, for example Cassandra's a column store, whatever, but it's not. Cassandra's not a column store in the sense that Vertica is. So Vertica was kind of built from the ground up to be... So it's the original column store. So back in the cStor project at Berkeley that Stonebraker was involved in, he said let's explore what kind of efficiencies we can get out of a real columnar database. And what he found was that, he and his grad students that started Vertica. What they found was that what they can do is they could build a database that gives orders of magnitude better query performance for the kinds of analytics I'm talking about here today. With orders of magnitude less data storage underneath. So building on top of machine data, as I mentioned, is hard, because it doesn't have any defined schemas. But we can use an RDBMS like Vertica once we've structured the data to do the analytics that we need to do. So I talked a little bit about this, but if you think about machine data in general, it's perfectly suited for a columnar store. Because, if you imagine laying out sort of all the attributes of an event type, right? So you can imagine that each occurrence is going to have- So there may be, say, three or four function names that are going to occur for all the instances of a given event type. And so if you were to sort all of those event instances by function name, what you would find is that you have sort of long, million long runs of the same function name over and over. So what you have, in general, in machine data, is lots and lots of slowly varying attributes, lots of low-cardinality data that it's almost completely compressed out when you use a real column store. So you end up with a massive footprint reduction on disk. And it also, that propagates through the analytical pipeline. Because Vertica does late materialization, which means it tries to carry that data through memory with that same efficiency, right? So the scale-out architecture, of course, is really suitable for petascale workloads. Also, I should point out, I was going to mention it in another slide or two, but we use the Vertica Eon architecture, and we have had no problems scaling that in the cloud. It's a beautiful sort of rewrite of the entire data layer of Vertica. The performance and flexibility of Eon is just unbelievable. And so I've really been enjoying using it. I was skeptical, you could get a real column store to run in the cloud effectively, but I was completely wrong. So finally, I should mention that if you look at column stores, to me, Vertica is the one that has the full SQL support, it has the ODBC drivers, it has the ACID compliance. Which means I don't need to worry about these things as an application developer. So I'm laying out the reasons that I like to use Vertica. So I touched on this already, but essentially what's amazing is that Vertica Eon is basically using S3 as an object store. And of course, there are other offerings, like the one that Vertica does with pure storage that doesn't use S3. But what I find amazing is how well the system performs using S3 as an object store, and how they manage to keep an actual consistent database. And they do. We've had issues where we've gone and shut down hosts, or hosts have been shut down on us, and we have to restart the database and we don't have any consistency issues. It's unbelievable, the work that they've done. Essentially, another thing that's great about the way it works is you can use the S3 as a shared object store. You can have query nodes kind of querying from that set of files largely independently of the nodes that are writing to them. So you avoid this sort of bottleneck issue where you've got contention over who's writing what, and who's reading what, and so on. So I've found the performance using separate subclusters for our UI and for the ingest has been amazing. Another couple of things that they have is they have a lot of in-database machine learning libraries. There's actually some cool stuff on their GitHub that we've used. One thing that we make a lot of use of is the sequence and time series analytics. For example, in our product, even though we do all of this stuff autonomously, you can also go create alerts for yourself. And one of the kinds of alerts you can do, you can say, "Okay, if this kind of event happens within so much time, and then this kind of an event happens, but not this one," Then you can be alerted. So you can have these kind of sequences that you define of events that would indicate a problem. And we use their sequence analytics for that. So it kind of gives you really good performance on some of these queries where you're wanting to pull out sequences of events from a fact table. And timeseries analytics is really useful if you want to do analytics on the metrics and you want to do gap filling interpolation on that. It's actually really fast in performance. And it's easy to use through SQL. So those are a couple of Vertica extensions that we use. So finally, I would like to encourage everybody, hey, come try us out. Should be up and running in a few minutes if you're using Kubernetes. If not, it's however long it takes you to run an installer. So you can just come to our website, pick it up and try out autonomous monitoring. And I want to thank everybody for your time. And we can open it up for Q and A.
SUMMARY :
Also, just a reminder that you can maximize your screen And one of the kinds of alerts you can do, you can say,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
David | PERSON | 0.99+ |
Larry Lancaster | PERSON | 0.99+ |
David Gill | PERSON | 0.99+ |
Vertica | ORGANIZATION | 0.99+ |
100% | QUANTITY | 0.99+ |
Sue LeClaire | PERSON | 0.99+ |
five times | QUANTITY | 0.99+ |
Larry | PERSON | 0.99+ |
S3 | TITLE | 0.99+ |
three minutes | QUANTITY | 0.99+ |
six times | QUANTITY | 0.99+ |
Sue | PERSON | 0.99+ |
100 services | QUANTITY | 0.99+ |
Zebrium | ORGANIZATION | 0.99+ |
today | DATE | 0.99+ |
three | QUANTITY | 0.99+ |
five years | QUANTITY | 0.99+ |
Today | DATE | 0.99+ |
yesterday | DATE | 0.99+ |
both | QUANTITY | 0.99+ |
Kubernetes | TITLE | 0.99+ |
one | QUANTITY | 0.99+ |
thousands | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
SQL | TITLE | 0.99+ |
one customer | QUANTITY | 0.98+ |
three lines | QUANTITY | 0.98+ |
three tables | QUANTITY | 0.98+ |
each event | QUANTITY | 0.98+ |
hundreds | QUANTITY | 0.98+ |
first people | QUANTITY | 0.98+ |
1,000 log streams | QUANTITY | 0.98+ |
20 years ago | DATE | 0.98+ |
eight incidents | QUANTITY | 0.98+ |
tens of thousands of customers | QUANTITY | 0.97+ |
later this week | DATE | 0.97+ |
thousands of users | QUANTITY | 0.97+ |
Stonebraker | ORGANIZATION | 0.96+ |
each occurrence | QUANTITY | 0.96+ |
Postgres | ORGANIZATION | 0.96+ |
One thing | QUANTITY | 0.95+ |
three event types | QUANTITY | 0.94+ |
million | QUANTITY | 0.94+ |
Vertica | TITLE | 0.94+ |
one thing | QUANTITY | 0.93+ |
4/2 | DATE | 0.92+ |
English | OTHER | 0.92+ |
four function names | QUANTITY | 0.86+ |
day one | QUANTITY | 0.84+ |
Prometheus | TITLE | 0.83+ |
one-stop | QUANTITY | 0.82+ |
Berkeley | LOCATION | 0.82+ |
Confluence | ORGANIZATION | 0.79+ |
double arrow | QUANTITY | 0.79+ |
last couple of months | DATE | 0.79+ |
one of | QUANTITY | 0.76+ |
cStor | ORGANIZATION | 0.75+ |
a billion | QUANTITY | 0.73+ |
Atlassian Stack | ORGANIZATION | 0.72+ |
Eon | ORGANIZATION | 0.71+ |
Bitbucket | ORGANIZATION | 0.68+ |
couple more examples | QUANTITY | 0.68+ |
Litmus | TITLE | 0.65+ |
Tyler Williams & Karthik Subramanian, SAIC | Splunk .conf19
>>Live from Las Vegas. That's the Q covering splunk.com 19 brought to you by Splunk. >>You know, kind of leaning on that heavily. Automation, certainly very important. But what does enterprise and what does enterprise security 6.0 bring to the table. So can you take us through the evolution of where you guys are at with, with Splunk, if you want to handle that enterprise security? So yeah, generally enterprise security has traditionally had really, really good use cases for like the external threats that we're talking about. But like you said, it's very difficult to crack the insider threat part. And so we leveraging machine learning toolkit has started to build that into Splunk to make sure that you know, you can protect your data. And, uh, you know, Tyler and I specifically did this because we saw that there was immaturity in the cybersecurity market for insider threat. And so one of the things that we're actually doing in this top, in addition to talking about what we've done, we're actually giving examples of actionable use cases that people can take home and do themselves. >>Like we're giving them an exact sample code of how to find some outliers. They give me an example of what, so the use case that we go over in the talk is a user logs in at a weird time of day outside of their baseline and they exfiltrate a large amount of data in a low and slow fashion. Um, but they're doing this obviously outside of the scope of their normal behavior. So we give some good searches that you can take home and look at how could I make a baseline, how could I establish that there's deviations from that baseline from a statistical standpoint, and identify this in the future and find the needle in the haystack using the machine learning toolkit. And then if I have a sock that I want to send notables to or some sort of some notification to how do we make that happen, how do we make the transition from machine learning toolkit over to enterprise security or however your SOC operates? >>How do you do that? Do you guys write your own code for that? Or you guys use Splunk? So Splunk has a lot of internal tools and there's a couple of things that need to be pointed out of how to make this happen because we're aggregating large amounts of data. We go through a lot of those finer points in the talk, but sending those through to make sure that they're high confidence is the, is the channel you guys are codifying the cross connect from the machine, learning to the other systems. All right, so I've got to ask, this is basically pattern recognition. You want to look at baselining, how do people, can people hide in that baseline data? So like I'll give you, if I'm saying I'm an evil genius, I say, Hey, I knew these guys looking for Romans anomalies in my baseline, so I'm going to go low and slow in my baseline. >>Can you look for that too? Yeah, there are. There absolutely are ways of, fortunately, uh, there's a lot of different people who are doing research in that space on the defensive side. And so there's a ton of use cases to look at and if you aggregate over a long enough period of time, it becomes incredibly hard to hide. And so the baselines that we recommend building generally look at your 90 day or 120 day out. Um, I guess viewpoint. So you really want to be able to measure that. And most insider threat that happen occur within that 30 to 90 day window. And so the research seems to indicate that those timelines will actually work. Now if you were in there and you read all the code and you did all of the work to see how all of the things come through and you really understood the machine learning minded, I'm sure there's absolutely a way to get in if you're that sophisticated. >>But most of the times they just trying to steal stuff and get out or compromise a system. Um, so is there other patterns that you guys have seen in terms of the that are kind of low hanging fruit priorities that people aren't paying attention to and what's the levels of importance to I guess get ahold of or have some sort of mechanism for managing insider threats? I passwords I've seen one but I mean like there's been a lot of recent papers that have come out in lateral movement and privilege escalation. I think it's an area where a lot of people haven't spent enough time doing research. We've looked into models around PowerShell, um, so that we can identify when a user's maliciously executing PowerShell scripts. I think there's stuff that's getting attention now that when it really needs to, but it is a little bit too late. >>Uh, the community is a bit behind the curve on it and see sharks becoming more of a pattern to seeing a lot more C sharp power shells kind of in hunted down kind of crippled or like identified. You can't operate that way, what we're seeing but, but is that an insider and do that. And do insiders come in with the knowledge of doing C sharp? Those are gonna come from the outside. So I mean, what's the sophistic I guess my question is what's the sophistication levels of an insider threat? Depends on the level a, so the cert inside of dread Institute has aggregated about 15,000 different events. And it could be something as simple as a user who goes in with the intent to do something bad. It could be a person who converted from the inside at any level of the enterprise for some reason. >>Or it could be someone who gets, you know, really upset after a bad review. That might be the one person who has access and he's being socially engineered as well as all kinds of different vectors coming in there. And so, you know, in addition to somebody malicious like that, that you know, there's the accidental, you're phishing campaigns here, somebody's important clicks on an email that they think is from somebody else important or something like that. And you know, we're looking fair for that as well. And that's definitely spear fishing's been very successful. That's a hard one to crack. It is. They have that malware and they're looking at, you can say HR data's out of this guy, just got a bad review, good tennis cinema, a resume or a job opening for, and that's got the hidden code built in. We've seen that move many times. >>Yeah, and natural language processing and more importantly, natural language understanding can be used to get a lot of those cases out. If you're ingesting the text of the email data, well you guys are at a very professional high end from Sai C I mean the history of storied history goes way back and a lot of government contracts do. They do a lot of heavy lifting from anywhere from development to running full big time OSS networks. So there's a lot of history there. What does sustain of the yard? What do you guys look at as state of the art right now in security? Given the fact that you have some visibility into some of the bigger contracts relative to endpoint protection or general cyber, what's the current state of the art? What's, what should people be thinking about or what are you guys excited about? What are some of the areas that is state of the art relative to cyber, cyber security around data usage. >>So, I mean, one of the things, and I saw that there were some talks about it, but not natural language processing and sentiment analysis has gotten, has come a long way. It is much easier to understand, you know, or to have machines understand what, what people are trying to say or what they're doing. And especially, for example, if somebody's like web searching history, you know, and you might think of somebody might do a search for how do I hide downloading a file or something like that. And, and that's something that, well, we know immediately as people, but you know, we have, our customer for example, has 1000000001.2 billion events a day. So you know, if the billion, a billion seconds, that's 30 years. Yeah. So like that's, it's, it's a big number. You know, we, we, we hear those numbers thrown around a lot, but it's a big number to put it in perspective. >>So we're getting that a day and so how do we pick out, it's hard to step of that problem. The eight staff, you can't put stamp on that. Most cutting edge papers that have come out recently have been trying to understand the logs. They're having them machine learning to understand the actual logs that are coming in to identify those anomalies. But that's a massive computation problem. It's a huge undertaking to kind of set that up. Uh, so I really have seen a lot of stuff actually at concierge, some of the innovations that they're doing to optimize that because finding the needle in the haystack is obviously difficult. That's the whole challenge. But there's a lot of work that's being done in Splunk to make that happen a lot faster. And there's some work that's being done at the edge. It's not a lot, but the cutting edge is actually logging and looking at every single log that comes in and understanding it and having a robot say, boom, check that one out. >>Yeah. And also the sentiment, it gets better with the data because we all crushed those billions of events. And you can get a, you know, smiley face or that'd be face depending upon what's happening. It could be, Oh this is bad. But this, this comes back down to the data points you mentioned logs is now beyond logs. I've got tracing other, other signals coming in across the networks. So that's not, that's a massive problem. You need automation, you've got to feed the beast by the machines and you got to do it within whatever computation capabilities you have. And I always say it's a moving train hard. The Target's moving all the time. You guys are standing on top of it. Um, what do you guys think of the event? What's the, what's the most important thing happening here@splunk.com this year? I'd love to have both of you guys take away in on that. >>There's a ton of innovation in the machine learning space. All of the pipelines really that I've, I've been working on in the last year are being augmented and improved by the staff. That's developing content in the machine learning and deep learning space that's belongs. So to me that's by far the most important thing. Your, your take on this, um, between the automation. I know in the last year or so, Splunk has just bought a lot of different companies that do a lot of things that now we can, instead of having to build it ourselves or having to go to three or four different people on top to build a complete solution for the federal government or for whoever your customer is, you can, you know, Splunk is becoming more of a one stop shop. And I think just upgrading all of these things to have all the capabilities working together so that, for example, Phantom, Phantom, you know, giving you that orchestration and automation after. >>For example, if we have an EMS notable events saying, Hey, possible insider threat, maybe they automate the first thing of checking, you know, pull immediately pulling those logs and emailing them or putting them in front of the SOC analyst immediately. So that in, in addition to, Hey, you need to check this person out, it's, you need to check this person out here is the first five pages of what you need to look at. Oh, talking about the impact of that because without that soar feature. Okay. The automation orchestration piece of it, security, orchestration and automation piece of it without where are you know, speed. What's the impact? What's the alternative? Yes. So when we're, right now, when we're giving information to our EES or analysts through yes, they look at it and then they have to click five, six, seven times to get up the tabs that they need to make it done. >>And if we can have those tabs pre populated or just have them, you know, either one click or just come up on their screen for once they open it up. I mean their time is important. Especially when we're talking about an insider threat whom might turn to, yeah, the alternative is five X increase in timespan by the SOC analyst and no one wants that. They want to be called vented with the data ready to go. Ready, alert on it. All right, so final few guys are awesome insights. Walking data upsets right here. Love the inside. Love the love the insights. So final question for the folks watching that are Splunk customers who are not as on the cutting edge, as you guys pioneering this field, what advice would you give them? Like if you had to, you know, shake your friend egg, you know, get off your button, do this, do that. What is the, what do people need to pay attention to that's super urgent that you would implore on them? What would you, what would your advice be once you start that one? >>One of the things that I would actually say is, you know, we can code really cool things. We can do really cool things, but one of the most important things that he and I do as part of our processes before we go to the machine and code, the really cool things. We sometimes just step back and talk for a half an hour talk for an hour of, Hey, what are you thinking about? Hey, what is a thing that you know or what are we reading? What and what are we? And you know, formulating a plan because instead of just jumping into it, if you formulate a plan, then you can come up with you know, better things and augmented and implemented versus a smash and grab on the other side of just, all right, here's the thing, let's let's dump it in there. So you're saying is just for you jump in the data pool and start swimming around, take a step back, collaborate with your peers or get some kind of a game thinking plan. >>We spent a lot of hours, white boarding, but I would to to add to that, it's augment that we spent a lot of time reading the scientific research that's being done by a lot of the teams that are out solving these types of problems. And sometimes they come back and say, Hey, we tried this solution and it didn't work. But you can learn from those failures just like you can learn from the successes. So I recommend getting out and reading. There's a ton of literature in that space around cyber. So always be moving. Always be learning. Always be collaborating. Yeah, it's moving training guys, thanks for the insights Epic session here. Thanks for coming on and sharing your knowledge on the cube, the cube. We're already one big data source here for you. All the knowledge here at.com our seventh year, their 10th year is the cubes coverage. I'm John furry with back after this short break.
SUMMARY :
splunk.com 19 brought to you by Splunk. that into Splunk to make sure that you know, you can protect your So we give some good searches that you can take home and to make sure that they're high confidence is the, is the channel you guys are codifying the cross connect from And so the research seems to indicate so is there other patterns that you guys have seen in terms of the that are kind of low hanging fruit Uh, the community is a bit behind the curve on it and see sharks becoming more of a pattern to And so, you know, in addition to somebody malicious like that, that you know, there's the accidental, Given the fact that you have some visibility into some of the bigger contracts relative to understand, you know, or to have machines understand what, actually at concierge, some of the innovations that they're doing to optimize that because finding the needle in the haystack I'd love to have both of you guys take away in on that. you know, giving you that orchestration and automation after. here is the first five pages of what you need to look at. Like if you had to, you know, shake your friend egg, you know, get off your button, do this, One of the things that I would actually say is, you know, we can code really cool failures just like you can learn from the successes.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
30 years | QUANTITY | 0.99+ |
Karthik Subramanian | PERSON | 0.99+ |
Splunk | ORGANIZATION | 0.99+ |
seventh year | QUANTITY | 0.99+ |
30 | QUANTITY | 0.99+ |
last year | DATE | 0.99+ |
90 day | QUANTITY | 0.99+ |
Tyler Williams | PERSON | 0.99+ |
120 day | QUANTITY | 0.99+ |
Las Vegas | LOCATION | 0.99+ |
Tyler | PERSON | 0.99+ |
10th year | QUANTITY | 0.99+ |
three | QUANTITY | 0.99+ |
four | QUANTITY | 0.99+ |
six | QUANTITY | 0.99+ |
PowerShell | TITLE | 0.99+ |
billion | QUANTITY | 0.99+ |
five | QUANTITY | 0.99+ |
one click | QUANTITY | 0.99+ |
first five pages | QUANTITY | 0.98+ |
both | QUANTITY | 0.98+ |
a day | QUANTITY | 0.98+ |
about 15,000 different events | QUANTITY | 0.98+ |
seven times | QUANTITY | 0.97+ |
half an hour | QUANTITY | 0.97+ |
dread Institute | ORGANIZATION | 0.97+ |
one | QUANTITY | 0.97+ |
billions of events | QUANTITY | 0.96+ |
an hour | QUANTITY | 0.96+ |
a billion seconds | QUANTITY | 0.95+ |
this year | DATE | 0.95+ |
one person | QUANTITY | 0.95+ |
EES | ORGANIZATION | 0.94+ |
eight staff | QUANTITY | 0.93+ |
Target | ORGANIZATION | 0.93+ |
1000000001.2 billion events a day | QUANTITY | 0.93+ |
at.com | ORGANIZATION | 0.93+ |
One | QUANTITY | 0.92+ |
first thing | QUANTITY | 0.9+ |
here@splunk.com | OTHER | 0.9+ |
SOC | ORGANIZATION | 0.87+ |
people | QUANTITY | 0.82+ |
Romans | OTHER | 0.81+ |
five X | QUANTITY | 0.81+ |
John furry | PERSON | 0.75+ |
Splunk .conf19 | OTHER | 0.74+ |
SAIC | ORGANIZATION | 0.73+ |
Phantom | ORGANIZATION | 0.7+ |
one stop shop | QUANTITY | 0.7+ |
literature | QUANTITY | 0.68+ |
one big data | QUANTITY | 0.68+ |
every single log | QUANTITY | 0.67+ |
things | QUANTITY | 0.66+ |
C | PERSON | 0.55+ |
Sai | ORGANIZATION | 0.48+ |
ton | QUANTITY | 0.43+ |
Splunk | TITLE | 0.35+ |
19 | QUANTITY | 0.35+ |
splunk.com | TITLE | 0.34+ |
Vaughn Stewart, Pure Storage & Bharath Aleti, Splunk | Pure Accelerate 2019
>> from Austin, Texas. It's Theo Cube, covering pure storage. Accelerate 2019. Brought to you by pure storage. >> Welcome back to the Cube. Lisa Martin Day Volante is my co host were a pure accelerate 2019 in Austin, Texas. A couple of guests joining us. Next. Please welcome Barack elected director product management for slunk. Welcome back to the Cube. Thank you. And guess who's back. Von Stewart. V. P. A. Technology from pure Avon. Welcome back. >> Hey, thanks for having us guys really excited about this topic. >> We are too. All right, so But we'll start with you. Since you're so excited in your nice orange pocket square is peeking out of your jacket there. Talk about the Splunk, your relationship. Long relationship, new offerings, joint value. What's going on? >> Great set up. So Splunk impure have had a long relationship around accelerating customers analytics The speed at which they can get their questions answered the rate at which they could ingest data right to build just more sources. Look at more data, get faster time to take action. However, I shouldn't be leading this conversation because Split Split has released a new architecture, a significant evolution if you will from the traditional Splunk architectural was built off of Daz and a shared nothing architecture. Leveraging replicas, right? Very similar what you'd have with, like, say, in H D. F s Work it load or H c. I. For those who aren't in the analytic space, they've released the new architecture that's disaggregated based off of cashing and an object store construct called Smart Store, which Broth is the product manager for? >> All right, tell us about that. >> So we release a smart for the future as part of spunk Enterprise. $7 to about a near back back in September Timeframe. Really Genesis or Strong Smart Strong goes back to the key customer problem that we were looking to solve. So one of our customers, they're already ingesting a large volume of data, but the need to retain the data for twice, then one of Peter and in today's architecture, what it required was them to kind of lean nearly scale on the amount of hardware. What we realized it. Sooner or later, all customers are going to run into this issue. But if they want in just more data or reading the data for longer periods, of time, they're going to run into this cost ceiling sooner or later on. The challenge is that into this architecture, today's distributes killer dark picture that we have today, which of all, about 10 years back, with the evolution of the Duke in this particular architecture, the computer and story Jacqui located. And because computer storage acqua located, it allows us to process large volumes of data. But if you look at the demand today, we can see that the demand for storage or placing the demand for computer So these are, too to directly opposite trans that we're seeing in the market space. If you need to basically provide performance at scale, there needs to be a better model. They need a better solution than what we had right now. So that's the reason we basically brought Smart store on denounced availability last September. What's Marceau brings to the table is that a D couples computer and storage, So now you can scale storage independent of computers, so if you need more storage or if you need to read in for longer periods of time, you can just kill independent on the storage and with level age, remote object stores like Bill Flash bid to provide that data depository. But most of your active data said still decides locally on the indexers. So what we did was basically broke the paradigm off computer storage location, and we had a small twist. He said that now the computer stories can be the couple, but you bring comfort and stories closer together only on demand. So that means that when you were running a radio, you know, we're running a search, and whenever the data is being looked for that only when we bring the data together. The other key thing that we do is we have an active data set way ensure that the smart store has ah, very powerful cash manager that allows that ensures that the active data set is always very similar to the time when your laptop, the night when your laptop has active data sets always in the cash always on memory. So very similar to that smarts for cash allows you to have active data set always locally on the index. Start your search performance is not impact. >> Yes, this problem of scaling compute and storage independently. You mentioned H. D. F s you saw it early on there. The hyper converged guys have been trying to solve this problem. Um, some of the database guys like snowflakes have solved it in the cloud. But if I understand correctly, you're doing this on Prem. >> So we're doing this board an on Prem as well as in Cloud. So this smart so feature is already available on tramp were also already using a host all off our spun cloud deployments as well. It's available for customers who want obviously deploy spunk on AWS as well. >> Okay, where do you guys fit in? So we >> fit in with customers anywhere from on the hate say this way. But on the small side, at the hundreds of terabytes up into the tens and hundreds of petabytes side. And that's really just kind of shows the pervasiveness of Splunk both through mid market, all the way up through the through the enterprise, every industry and every vertical. So where we come in relative to smart store is we were a coat co developer, a launch partner. And because our object offering Flash Blade is a high performance object store, we are a little bit different than the rest of the Splunk s story partner ecosystem who have invested in slow more of an archive mode of s tree right, we have always been designed and kind of betting on the future would be based on high performance, large scale object. And so we believe smart store is is a ah, perfect example, if you will, of a modern analytics platform. When you look at the architecture with smart store as brush here with you, you want to suffice a majority of your queries out of cash because the performance difference between reading out a cash that let's say, that's NAND based or envy. Emmy based or obtain, if you will. When you fall, you have to go read a data data out of the Objects store, right. You could have a significant performance. Trade off wean mix significantly minimized that performance drop because you're going to a very high bandwith flash blade. We've done comparison test with other other smart store search results have been published in other vendors, white papers and we show Flash blade. When we run the same benchmark is 80 times faster and so what you can now have without architecture is confidence that should you find yourself in a compliance or regulatory issue, something like Maybe GDP are where you've got 72 hours to notify everyone who's been impacted by a breach. Maybe you've got a cybersecurity case where the average time to find that you've been penetrated occurs 206 days after the event. And now you gotta go dig through your old data illegal discovery, you know, questions around, you know, customer purchases, purchases or credit card payments. Any time where you've got to go back in the history, we're gonna deliver those results and order of magnitude faster than any other object store in the market today. That translates from ours. Today's days, two weeks, and we think that falls into our advantage. Almost two >> orders of magnitude. >> Can this be Flash Player >> at 80%? Sorry, Katie. Time 80 x. Yes, that's what I heard. >> Do you display? Consider what flashlight is doing here. An accelerant of spunk, workloads and customer environment. >> Definitely, because the forward with the smart, strong cash way allow high performance at scale for data that's recites locally in the cash. But now, by using a high performance object store like your flash played. Customers can expect the same high performing board when data is in the cash as well as invented sin. Remorseful >> sparks it. Interesting animal. Um, yeah, you have a point before we >> subjects. Well, I don't want to cut you off. It's OK. So I would say commenting on the performance is just part of the equation when you look at that, UM, common operational activities that a splitting, not a storage team. But a Splunk team has to incur right patch management, whether it's at the Splunk software, maybe the operating system, like linen store windows, that spunk is running on, or any of the other components on side on that platform. Patch Management data Re balancing cause it's unequal. Equally distributed, um, hardware refreshes expansion of the cluster. Maybe you need more computer storage. Those operations in terms of time, whether on smart store versus the classic model, are anywhere from 100 to 1000 times faster with smart store so you could have a deployment that, for example, it takes you two weeks to upgrade all the notes, and it gets done in four hours when it's on Smart store. That is material in terms of your operational costs. >> So I was gonna say, Splunk, we've been watching Splunk for a long time. There's our 10th year of doing the Cube, not our 10th anniversary of our 10th year. I think it will be our ninth year of doing dot com. And so we've seen Splunk emerged very cool company like like pure hip hip vibe to it. And back in the day, we talked about big data. Splunk never used that term, really not widely in its marketing. But then when we started to talk about who's gonna own the big data, that space was a cloud era was gonna be mad. We came back. We said, It's gonna be spunk and that's what's happened. Spunk has become a workload, a variety of workloads that has now permeated the organization, started with log files and security kind of kind of cumbersome. But now it's like everywhere. So I wonder if you could talk to the sort of explosion of Splunk in the workloads and what kind of opportunity this provides for you guys. >> So a very good question here, Right? So what we have seen is that spunk has become the de facto platform for all of one structure data as customers start to realize the value of putting their trying to Splunk on the watch. Your spunk is that this is like a huge differentiate of us. Monk is the read only skim on reed which allows you to basically put all of the data without any structure and ask questions on the flight that allows you to kind of do investigations in real time, be more reactive. What's being proactive? We be more proactive. Was being reactive scaleable platform the skills of large data volumes, highly available platform. All of that are the reason why you're seeing an increase that option. We see the same thing with all other customers as well. They start off with one data source with one use case and then very soon they realize the power of Splunk and they start to add additional use cases in just more and more data sources. >> But this no >> scheme on writer you call scheme on Reed has been so problematic for so many big data practitioners because it just became the state of swamp. >> That didn't >> happen with Splunk. Was that because you had very defined use cases obviously security being one or was it with their architectural considerations as well? >> They just architecture, consideration for security and 90 with the initial use cases, with the fact that the scheme on Reid basically gives open subject possibilities for you. Because there's no structure to the data, you can ask questions on the fly on. You can use that to investigate, to troubleshoot and allies and take remedial actions on what's happening. And now, with our new acquisitions, we have added additional capabilities where we can talk, orchestrate the whole Anto and flow with Phantom, right? So a lot of these acquisitions also helping unable the market. >> So we've been talking about TAM expansion all week. We definitely hit it with Charlie pretty hard. I have. You know, I think it's a really important topic. One of things we haven't hit on is tam expansion through partnerships and that flywheel effect. So how do you see the partners ship with Splunk Just in terms of supporting that tam expansion the next 10 years? >> So, uh, analytics, particularly log and Alex have really taken off for us in the last year. As we put more focus on it, we want to double down on our investments as we go through the end of this year and in the next year with with a focus on Splunk um, a zealous other alliances. We think we are in a unique position because the rollout of smart store right customers are always on a different scale in terms of when they want to adopt a new architecture right. It is a significant decision that they have to make. And so we believe between the combination of flash array for the hot tear and flash played for the cold is a nice way for customers with classic Splunk architecture to modernize their platform. Leverage the benefits of data reduction to drive down some of the cost leverage. The benefits of Flash to increase the rate at which they can ask questions and get answers is a nice stepping stone. And when customers are ready because Flash Blade is one of the few storage platforms in the market at this scale out band with optimized for both NFS and object, they can go through a rolling nondestructive upgrade to smart store, have you no investment protection, and if they can't repurpose that flash rate, they can use peers of service to have the flesh raise the hot today and drop it back off just when they're done within tomorrow. >> And what about C for, you know, big workloads, like like big data workloads. I mean, is that a good fit here? You really need to be more performance oriented. >> So flash Blade is is high bandwith optimization, which really is designed for workload. Like Splunk. Where when you have to do a sparse search, right, we'll find that needle in the haystack question, right? Were you breached? Where were you? Briefed. How were you breached? Go read as much data as possible. You've gotta in just all that data, back to the service as fast as you can. And with beast Cloud blocked, Teresi is really optimized it a tear to form of NAND for that secondary. Maybe transactional data base or virtual machines. >> All right, I want more, and then I'm gonna shut up sick. The signal FX acquisition was very interesting to me for a lot of reasons. One was the cloud. The SAS portion of Splunk was late to that game, but now you're sort of making that transition. You saw Tableau you saw Adobe like rip the band Aid Off and it was somewhat painful. But spunk is it. So I wonder. Any advice that you spend Splunk would have toe von as pure as they make that transition to that sass model. >> So I think definitely, I think it's going to be a challenging one, but I think it's a much needed one in there in the environment that we are in. The key thing is to always because two more focus and I'm sure that you're already our customer focus. But the key is key thing is to make sure that any service is up all the time on make sure that you can provide that up time, which is going to be crucial for beating your customers. Elise. >> That's good. That's good guidance. >> You >> just wanted to cover that for you favor of keeping you date. >> So you gave us some of those really impressive stats In terms of performance. >> They're almost too good to be true. >> Well, what's customer feedback? Let's talk about the real world when you're talking to customers about those numbers. What's the reaction? >> So I don't wanna speak for Broth, so I will say in our engagements within their customer base, while we here, particularly from customers of scale. So the larger the environment, the more aggressive they are to say they will adopt smart store right and on a more aggressive scale than the smaller environments. And it's because the benefits of operating and maintaining the indexer cluster are are so great that they'll actually turn to the stores team and say, This is the new architecture I want. This is a new storage platform and again. So when we're talking about patch management, cluster expansion Harbor Refresh. I mean, you're talking for a large sum. Large installs weeks, not two or 3 10 weeks, 12 weeks on end so it can be. You can reduce that down to a couple of days. It changes your your operational paradigm, your staffing. And so it has got high impact. >> So one of the message that we're hearing from customers is that it's far so they get a significant reduction in the infrastructure spent it almost dropped by 2/3. That's really significant file off our large customers for spending a ton of money on infrastructure, so just dropping that by 2/3 is a significant driver to kind of move too smart. Store this in addition to all the other benefits that get smart store with operational simplicity and the ability that it provides. You >> also have customers because of smart store. They can now actually bursts on demand. And so >> you can think of this and kind of two paradigms, right. Instead of >> having to try to avoid some of the operational pain, right, pre purchase and pre provisional large infrastructure and hope you fill it up. They could do it more of a right sides and kind of grow in increments on demand, whether it's storage or compute. That's something that's net new with smart store um, they can also, if they have ah, significant event occur. They can fire up additional indexer notes and search clusters that can either be bare metal v ems or containers. Right Try to, you know, push the flash, too. It's Max. Once they found the answers that they need gotten through. Whatever the urgent issues, they just deep provisionals assets on demand and return back down to a steady state. So it's very flexible, you know, kind of cloud native, agile platform >> on several guys. I wish we had more time. But thank you so much fun. And Deron, for joining David me on the Cube today and sharing all of the innovation that continues to come from this partnership. >> Great to see you appreciate it >> for Dave Volante. I'm Lisa Martin, and you're watching the Cube?
SUMMARY :
Brought to you by Welcome back to the Cube. Talk about the Splunk, your relationship. if you will from the traditional Splunk architectural was built off of Daz and a shared nothing architecture. What's Marceau brings to the table is that a D couples computer and storage, So now you can scale You mentioned H. D. F s you saw it early on there. So this smart so feature is And now you gotta go dig through your old data illegal at 80%? Do you display? Definitely, because the forward with the smart, strong cash way allow Um, yeah, you have a point before we on the performance is just part of the equation when you look at that, Splunk in the workloads and what kind of opportunity this provides for you guys. Monk is the read only skim on reed which allows you to basically put all of the data without scheme on writer you call scheme on Reed has been so problematic for so many Was that because you had very defined use cases to the data, you can ask questions on the fly on. So how do you see the partners ship with Splunk Flash Blade is one of the few storage platforms in the market at this scale out band with optimized for both NFS And what about C for, you know, big workloads, back to the service as fast as you can. Any advice that you But the key is key thing is to make sure that any service is up all the time on make sure that you can provide That's good. Let's talk about the real world when you're talking to customers about So the larger the environment, the more aggressive they are to say they will adopt smart So one of the message that we're hearing from customers is that it's far so they get a significant And so you can think of this and kind of two paradigms, right. So it's very flexible, you know, kind of cloud native, agile platform And Deron, for joining David me on the
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Lisa Martin | PERSON | 0.99+ |
Dave Volante | PERSON | 0.99+ |
$7 | QUANTITY | 0.99+ |
Katie | PERSON | 0.99+ |
David | PERSON | 0.99+ |
Barack | PERSON | 0.99+ |
two weeks | QUANTITY | 0.99+ |
80 times | QUANTITY | 0.99+ |
ninth year | QUANTITY | 0.99+ |
four hours | QUANTITY | 0.99+ |
Deron | PERSON | 0.99+ |
12 weeks | QUANTITY | 0.99+ |
72 hours | QUANTITY | 0.99+ |
Austin, Texas | LOCATION | 0.99+ |
twice | QUANTITY | 0.99+ |
10th year | QUANTITY | 0.99+ |
Von Stewart | PERSON | 0.99+ |
Elise | PERSON | 0.99+ |
last year | DATE | 0.99+ |
hundreds of terabytes | QUANTITY | 0.99+ |
Today | DATE | 0.99+ |
2019 | DATE | 0.99+ |
today | DATE | 0.99+ |
Vaughn Stewart | PERSON | 0.99+ |
tomorrow | DATE | 0.99+ |
Bharath Aleti | PERSON | 0.99+ |
next year | DATE | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
one | QUANTITY | 0.99+ |
Splunk | ORGANIZATION | 0.99+ |
September | DATE | 0.98+ |
10th anniversary | QUANTITY | 0.98+ |
80% | QUANTITY | 0.98+ |
two | QUANTITY | 0.98+ |
Avon | ORGANIZATION | 0.98+ |
Peter | PERSON | 0.98+ |
Alex | PERSON | 0.98+ |
last September | DATE | 0.98+ |
100 | QUANTITY | 0.98+ |
Jacqui | PERSON | 0.98+ |
Lisa Martin Day Volante | PERSON | 0.98+ |
hundreds of petabytes | QUANTITY | 0.97+ |
Splunk | PERSON | 0.97+ |
Spunk | ORGANIZATION | 0.97+ |
Charlie | PERSON | 0.96+ |
Tableau | TITLE | 0.96+ |
both | QUANTITY | 0.96+ |
206 days | QUANTITY | 0.95+ |
One | QUANTITY | 0.95+ |
Adobe | ORGANIZATION | 0.95+ |
end of this year | DATE | 0.95+ |
two paradigms | QUANTITY | 0.94+ |
about 10 years back | DATE | 0.93+ |
1000 times | QUANTITY | 0.93+ |
Reed | ORGANIZATION | 0.9+ |
one use case | QUANTITY | 0.89+ |
3 10 weeks | QUANTITY | 0.88+ |
Reid | ORGANIZATION | 0.88+ |
90 | QUANTITY | 0.87+ |
couple of guests | QUANTITY | 0.87+ |
Phantom | ORGANIZATION | 0.87+ |
Flash | PERSON | 0.85+ |
2/3 | QUANTITY | 0.84+ |
Marceau | PERSON | 0.83+ |
TAM | ORGANIZATION | 0.83+ |
days | QUANTITY | 0.82+ |
couple | QUANTITY | 0.82+ |
Jeff Mathis, Scalyr & Steve Newman, Scalyr | Scalyr Innovation Day 2019
from San Mateo its the cube covering scalar innovation day brought to you by scaler but I'm John four with the cube we are here in San Mateo California official innovation day at Skylar's headquarters with Steve Neumann the founder of scalar and Jeff Mathis a software engineer guys thanks for joining me today thanks for having us thanks great to have you here so you guys introduced power queries what is all this about yes so the vision for scalar is to become the platform users trust when they want to observe their systems and power queries is a really important step along that journey power queries provide new insights into data with a powerful and expressive query language that's still easy to use so why is this important so we like to scaler we like to think that we're all about speed and a lot of what we're known for is the kind of the raw performance of the query engine that we've built that's sitting underneath this product which is one measure of speed but really we like to think of speed as the time from a question in someone's head to an answer on their screen and so the whole kind of user journey is part of that and you know kind of traditionally in our product we've we provided a set of basic capabilities for searching and counting and graphing that are kind of very easy for people to access and so you can get in quickly pose your question get an answer without even having to learn a query language and and that's been great but there are sometimes the need goes a little bit beyond that the question that some wants to ask is a little bit more complicated or the data needs a little bit of massaging and it just goes beyond the boundaries what you can do in kind of those basic you know sort of basic set of predefined abilities and so that's where we wanted to take a step forward and you know kind of create this more advanced language for for those more advanced cases you know I love the name power query so they want power and it's got to be fast and good so that aside you know queries been around people know search engines search technology discovery finding stuff but as ai/an comes around and more scales and that the system this seems to be a lot more focus on like inference into intuiting what's happening this has been a big trend what do you what's your opinion on that because this has become a big opportunity using data we've seen you know file companies go public we know who they are and they're out there but there's more data coming I mean it's not like it's stopping anytime soon so what's the what's the innovation that that just gonna take power queries to the next level yes so one of the features that I'm really excited about in the future of power queries is our autocomplete feature we've taken a lot of inspiration from just what your navbar does in the browser so the idea is to have a context-sensitive predictive autocomplete feature that's going to take into account a number of individual the syntactic context of where you are in the query what fields you have available to you what fields you've searched recently those kinds of factors Steve what's your take before we get to the customer impact what's the what's the difference it different what's weird whereas power queries gonna shine today and tomorrow so it's some it was a kind of both an interesting and fun challenge for us to design and build this because you're you know we're trying to you know by definition this is for the you know the more advanced use cases the more you know when you need something more powerful and so a big part of the design question for us is how do we how do we let people you know do more sophisticated things with their logs when the when they have that that use case while still making it some you know kind of preserving that that's speed and ease of use that that we like to think we're known for and and in particular you know they've been you know something where you know step one is go you know read this 300 page reference manual and you know learn this complicated query language you know if that was the approach then you know then we would have failed before we started and we had we have the benefit of a lot of hindsight you know there a lot of different sister e of people manipulating data you know working with these sophisticated different and different kinds of systems so there are you know we have users coming to us who are used to working with other other log management tools we have users or more comfortable than SQL we have users who really you know their focus is just a more conventional programming languages especially because you know one of the constituencies we serve our you know it's a trend nowadays that development engineers are responsible also for keeping their code working well in production so they're not experts in this stuff they're not log management experts they're not you know uh telemetry experts and we want them to be able to come in and kind of casual you know coming casually to this tool and get something done but we had all that context of drawn with these different history of languages that people are used to so we came up with about a dozen use cases that we thought kind of covered the spectrum of you know what would people bring bring people into a scenario like this and we actually game to those out well how would you solve this particular question if we were using an SQL like approach or an approach based on this tool or which based on that tool and so we we did this like big exploration and we were able to boil down boil everything down to about ten fairly simple commands that they're pretty much covered the gamut by comparison you know there are there other solutions that have over a hundred commands and it obviously if it's just a lot to learn there at the other end of the spectrum um SQL really does all this with one command select and it's incredibly powerful but you also really have to be a wizard sometimes to kind of shoehorn that into yeah even though sequels out there people know that but people want it easier ultimately machines are gonna be taking over you get the ten commands you almost couldn't get to the efficiency level simplifying the use cases what's the customer scenario looked like what's that why is design important what's what's in it for the customer yeah absolutely so the user experience was a really important focus for us when designing power queries we knew from the start that if tool took you ten minutes to relearn every time you wanted to use it then the query takes ten minutes to execute it doesn't take seconds to execute so one of the ways we approached this problem was to make sure we're constantly giving the user feedback that starts as soon you load the page you've immediately got access to some of the documentation you need you use the feature if you have type in correct syntax you'll get feedback from the system about how to fix that problem and so really focusing on the user experience was a big part of the yeah people gonna factor in the time it takes to actually do the query write it up if you have to code it up and figure it out that's time lag right there you want be as fast as possible interesting design point radical right absolutely so Steve how does it go fast Jeff how does it go fast what are you guys looking at here what's the magic so let me I'm going to step over to the whiteboard shock board here and we'll so chog in one hand Mike in the other will will evaluate my juggling skills but I wanted to start by showing an example of what one of these queries looks like you know I talked about how we kind of boil everything down to about 10 commands so so let's talk through a simple scenario let's say I'm running a tax site you know people come to our web site and they're you know they're putting their taxes together and they're downloading forms and tax laws are different in every state so I have different code that's running for you know you know people in California versus people in Michigan or whatever and I can you know it's easy to do things like graph the overall performance and error rate for my site but I might have a problem with the code for one specific state and it might not show up in those overall statistics very clearly so I don't know I want to get a sense of how well I'm how I am performing for each of the 50 states so I'm gonna and I'm gonna simplify this a little bit but you know I might have an access log for this system where we'll see entries like you know we're loading the tax form and it's for the state of California and the status code was 200 which means that was successful and then we load the tax form and the state is Texas and again that was a success and then we load the tax form for Michigan and the status was a 502 which is a server error and then you know and millions of these mixing with other kinds of logs from other parts of my system and so I want to pull up a report what percentage of requests are succeeding or failing by state and so let me sketch for it first with the query would look like for that and then I'll talk about how how we execute this at speed so so first of all I have to say what which you know of all my other you know I've drawn just the relevant logs but this is gonna be mixed in with all the other logs for my system I need to say which which logs I care about well maybe as simple as just calling out they all have the this page name in them tax form so that that's the first step of my query I'm searching for tax form and now I want to count these count how many of these there are how many of them succeeded or failed and I want to cluster that by state so I'm gonna clustering is with the group command so I'm gonna say I want to count the total number of requests which is just the count so count is a part of the language total is what I'm choosing to name that and I want to count the errors which is also going to be the count command but now I'm going to give it a condition I want to only count where the status is at least 500 and I rather you can see that but behind the plant is a 500 and I'm gonna group that by state so we're we're counting up how many of these values were above 500 and we're grouping it by this field and what's gonna come out of that is a table that'll say for each state the total number of requests the number of errors oh and sorry I actually left out a couple of steps but so it's but actually let's draw what this would give us so far so it's gonna show me for California maybe I had nine thousand one hundred and fifty two requests thirteen of them were errors for Texas I had and so on but I'm still not really there you know that might show me that California had you know maybe California had thirteen errors and Rodi had 12 errors but only there were only 12 requests for Rhode Island Rhode Island is broke you know I've broken my code for Rhode Island but it's only 12 errors because it's a smaller population so that's you know this analysis is still not quite gonna get me where I need to go so I can now add another command I've done this group now I'm gonna say I'm gonna say let which triggers a calculation let error rate equal errors divided by total and so that's going to give me the fraction and so for California you know that might be 0.01 or whatever but for Rhode Island it's gonna be one 100% of the requests are failing and then I can add another command to sort by the error rate and now my problem states are gonna pop to the top so real easy to use language it's great for the data scientists digging in their practitioners you don't need to be hard core coder to get into this exactly that's the idea you know groups or you know very simple commands that just directly you know kind of match the English description of what you're trying to do so then but you know yeah asked a great question then which is how do we take this whole thing and execute it quickly so I'm gonna erase here you're getting into speed now right so yeah bit like that how you get the speed exactly speed is good so simplicity to use I get that it's now speed becomes the next challenge exactly and the speed feeds into the simplicity also because you know step one for anything any tool like this is learning the tool yeah and that involves a lot of trial and error and if the trial and error involves waiting and then at the end of the wait for a query to run you learn that oh you did the query wrong that's very discouraging to people and so we actually think of speed really then becomes some ease of use but all right so how do we actually do this so you've got you know you'll have your whole mass of log data tax forms other forms internal services database logs that are you got your whole you know maybe terabytes of log data somewhere in there are the the really important stuff the tax form errors as well as all the other tax form logs mixed in with a bigger pile of everything else so step one is to filter from that huge pile of all your logs down to just the tax form logs and for that we were able to leverage our existing query engine and one of the main things that makes that engine there's kind of two things that make that that engine as fast it is as it is it's massively parallel so we we segment the data across hundreds of servers our servers so all this data is already distributed across all these servers and once your databases you guys build your own in-house ok got it exactly so this is on our system so we've already collected we're collecting the logs in real time so by the time the user comes and types in that query we already have the data and it's already spread out across all these service then the you know the first step of that query was just a search for tax form and so that's our existing query engine that's not the new thing we've built for power queries so that existing very highly optimized engine this server scans through these logs this service insula these logs each server does its share and they collectively produce a smaller set of data which is just the tax form logs and that's still distributed by the way so really each server is doing this independently and and is gonna continue locally doing the next step so so we're harnessing the horsepower of all these servers each page I only have to work with a small fraction of the data then the next step was that group command we were counting the requests counting the errors and rolling that up by state so that's the new engine we've built but again it each server can do just its little share so this server is gonna take whichever tax form logs it found and produce a little table of counts in it by state this server is gonna do the same thing so at each produce they're a little grouping table with just their share of the logs and then all of that funnels down to one central server where we do the later steps we do the division divide number of errors by total count and and then sort it but by now you know here we might have you might have trillions of log messages down to millions or billions of messages that are relevant to your query now we here we have 50 records you know just one for each state so suddenly the amount of data is very small and so the you know the later steps may be kind of interesting from a processing perspective but they're easy from a speed perspective so you solve a lot of database challenges by understanding kind of how things flow once you've got everything with the columnar database is there just give up perspective of like what if the alternative would be if we this is like I just drew this to a database and I'm running sequel trillions of log files I mean it's not trivial I mean it's a database problem then it's a user problem kind of combine what's order of magnitude difference if I was gonna do the old way yeah so I mean I mean the truth is there's a hundred old ways know how much pain yes they're healthy you know if you're gonna you know if you try to just throw this all into one you know SQL sir you know MySQL or PostgreSQL bytes of data and and by the way we're glossing over the data has to exist but also has to get into the system so you know in you know when you're checking you know am i letting everyone in Rhode Island down on the night before you know the 15th you need up to the moment information but the date you know your database is not necessarily even if it could hold the data it's not necessarily designed to be pulling that in in real time so you know just sort of a simple approach like let me spin up my SQL and throw all the data in it's it's just not even gonna happen I'm gonna have so now you're sharding the data or you're looking at some you know other database solution or ever in it it's a heavy lift either way it's a lot of extra effort taxing on the developers yeah you guys do the heavy lifting yeah okay what's next where's the scale features come in what do you see this evolving for the customers so you know so Jeff talked about Auto complete which you were really excited about because it's gonna again you know a lot of this is for the casual user you know they're you know they're a power user of you know JavaScript or Java or something you're they're building the code and then they've got to come in and solve the problem and get back to what they think of as their real job and so you know we think autocomplete and the way we're doing it we're we're really leveraging both the context of what you're typing as well as the history of what you and your team have done in queried in the past as well as the content of your data every think of it a little bit like the the browser location bar which somehow you type about two letters and it knows exactly which page you're looking for because it's relying on all those different kinds of cues yeah it seems like that this is foundational heavy-lift you myself minimize all that pain then you get the autocomplete start to get in a much more AI machine learning kicks in more intelligent reasoning you start to get a feel for the data it seems like yeah Steve thanks for sharing that there it is on the whiteboard I'm trying for a year thanks for watching this cube conversation
SUMMARY :
small and so the you know the later
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Jeff Mathis | PERSON | 0.99+ |
Jeff | PERSON | 0.99+ |
California | LOCATION | 0.99+ |
ten minutes | QUANTITY | 0.99+ |
Michigan | LOCATION | 0.99+ |
Steve Neumann | PERSON | 0.99+ |
50 records | QUANTITY | 0.99+ |
Rhode Island | LOCATION | 0.99+ |
12 errors | QUANTITY | 0.99+ |
thirteen errors | QUANTITY | 0.99+ |
Steve | PERSON | 0.99+ |
Texas | LOCATION | 0.99+ |
nine thousand | QUANTITY | 0.99+ |
San Mateo | LOCATION | 0.99+ |
millions | QUANTITY | 0.99+ |
Steve Newman | PERSON | 0.99+ |
thirteen | QUANTITY | 0.99+ |
Java | TITLE | 0.99+ |
MySQL | TITLE | 0.99+ |
two things | QUANTITY | 0.99+ |
ten commands | QUANTITY | 0.99+ |
50 states | QUANTITY | 0.99+ |
each page | QUANTITY | 0.99+ |
0.01 | QUANTITY | 0.99+ |
300 page | QUANTITY | 0.99+ |
Rhode Island | LOCATION | 0.99+ |
today | DATE | 0.99+ |
each server | QUANTITY | 0.99+ |
each server | QUANTITY | 0.99+ |
hundreds of servers | QUANTITY | 0.98+ |
500 | QUANTITY | 0.98+ |
first step | QUANTITY | 0.98+ |
over a hundred commands | QUANTITY | 0.98+ |
tomorrow | DATE | 0.98+ |
JavaScript | TITLE | 0.98+ |
Rodi | PERSON | 0.98+ |
502 | OTHER | 0.98+ |
one | QUANTITY | 0.98+ |
step one | QUANTITY | 0.97+ |
Mike | PERSON | 0.97+ |
PostgreSQL | TITLE | 0.97+ |
billions of messages | QUANTITY | 0.97+ |
12 requests | QUANTITY | 0.96+ |
both | QUANTITY | 0.96+ |
100% | QUANTITY | 0.96+ |
each state | QUANTITY | 0.96+ |
200 | OTHER | 0.95+ |
a year | QUANTITY | 0.95+ |
one command | QUANTITY | 0.95+ |
John | PERSON | 0.95+ |
first | QUANTITY | 0.95+ |
about a dozen use cases | QUANTITY | 0.95+ |
about ten fairly simple commands | QUANTITY | 0.95+ |
trillions of log messages | QUANTITY | 0.95+ |
SQL | TITLE | 0.95+ |
English | OTHER | 0.93+ |
about 10 commands | QUANTITY | 0.93+ |
one central server | QUANTITY | 0.92+ |
one measure | QUANTITY | 0.92+ |
above 500 | QUANTITY | 0.9+ |
one of the main things | QUANTITY | 0.89+ |
each | QUANTITY | 0.89+ |
one specific state | QUANTITY | 0.89+ |
15th | QUANTITY | 0.89+ |
scalar innovation day | EVENT | 0.88+ |
Skylar | ORGANIZATION | 0.88+ |
Scalyr | PERSON | 0.84+ |
at least 500 | QUANTITY | 0.84+ |
Shia Liu, Scalyr | Scalyr Innovation Day 2019
>> from San Matteo. It's the Cube covering scaler Innovation Day Brought to you by scaler. >> I'm John for the Cube. We are here in San Mateo, California, for special Innovation Day with scaler at their headquarters. Their new headquarters here. I'm here. She here. Lou, Who's Xia Liu? Who's the software engineering team? Good to see you. Thanks for joining. >> Thank you. >> So tell us, what do you do here? What kind of programming? What kind of engineering? >> Sure. Eso i'ma back and suffer engineer at scaler. What I work on from the day to day basis is building our highly scaleable distributed systems and serving our customers fast queries. >> What's the future that you're building? >> Yeah. So one of the project that I'm working on right now is it will help our infrastructure to move towards a more stateless infrastructure s o. The project itself is a meta data storage component and a series of AP ice that Comptel are back and servers where to find a lock file. That might sound really simple, but at the massive scale of ours, it is actually a significant challenge to do it fast and reliably. >> And you're getting date is a big challenge or run knows that data is the new oil date is the goal. Whatever the people saying, the states is super important. You guys have a unique architecture around data ingest What's so unique about it? You mind sharing? >> Of course, s O. We have a lot of things that we do or don't do. Uniquely. I would like to start with the ingestion front of things and what we don't do on that front. So we don't do keywords indexing which most other extinct existing solutions, too. By not doing that, not keeping the index files up to date with every single log message that's incoming. We saved a lot of time and resource, actually, from the moment that our customers applications generate a logline Teo that logline becoming available to for search in scaler. You y that takes just a couple of seconds on DH on other existing solutions. That can take hours. >> So that's the ingests I What about the query side? Because you got in just now. Query. What's that all about? >> Yeah, of course. Actually. Do you mind if we go to black board a little bit? >> Take a look. >> Okay. Grab a chart real quick. Um, so we have a lot of servers around here. We have, uh, Q >> servers. Let's see. >> These are accused servers and, um, a lot of back and servers, Um, just to reiterate on the interest inside a little bit. When locks come in, they will hit one of these Q servers, and you want them Any one of them. And the Q server will kind of batch the log messages together and then pick one of the bag and servers at random and send the batch of locks. Do them any Q can reach any back in servers. And that's how we kind of were able to handle gigs of laughs. How much ever log that you give us way in jazz? Dozens of terabytes of data on a daily basis. Um, and then it is this same farm of back and servers. That's kind of helping us on the query funds crave front. Um, our goal is when a query comes in, we summon all of these back and servers at once. We get all of their computation powers, all of their CPU cores, to serve this one queer Ari, And that is just a massively scalable multi tenant model and in my mind is really economies of scale at its best. >> So scales huge here. So they got the decoupled back in and accused Q system. But yet they're talking to each other. So what's the impact of the customer? What some of the order of magnitude scale we're talking about here? >> Absolutely. So for on the loch side, we talked about seconds response time from logs being generated, too. They see the lock show up and on the query side, um, the median response time of our queries is under 100 milli second. And we defined that response time from the moment the customer hit in the return button on their laptop to they see results show up and more than 90% of our queries return results in under one second. >> So what's the deployment model for the customers? So I'm a customer. Oh, that sounds great. Leighton sees a huge issue one of low late and seek. His legacy is really the lag issue for data. Do I buy it as a service on my deploying boxes? What does this look like here? >> Nope. Absolutely. Adult were 100 plan cloud native. All of this is actually in our cloud infrastructure and us a customer. You just start using us as a sulfur is a service, and when you submit a query, all of our back and servers are at your service. And what's best about this model is that asks Keller's business girls. We will add more back and servers at more computation power and you as a customer's still get all of that, and you don't need to pay us any extra for the increased queries. >> What's the customer news case for this given you, given example of who would benefit from this? >> Absolutely. So imagine your e commerce platform and you're having this huge black Friday sales. Seconds of time might mean millions of revenues to you, And you don't wantto waste any time on the logging front to debug into your system to look at your monitoring and see where the problem is. If you ever have a problem, so we give you a query response time on the magnitude of seconds versus other is existing solutions. Maybe you need to wait for minutes anxiously in front of your computer. >> She What's the unique thing here? This looks like a really good actor, decoupling things that might make sense. But what's the What's the secret sauce? You? What's the big magic here? >> Yeah, absolutely. So anyone can kind of do a huge server farm Route Fours query approach. But the 1st 80% of a brute force algorithm is easy. It's really the last 20%. That's kind of more difficult, challenging and really differentiate. That's from the rest of others. Solutions. So to start with, we make every effort we can teo identify and skip the work that we don't have to do. S O. Maybe we can come back to your seats. >> Cut. >> Okay, so it's so it's exciting. >> Yeah. So we there are a couple things we do here to skip the work that we don't have to do. As we always say, the fastest queries are those we don't even have to run, which is very true. We have this Colin, our database that wee boat in house highly performance for our use case that can lead us only scan the columns that the customer cares about and skipped all the rest. And we also build a data structure called bloom Filters And if a query term does not occur in those boom filters, we can just skip the whole data set that represents >> so that speed helps on the speed performance. >> Absolutely. Absolutely. If we don't even have to look at that data set, >> You know, I love talking to suffer engineers, people on the cutting edge because, you know, you guys were startup. Attracting talent is a big thing, and people love to work on hard problems. What's the hard problem that you guys are solving here? >> Yeah, absolutely. S o we we have this huge server farm at at our disposal. It's, however, as we always say, the key to brute force algorithms is really to recruit as much force as possible as fast as we can. If you have hundreds thousands, of course lying around. But you don't have an effective way to some of them around when you need them. Then there's no help having them around 11 of the most interesting things that my team does is we developed this customised scatter gather algorithm in order to assign the work in a way that faster back and servers will dynamically compensate for slower servers without any prior knowledge. And I just love that >> how fast is going to get? >> Well, I have no doubt that will one day reach light speed. >> Specialist. Physics is a good thing, but it's also a bottleneck. Just what? Your story. How did you get into this? >> Yeah, s o. I joined Scaler about eight months ago as an ap s server, Actually. Sorry. As an FBI engineer, actually eso during my FBI days. I use scaler, the product very heavily. And it just became increasingly fascinated about the speed at which our queria runs. And I was like, I really want to get behind the scene and see what's going on in the back end. That gives us such fast query. So here I am. Two months ago, I switched the back and team. >> Well, congratulations. And thanks for sharing that insight. >> Thank you, John. Thank >> jumper here with Cuban Sites Day and Innovation Day here in San Mateo. Thanks for watching
SUMMARY :
Day Brought to you by scaler. I'm John for the Cube. basis is building our highly scaleable distributed systems and serving That might sound really simple, but at the massive scale of ours, Whatever the people saying, not keeping the index files up to date with every single log message that's incoming. So that's the ingests I What about the query side? Yeah, of course. so we have a lot of servers around here. And the Q server will kind of batch the log messages together and What some of the order of magnitude scale we're So for on the loch side, we talked about seconds His legacy is really the lag issue for data. for the increased queries. so we give you a query response time on the magnitude of seconds versus She What's the unique thing here? the work that we don't have to do. the work that we don't have to do. If we don't even have to look at that data set, What's the hard problem that you guys are solving here? of the most interesting things that my team does is we developed this customised How did you get into this? behind the scene and see what's going on in the back end. And thanks for sharing that insight. Thanks for watching
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
San Mateo | LOCATION | 0.99+ |
FBI | ORGANIZATION | 0.99+ |
John | PERSON | 0.99+ |
Xia Liu | PERSON | 0.99+ |
Comptel | ORGANIZATION | 0.99+ |
Colin | PERSON | 0.99+ |
Two months ago | DATE | 0.99+ |
Lou | PERSON | 0.99+ |
San Mateo, California | LOCATION | 0.99+ |
more than 90% | QUANTITY | 0.99+ |
Keller | PERSON | 0.98+ |
millions | QUANTITY | 0.98+ |
Cuban Sites Day | EVENT | 0.98+ |
black Friday | EVENT | 0.98+ |
under 100 milli second | QUANTITY | 0.97+ |
1st 80% | QUANTITY | 0.97+ |
Shia Liu | PERSON | 0.97+ |
Dozens of terabytes of data | QUANTITY | 0.96+ |
hundreds thousands | QUANTITY | 0.96+ |
under one second | QUANTITY | 0.95+ |
Innovation Day | EVENT | 0.94+ |
one | QUANTITY | 0.94+ |
Innovation Day | EVENT | 0.92+ |
around 11 | QUANTITY | 0.88+ |
San Matteo | ORGANIZATION | 0.87+ |
Seconds | QUANTITY | 0.85+ |
20% | QUANTITY | 0.83+ |
one day | QUANTITY | 0.83+ |
eight months ago | DATE | 0.81+ |
Scalyr | PERSON | 0.78+ |
Leighton | PERSON | 0.77+ |
Route Fours | OTHER | 0.75+ |
single log message | QUANTITY | 0.75+ |
100 | QUANTITY | 0.74+ |
Scalyr Innovation Day 2019 | EVENT | 0.73+ |
couple of seconds | QUANTITY | 0.73+ |
about | DATE | 0.61+ |
Cube | ORGANIZATION | 0.57+ |
seconds | QUANTITY | 0.56+ |
plan | ORGANIZATION | 0.51+ |
minutes | QUANTITY | 0.49+ |
Scaler | ORGANIZATION | 0.49+ |
scaler | TITLE | 0.38+ |
Jack Gold, Jack Gold & Associates | Citrix Synergy 2019
(upbeat theme music plays) >> Live from Atlanta, Georgia, it's theCube. Covering Citrix Synergy, Atlanta, 2019, brought to you by Citrix. >> Hi, welcome back to theCube. Lisa Martin with Keith Townsend, and we are live in Atlanta, Georgia for Citrix Synergy, 2019. We are pleased to welcome Jack Gold to The Cube, President and founder of Jack Gold & Associates. Jack, it's great to have you join Keith and me this afternoon. >> Thank you for having me. >> So, we had a great day. We've talked to eight or nine folks or so, lot's of really relevant exciting news from Citrix this morning. Talking about the employee experience as, and how I kind of interpreted it, as a catalyst for digital transformation, cultural transformation. You've been working with Citrix for a long time. I'd love to get your perspective on not just what you heard today from Citrix, and with Google and Microsoft, but in the last year or so since they've really kind of done a re-brand effort. What're your thoughts on that? >> Yeah, it's interesting from a Citrix perspective. Citrix, the old Citrix I guess I would put in quotes right, was always known as the VDI company. I've got, you know, the screen that will talk to the server, that will talk to whatever other apps I need it to talk to, and I can have a nice thin client sitting on my desktop and I don't have to spend a lot of money. And I also don't have to worry about if I'm going to bank people stealing stuff off the hard drive, or whatever. They've made a pretty significant transition that was the old work space, if you will. The modern work spaces which is where Citrix is really moving is one where, look we've all grown up with smart phones for the last ten or fifteen years, our kids don't know anything different. They're not going to deal with anything that's complex, anything where I have to log in and out of applications, anything where I have to switch between screens, this just doesn't make any sense for them. And so, what we're seeing Citrix do is move into an environment where, as I said, it's about the modern workspace, it's about being able to help me do my job not getting in my way of me doing my job, and that's really the transition. It's not just Citrix, the industry is moving in that direction as well, but Citrix is really at the forefront of making a lot of that work now. >> So, Jack, talk to us about the new promise of the new Citrix. The, if you remember me, it had to have be about seven years ago, I did a blog post of running Windows XP on your iPad. It was taking, you know, the then desktop solution and running it on your iPad. >> (Jack) Sure. >> And it was a cool trick. But we talked about, today, we would hope by today, that mobile technology would of forced companies to rewrite applications, for a mobile first experience. But that simply hasn't happened. So, presenting a bad application on to a mobile dot, to a mobile work station, or a mobile device, doesn't work. We end up packing in, trying it, and squeezing, and trying to get our work done, how is Citrix promising to change that experience, even versus their competitors? >> Sure. Well, first of all so two bad's don't make a good. Right. Having a bad app on a bad device doesn't make it good. >> (Keith) Right. >> Doesn't make it easy to use, doesn't help me get my job done. What we really are talking about, now, is the ability to build a workspace. Something where I sit and look at, that helps me get my job done, as opposed to getting in the way. Which means that, instead of having to punch fourteen different holes, or you know, icons and sit at my keyboard and type forty-eight different commands and do thirty-eight different log-ins as each one is different, and by the way I couldn't remember them so I just called the help desk in-between, and that's another half an hour of my time that I didn't want to, that I wasted. >> (laughs) Give me my word perfect templates. >> (Lisa laughs) >> (Jack) There you go, there you go, word perfect I remember that no so well. I remember it well not so nicely. What we're really trying to focus on now is user experience, right. What we're really trying to focus on is if, if you wanted to get your work done, I want to make it easy. Think about it as going to a grocery store. If you can't, if you've got a list of groceries and you can't find what you want in five minutes, you leave, you go somewhere else. You go to another grocery store where things are much easier to find. It's the same at work, or it should be the same at work. Now, that said, a lot of apps and organizations, especially big enterprises where they have, some can have literally thousands of apps, are not going away. The notion that everything is going to go into the modern workspace, where everything looks like a phone, it's a nice idea, it's properly not going to happen. Legacy apps will be legacy apps for a very long time, it's like mainframes are dead, guess what, they're still around. That said, that doesn't mean that you can't take some of those legacy apps and make them easier to use with the proper front-end. And that's really what Citrix is trying to do with the workspaces, and other's again, it's not just Citrix in this, we have to be fair there are lots people working in this space. But, if you can make the front-end workspace more attractive, easier to use, easier to navigate, even if I've got old, clunky stuff in the background. For me as a user, you can give me back fifteen, twenty, thirty minutes a day, an hour a day, that's really productivity. Look, if you're paying me a hundred dollars an hour, and you save me an hour a day you just made a hundred dollars every day that I'm working at that company, that sounds like a lot, but there are people who make that kind of money. Or even a fifty or twenty-five dollars, it all adds up. And so, what we're really doing is trying to move into an environment where if I can make you more productive but making things more easier for you to navigate, and getting in and our of applications more quickly, getting more information to me more quickly, which makes the overall organization more productive because I'm sharing more information with you, then that's a real win-win, and that's where I think Citrix is really trying to position itself, and doing a fairly good job at doing that. Clearly they don't have all of the components yet, but then no one does. This is an ongoing process. >> So, employee experience is table-stakes for any business, as we look at the modern workforce it's highly disrupted. >> (Jack) Yes. >> It's composed of five different generations. >> (Jack) Yes. >> Who have varying expertise with technology. It is also demanding because we're all consumers. >> (Jack) Yes. >> And so we have this expectation, or this, yeah I'd say expectation that I want to be able to go in and have this personalized experience. I don't want to have to become an expert in sales-force because I might need to understand, can I talk to that costumer and ask them to be a reference? How much time are you going to take? But this personalization is becoming more and more critical as we see this influence from the consumer side. >> Right. >> Were some of the things that you heard today from Citrix, what are your thoughts on how their going to be able to improve that more personalized employee experience? >> So people think of personalization, I think sometimes, too narrowly. For some people personalization is, you know, I've got my phone out, and I have the apps that I want on my phone and that's personalization. I think of it a little bit differently. We need to extend personalization. When I'm at work, what I want is not just the apps I want, clearly I want those, right, but also the ability, to get help with those apps as I need it, right. And so where Citrix is going is trying to put intelligence into the system, so that when I'm interacting with back-end solutions or my neighbors, or with teams collaboration, I get the assistance I need to make it easier for me to do that work. It's not just the apps, it's also help with the apps. And if we can do that, that's really what we want. We go, you know, if I have a problem with my laptop I'm going to come to you and say, hey, you did this yesterday what was the result, can you help me for five minutes? Five minutes is never five minutes, it's usually an hour and a half, but still. I'll come to you. Why can't I have an app on my desk that does the same thing? I'm having trouble. Help me. Fix it. Let me know what I'm doing wrong, or let me know how I can do it better. And that's where Citrix is trying to go with the analytics that they've got in place. Which is huge, I think they're underplaying that, because I think that the whole analytic space in making things easier for people to use, because in understanding where my problems are is huge, and that's going to pick up. The notion of having a nice pretty, pretty may not be the right word, but attractive at least, workspace for me to go in that doesn't get frustrated, frustration is a killer in productivity, as everyone knows. There are examples I've heard multiple people tell me now that they go out and hire, especially with millennials, that go out an hire twenty or thirty new employees, and half of them quit within a week because their systems are so bad that they get so frustrated that they're not going to work there. So, the notion of having a modern workspace where I get the applications that I need, I get the assistance I need, because of the analytics of that backend telling the systems what I need, and making it easier for me to do it. And then allowing me to be productive not just for myself, but for the organization, is where we all need to go and I think that Citrix is making some real progress going that way. >> (Keith) Well Jack, we're talking about products that haven't quite been released yet, so I'm trying to get a sense or, worth's the right built versus buy stage, in complexity Citrix should be? You know, I can make it apple pie by going out and picking the apple. >> (Jack laughs) Right. >> And making my own crust or, I can go buy filling, or I can just go buy any mince pie, stick it in the oven and warm it up. Three very different experiences. Three different layers of investment, and outcomes frankly. In this world, I can go hire application developers to write these many apps, to write these customizations, to write these integrations, but that's, I think that's akin to picking the apple and that just simply doesn't scale. But, also while any mince pie is okay every now and again, I want, you know, something of higher quality. Where do you think Citrix is on the kind of range of built versus buy with this intelligent experience? >> So built versus buy is a very interesting phenomenon. And it's interesting because a lot of it has to do with where you think you are right now in the world, right. You know you mention going out and getting developers and building your own, that's all well and good, it doesn't scale, and by the way in today's market you can't find them to begin with. So you often don't even have a choice. So that's number one. Number two is that there are companies out there that still think for competitive advantage that they have to do everything from scratch, like building your pie. Yes, you probably make the best pie in the world, but guess what, sometimes a good enough pie is good enough. Right, and if you're in business sometimes good enough is the only way you survive. It doesn't have to be a hundred percent perfect, ninety percent's okay too. People can deal with that. So that's the other piece. The third piece of it is, from an end-user perspective, right, if end-users are accustom to having an interaction in a certain way and the you go out and get developers that come in and do it, something completely different, which they're apt to do because each will have their own kind of flavor to it, then you just force them to learn one, two, three, four, five different interface interactions I'm not going to do that. I'm going to get frustrated as heck, and I'm going to go call the help desk or I'm going to go get my app and say go do this for me. Both of which are counterproductive to the company and to me. So, it really depends on where you are in the stage of where your company is, I would say built versus buy it's not a one or a zero. There's lots of shades of gray in between, it's also not all or nothing. So, some applications might be built internally, some you may want to buy externally, some you may have a hybrid, and the nice thing about where workspaces is going now is that you plug all of those into the same environment. That's really the ultimate goal, is to make it as easy and transparent for the organization as possible, and also for the user because the user ultimately is the end consumer. And if it's not good for the end consumer, it's not good for the company either. >> (Lisa) So delivering this great game-changing customer experience for this, as we talked about before this distributed modern work force that wants to be able to access mobile apps, Sass apps. >> (Jack) Right. >> Web apps from tablets, PC, phone, desktop. >> (Jack) Your car, your refrigerator >> Exactly. >> (Jack) Anything with a screen on it. >> Oh yeah, the refrigerators. Wherever you are, I think, okay people >> (Jack) Sure. >> We're people, and we are the biggest single security threat there is. >> (Jack laughs) >> So in your perspective, how is what Citrix is talking about balancing security as an essential component of this employee experience? >> So there are a few things, number one is a lot of companies think that if they limit the end user experience they're more secure. The truth of the matter is, yes, I mean if you don't let me get in to an app I can't steal application or information, or lose it somehow. But I also can't get my work done. So there's a balance between security and privacy which many companies don't talk about which is not exactly the same thing, there are two unique things, more and more privacy is becoming as big or bigger an issue than security, but you know we can get at that in a minute. But, the notion of security really relates to what I was talking about earlier which is analytics. If I know what you're suppose to be doing, you're here at Synergy. If someone just got your credentials and logged in from Los Angeles or New York or Chicago or Denver or wherever, I know it's not you. I can shut that thing down very quickly and not have to worry about them stealing information, also if you're, if I know you're not suppose to be in a certain version of SAP, you're not suppose to be doing some ERP system and you're in it, then again the analytics tells me that there's something going on, there's something anomalous going on that I need to investigate. So, having a system that protects because there's a kind of a front end to everything that's going on in the back end, and a realization of what's going on behind that screen gives me a much higher sense of security from a corporate perspective, it's not perfect there is no such thing as perfect security, but it's a lot better than just letting us kind of do our own thing, and loading, you know, semantic or McAfee or whatever on your PC. And that's where the industry ultimately has to go. That becomes part of the new modern workspace. It's not just about more productive it's about more secure. It's about more private. It's about not letting information escape that shouldn't be there to begin with. >> (Keith) So last question on data grabbing. Because we haven't talked about data and data is, you know, probably the most important thing in this topic. The importance of the (unintelligible) and Google announcement. You know, we, the yottabyte, the first time I've heard that term, yottabye of data that data's going to be spread across the world and this, this ideal of centralized compute and us being able to present, compute into data centers, no longer going to work, that we're going to have to, applications are going to be spread across the world. Where do you Citrix advancing that discipline of providing apps where they need to be with these relationships? >> So, it's an interesting phenomenon what we're going through right now, if you look back a couple of decades ago everything was centralized, people were centralized, they all work in one building, computing was centralized it was all in the data center, IT was centralized, it was all, you know, working around the servers. The Cloud is the opposite direction, although I would argue The Cloud isn't new, The Cloud is just time-share in a different environment, for us old people who remember the old IBM time-share computers. But everything is becoming distributed, data is distributed, people are distributed, applications are distributed, networks are distributed, you name it. The key critical factor for companies in keeping their productivity, keeping up the productivity is to make sure that the distributed environment doesn't get in the way of doing work. So you've got things like latency, if it takes me, if I'm in. (crowd cheers) >> They're having a party behind us. >> No, they agree with you! >> (Jack laughs) Yes, apparently. I, you know, if I'm here at Synergy but I have to work back at my offices near Boston, I can't wait five minutes for information to come back and forth, it's like the old days. Latency now has to be within five microseconds or people get frustrated, so that becomes a network issue, applications, same way, if I have to go to a data center, the data isn't local to my server here, it has to go to London, I'm not going to wait three minutes for it to come back like we use to, or ten minutes or an hour and a half. Or come back the next morning. You know, you want to book a flight on an airline, are you going to wait thirty minutes for them to find you a seat? You're going to go to another airline. So the whole notion of distributed means that it's very different now, even though it's distributed, everything is local. And by local, keeping it local means that you have to have latency below a certain point (crowd cheers) so that I don't realize that it's distributed, or I don't care that it's distributed. Yottabyte's of data means that we're going to have data everywhere, accessible all the time, and we're going to produce data like crazy. You know, a typical car, an autonomous car will produce a gigabyte of data every minute. Hundreds every, you know, hour. So, the amount of data is going to be fantastic that we have to deal with. Then, the big question becomes, okay so, I can't personally deal with all this data, it's impossible, I have to have the assistance, the intelligence within the system to go off and make something of that data so that I can actually interact with it in a meaningful fashion. That's where Citrix would like to go, that's where other's would like to go. They can't do it alone, because the problem is just too darn big. But, it will, we will get there, companies will get there eventually, not all of them perhaps, only the ones that are going to be successful long term are going to get there. >> Well, Jack, I wish we had more time to chat with you. This has, I just feel like going dot, dot, dot, to be continued. And I want to say, coincidence, I don't know, there were two rounds of applause when you talked about latency. (Keith laughs) >> There we go. They're just waiting for the bar to open, it's taking too long. >> (Lisa laughs) You think that's what it is? >> (Jack) Properly. >> All right well we'll get you over there, and thank-you again for joining Keith and me this afternoon. >> Thank-you very much. >> (Lisa) Our pleasure. For Keith Townsend, I'm Lisa Martin, you're watching theCube live from Citrix Synergy, 2019. Thanks for watching. (upbeat theme music plays)
SUMMARY :
brought to you by Citrix. Jack, it's great to have you join Keith and me not just what you heard today from Citrix, and with They're not going to deal with anything that's complex, you know, the then desktop solution and running it on your how is Citrix promising to change that experience, Having a bad app on a bad device is the ability to build a workspace. and make them easier to use with the proper front-end. So, employee experience is table-stakes for Who have varying expertise with technology. to that costumer and ask them to be a reference? I'm going to come to you and say, hey, you did this yesterday make it apple pie by going out and picking the apple. and again, I want, you know, something of higher quality. is the only way you survive. to access mobile apps, Sass apps. Wherever you are, We're people, and we are the biggest single But, the notion of security really relates to what I was The importance of the is to make sure that the distributed environment doesn't So, the amount of data is going to be fantastic to be continued. it's taking too long. All right well we'll get you over there, and thank-you For Keith Townsend, I'm Lisa Martin, you're watching theCube
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Keith | PERSON | 0.99+ |
Keith Townsend | PERSON | 0.99+ |
twenty | QUANTITY | 0.99+ |
Jack | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.99+ | |
five minutes | QUANTITY | 0.99+ |
Chicago | LOCATION | 0.99+ |
Denver | LOCATION | 0.99+ |
ten minutes | QUANTITY | 0.99+ |
Boston | LOCATION | 0.99+ |
three minutes | QUANTITY | 0.99+ |
Citrix | ORGANIZATION | 0.99+ |
ninety percent | QUANTITY | 0.99+ |
an hour and a half | QUANTITY | 0.99+ |
thirty minutes | QUANTITY | 0.99+ |
iPad | COMMERCIAL_ITEM | 0.99+ |
New York | LOCATION | 0.99+ |
eight | QUANTITY | 0.99+ |
Jack Gold | PERSON | 0.99+ |
Los Angeles | LOCATION | 0.99+ |
London | LOCATION | 0.99+ |
2019 | DATE | 0.99+ |
Five minutes | QUANTITY | 0.99+ |
fifteen | QUANTITY | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Synergy | ORGANIZATION | 0.99+ |
an hour and a half | QUANTITY | 0.99+ |
Atlanta, Georgia | LOCATION | 0.99+ |
Both | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
Lisa | PERSON | 0.99+ |
Citrix Synergy | ORGANIZATION | 0.99+ |
Windows XP | TITLE | 0.99+ |
third piece | QUANTITY | 0.99+ |
two rounds | QUANTITY | 0.99+ |
yesterday | DATE | 0.99+ |
two unique things | QUANTITY | 0.99+ |
McAfee | ORGANIZATION | 0.98+ |
hundred percent | QUANTITY | 0.98+ |
Atlanta | LOCATION | 0.98+ |
an hour a day | QUANTITY | 0.98+ |
each | QUANTITY | 0.98+ |
two | QUANTITY | 0.98+ |
last year | DATE | 0.98+ |
first experience | QUANTITY | 0.98+ |
next morning | DATE | 0.98+ |
five microseconds | QUANTITY | 0.98+ |
one | QUANTITY | 0.97+ |
half an hour | QUANTITY | 0.97+ |
Daniel Berg, IBM Cloud & Norman Hsieh, LogDNA | KubeCon 2018
>> Live from Seattle, Washington it's theCUBE, covering KubeCon and CloudNativeCon North America 2018. Brought to you by Red Hat, the Cloud Native Computing Foundation, and its ecosystem partners. >> Hey, welcome back everyone, it's theCUBE live here in Seattle for day three of three of wall-to-wall coverage. We've been analyzing here on theCUBE for three days, talking to all the experts, the CEOs, CTOs, developers, startups. I'm John Furrier, Stu Miniman, with theCUBE coverage of here at dock, not DockerCon, KubeCon and CloudNativeCon. Getting down to the last Con. >> So close, John, so close. >> Lot of Docker containers around here. We'll check it on the Kubernetes. Our next two guests got a startup, hot startup here. You got Norman Hsieh, head of business development, LogDNA. New compelling solution on Kubernetes give them a unique advantage, and of course, Daniel Berg who's distinguished engineer at IBM. They have a deal. We're going to talk about the startup and the deal with IBM. The highlights, kind of a new model, a new world's developing. Thanks for joining us. >> Yeah, no problem, thanks for having us. >> May get you on at DockerCon sometimes. (Daniel laughing) Get you DockerCon. The container certainly been great, talk about your product first. Let's get your company out there. What do you guys do? You got something new and different. Something needed. What's different about it? >> Yeah, so we started building this product. One thing we were trying to do is finding a login solution that was built for developers, especially around DevOps. We were running our own multi-tenant SaaS product at the time and we just couldn't find anything great. We tried open source Elastic and it turned out to be a lot to manage, there was a lot of configuration we had to do. We tried a bunch of the other products out there which were mostly built for log analysis, so you'd analyze logs, maybe a week or two after, and there was nothing just realtime that we wanted, and so we decided to build our own. We overcame a lot of challenges where we just felt that we could build something that was easier to use than what was out there today. Our philosophy is for developers in the terms of we want to make it as simple as possible. We don't want you to manage where you're going to think about how logs work today. And so, the whole idea, even you can go down to some of the integrations that we have, our Kubernetes integration's two lines. You essentially hit two QCTL lines, your entire cluster will get logged, directly logged in in seconds. That's something we show often times at demos as well. >> Norman, I wonder if you can drill in a little bit more for us. Always look at is a lot of times the new generation, they've got just new tools to play with and new things to do. What was different, what changes? Just the composability and what a small form factor. I would think that you could just change the order of magnitude in some of the pricing of some of these. Tell us why it's different. >> Yeah, I mean, I think there's, three major things was speed. So what we found was that there weren't a lot of solutions that were optimized really, really well for finding logs. There were a lot of log solutions out there, but we wanted to optimize that so we fine-tuned Elasticsearch. We do a lot of stuff around there to make that experience really pleasurable for our users. The other is scale. So we're noticing now is if you kind of expand on the world of back in the day we had single machines that people got logs off of, then you went to VMware where you're taking a single machine and splitting up to multiple different things, and now you have containers, and all of a sudden you have Kubernetes, you're talking about thousands and thousands of nodes running and large production service. How do you find logs in those things? And so we really wanted to build for that scale and that usability where, for Kubernetes, we'll automatically tag all your logs coming through. So you might get a single log line, but we'll tag it with all the meta-data you need to find exactly what you want. So if I want to, if my container dies and I no longer know that containers around, how am I going to get the logs off of that, well, you can go to LogDNA, find the container that you're looking for, know exactly where that error's coming from as well. >> So you're basically storing all this data, making it really easy for the integration piece. Where does the IBM relationship fit in? What's the partnership? What are you guys doing together? >> I don't know if Dan wants to-- >> Go ahead, go ahead. >> Yeah, so we're partnering with IBM. We are one of their major partners for login. So if you go into Observability tab under IMB Cloud and click on Login, login is there, you can start the login instance. What we've done is, IBM's brought us a great opportunity where we could take our product and help benefit their own customers and also IBM themselves with a lot of the login that we do. They saw that we are very simplistic way of thinking about logs and it was really geared towards when you think about IBM Cloud and the shift that they're moving towards, which is really developer-focused, it was a really, really good match for us. It brought us the visibility into the upmarket with larger customers and also gives us the ability to kind of deploy globally across IBM Cloud as well. >> I mean, IBMs got a great channel on the sales side too, and you guys got a great relationship. We've seen that playbook before where I think we've interviewed in all the other events with IBM. Startups can really, if they fit in with IBM, it's just massive, but what's the reason? Why the partnership? Explain. >> Well, I mean, first of all we were looking for a solution, a login solution, that fit really well with IKS, our Kubernetes service. And it's cloud-native, high scale, large number of cluster, that's what our customers are building. That's what we want to use internally as well. I mean, we were looking for a very robust cloud-native login service that we could use ourselves, and that's when we ran across these guys. What, about a year ago? >> Yeah, I mean, I think we kind of first got introduced at last year's KubeCon and then it went to Container World, and we just kept seeing each other. >> And we just kept on rolling with it so what we've done with that integration, what's nice about the integration, is it's directly in the catalog. So it's another service in the catalog, you go and select it, and provision it very easily. But what's really cool about it is we wanted to have that integration directly with the Kubernetes services as well, so there's the tab on the Integration tab on the Kubernetes, literally one button, two lines of code that you just have to execute, bam! All your logs are now streaming for the entire cluster with all the index and everything. It just makes it a really nice, rich experience to capture your logs. >> This is infrastructure as code, that's what the promise was. >> Absolutely, yes. >> You have very seamless integration and the backend just works. Now talk about the Kubernetes pieces. I think this is fascinating 'cause we've been pontificating and evaluating all the commentary here in theCUBE, and we've come to the conclusion that cloud's great, but there's other new platform-like things emerging. You got Edge and all these things, so there's a whole new set, new things are going to come up, and it's not going to be just called cloud, it's going to be something else. There's Edge, you got cameras, you got data, you got all kinds of stuff going on. Kubernetes seems to fit a lot of these emerging use cases. Where does the Kubernetes fit in? You say you built on Kubernetes, just why is that so important? Explain that one piece. >> Yeah, I mean, I think there's, Kubernetes obviously brought a lot of opportunities for us. The big differentiator for us was because we were built on Kubernetes from the get go, we made that decision a long time ago, we didn't realize we could actually deploy this package anywhere. It didn't have to be, we didn't have to just run as a multi-tenant SaaS product anymore and I think part of that is for IBM, their customers are actually running, when they're talking about an integrated login service, we're actually running on IBM Cloud, so their customers can be sure that the data doesn't actually move anywhere else. It's going to stay in IBM Cloud and-- >> This is really important and because they're on the Kubernetes service, it gives them the opportunity, running on Kubernetes, running automatic service, they're going to be able to put LogDNA in each of the major regions. So customer will be able to keep their logged data in the regions that they want it to stay. >> Great for compliance. >> Absolutely. >> I mean, compliance, dreams-- >> Got to have it. >> Especially with EU. >> How about search and discovery, that's fit in too? Just simple, what's your strategy on that? >> Yeah, so our strategy is if you look at a lot of the login solutions out there today, a lot of times they require you to learn complex query languages and things like that. And so the biggest thing we were hearing was like, man, onboarding is really hard because some of our developers don't look at logs on a daily basis. They look at it every two weeks. >> Jerry Chen from Greylock Ventures said machine learning is the new, ML is the new SQL. >> Yup. (Daniel laughing) >> To your point, this complex querying is going to be automated away. >> Yup. >> Yes. >> And you guys agree with that. >> Oh, yeah. >> You actually, >> Totally agree with that. >> you talked about it on our interview. >> Norman, wonder if you can bring us in a little bit of compliance and what discussions you're having with customers. Obviously GDPR, big discussion point we had. We've got new laws coming from California soon. So how important is this to your customers, and what's the reality kind of out there in your user base? >> Yeah, compliance was, our founders had run a lot of different businesses before. They had two major startups where they worked with eBay, compliance was the big thing, so we made a decision early on to say, hey, look, we're about 50 people right now, let's just do compliance now. I've been at startups where we go, let's just keep growing and growing and we'll worry about compliance later-- >> Yeah, bite you in the ass, big time. >> Yeah, we made a decision to say, hey, look, we're smaller, let's just implement all the processes and necessary needs, so. >> Well, the need's there too, that's two things, right? I mean, get it out early. Like security, build it up front and you got it in. >> Exactly. >> And remember earlier we were talking and I was telling you how within the Kubernetes service we like to use our own services to build expertise? It's the same thing here. Not only are they running on top of IKS, we're using LogDNA to manage the logs and everything, and cross the infrastructure for IKS as well. So we're heavily using it. >> This also highlights, Daniel, the ecosystem dynamic of having when you break down this monolithic type of environments and their sets of services, you benefit because you can tap into a startup, they can tap in to IBM's goodness. It's like somewhat simple Biz Dev deal other than the RevShare component of the sales, but technically, this is what customers want at the endgame is they want the right tool, the right job, the right product. If it comes from a startup, you guys don't have to build it. >> I mean, exactly. Let the experts do it, we'll integrate it. It's a great relationship. And the teams work really well together which is fantastic. >> What do you guys do with other startups? If a startup watches and says, hey, I want to be like LogDNA. I want to plug into IBM's Cloud. I want to be just like them and make all that cash. What do they got to do? What's the model? >> I mean, we're constantly looking at startups and new business opportunities obviously. We do this all the time. But it's got to be the right fit, alright? And that's important. It's got to be the right fit with the technology, it's got to be the right fit as far as culture, and team dynamics of not only my team but the startup's teams and how we're going to work together, and this is why it worked really great with LogDNA. I mean, everything, it just all fit, it all made sense, and it had a good business model behind that as well. So, yes, there's opportunities for others but we have to go through and explore all those. >> So, Norman, wonder if you can share, how's your experience been at the show here? We'd love to hear, you're going to have so many startups here. You got record-setting attendance for the show. What were your expectations coming in? What are the KPIs you're measuring with and how has it met what you thought you were going to get? >> No, it's great, I mean, previous to the last year's KubeCon we had not really done any events. We're a small company, we didn't want to spend the resources, but we came in last year and I think what was refreshing was people would talk to us and we're like, oh, yeah, we're not an open source technology, we're actually a log vendor and we can, and we'll-- (Stu laughing) So what we said was, hey, we'll brush that into an experience, and people were like, oh, wow, this is actually pretty refreshing. I'm not configuring my fluentd system, fluentd to tap into another Elasticsearch. There was just not a lot of that. I think this year expectation was we need the size doubled. We still wanted to get the message out there. We knew we were hot off the presses with the IMB public launch of our service on IBM Cloud. And I think we we're expecting a lot. I mean, we more than doubled what our lead count was and it's been an amazing conference. I mean, I think the energy that you get and the quality of folks that come by, it's like, yeah, everybody's running Kubernetes, they know what they're talking about, and it makes that conversation that much easier for us as well. >> Now you're CUBE alumni now too. It's the booth, look at that. (everyone laughing) Well, guys, thanks for coming on, sharing the insight. Good to see you again. Great commentary, again, having distinguished engineering, and these kinds of conversations really helps the community figure out kind of what's out there, so I appreciate that. And if everything's going to be on Kubernetes, then we should put theCUBE on Kubernetes. With these videos, we'll be on it, we'll be out there. >> Hey, yeah, absolutely, that'd be great. >> TheCUBE covers day three. Breaking it down here. I'm John Furrier, Stu Miniman. That's a wrap for us here in Seattle. Thanks for watching and look for us next year, 2019. That's a wrap for 2018, Stu, good job. Thanks for coming on, guys, really appreciate it. >> Thanks. >> Thank you. >> Thanks for watching, see you around. (futuristic instrumental music)
SUMMARY :
Brought to you by Red Hat, the CEOs, CTOs, developers, startups. We're going to talk about the startup and the deal with IBM. What do you guys do? And so, the whole idea, even you can go down and new things to do. and all of a sudden you have Kubernetes, What are you guys doing together? about IBM Cloud and the shift that they're moving towards, and you guys got a great relationship. Well, I mean, first of all we were looking for a solution, Yeah, I mean, I think we kind of first got introduced And we just kept on rolling with it so what we've done that's what the promise was. and it's not going to be just called cloud, It didn't have to be, we didn't have to just run in each of the major regions. And so the biggest thing we were hearing was like, machine learning is the new, ML is the new SQL. is going to be automated away. you talked about it So how important is this to your customers, so we made a decision early on to say, Yeah, we made a decision to say, and you got it in. And remember earlier we were talking and I was telling you of having when you break down this monolithic type And the teams work really well together which is What do you guys do It's got to be the right fit with the technology, and how has it met what you thought you were going to get? I mean, I think the energy that you get Good to see you again. Hey, yeah, absolutely, That's a wrap for us here in Seattle. see you around.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
IBM | ORGANIZATION | 0.99+ |
Jerry Chen | PERSON | 0.99+ |
Daniel Berg | PERSON | 0.99+ |
Norman Hsieh | PERSON | 0.99+ |
Norman | PERSON | 0.99+ |
Seattle | LOCATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
Cloud Native Computing Foundation | ORGANIZATION | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
California | LOCATION | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
eBay | ORGANIZATION | 0.99+ |
John | PERSON | 0.99+ |
two lines | QUANTITY | 0.99+ |
last year | DATE | 0.99+ |
Dan | PERSON | 0.99+ |
Greylock Ventures | ORGANIZATION | 0.99+ |
2018 | DATE | 0.99+ |
Daniel | PERSON | 0.99+ |
three days | QUANTITY | 0.99+ |
KubeCon | EVENT | 0.99+ |
Elastic | TITLE | 0.99+ |
One | QUANTITY | 0.99+ |
IBMs | ORGANIZATION | 0.99+ |
two things | QUANTITY | 0.99+ |
Seattle, Washington | LOCATION | 0.99+ |
DockerCon | EVENT | 0.99+ |
LogDNA | ORGANIZATION | 0.99+ |
two guests | QUANTITY | 0.98+ |
one piece | QUANTITY | 0.98+ |
IMB | ORGANIZATION | 0.98+ |
Stu | PERSON | 0.98+ |
IKS | ORGANIZATION | 0.98+ |
single machines | QUANTITY | 0.98+ |
single machine | QUANTITY | 0.98+ |
IBM Cloud | ORGANIZATION | 0.98+ |
IMB Cloud | TITLE | 0.97+ |
one button | QUANTITY | 0.97+ |
Kubernetes | TITLE | 0.97+ |
two | QUANTITY | 0.97+ |
each | QUANTITY | 0.96+ |
one | QUANTITY | 0.96+ |
CUBE | ORGANIZATION | 0.96+ |
CloudNativeCon | EVENT | 0.96+ |
today | DATE | 0.94+ |
CloudNativeCon North America 2018 | EVENT | 0.94+ |
single log line | QUANTITY | 0.93+ |
KubeCon 2018 | EVENT | 0.93+ |
thousands | QUANTITY | 0.92+ |
first | QUANTITY | 0.91+ |
GDPR | TITLE | 0.91+ |
about 50 people | QUANTITY | 0.91+ |
Container World | ORGANIZATION | 0.91+ |
day three | QUANTITY | 0.9+ |
this year | DATE | 0.9+ |
two major startups | QUANTITY | 0.9+ |
three | QUANTITY | 0.89+ |
Edge | TITLE | 0.88+ |
DevOps | TITLE | 0.88+ |
EU | ORGANIZATION | 0.87+ |
about a year ago | DATE | 0.86+ |
a week | QUANTITY | 0.86+ |
Elasticsearch | TITLE | 0.85+ |
Jon Rooney, Splunk | Splunk .conf18
>> Announcer: Live from Orlando, Florida. It's theCube. Covering .conf18, brought to you by Splunk. >> We're back in Orlando, Dave Vellante with Stu Miniman. John Rooney is here. He's the vice president of product marketing at Splunk. Lot's to talk about John, welcome back. >> Thank you, thanks so much for having me back. Yeah we've had a busy couple of days. We've announced a few things, quite a few things, and we're excited about what we're bringing to market. >> Okay well let's start with yesterday's announcements. Splunk 7.2 >> Yup. _ What are the critical aspects of 7.2, What do we need to know? >> Yeah I think first, Splunk Enterprise 7.2, a lot of what we wanted to work on was manageability and scale. And so if you think about the core key features, the smart storage, which is the ability to separate the compute and storage, and move some of that cool and cold storage off to blob. Sort of API level blob storage. A lot of our large customers were asking for it. We think it's going to enable a ton of growth and enable a ton of use cases for customers and that's just sort of smart design on our side. So we've been real excited about that. >> So that's simplicity and it's less costly, right? Free storage. >> Yeah and you free up the resources to just focus on what are you asking out of Splunk. You know running the searches and the safe searches. Move the storage off to somewhere else and when you need it you pull it back when you need it. >> And when I add an index or I don't have to both compute and storage, I can add whatever I need in granular increments, right? >> Absolutely. It just enables more graceful and elastic expansiveness. >> Okay that's huge, what else should we know about? >> So workload management, which again is another manageability and scale feature. It's just the ability to say the great thing about Splunk is you put your data in there and multiple people can ask questions of that data. It's just like an apartment building that has ... You know if you only have one hot water heater and a bunch of people are taking a shower at the same time, maybe you want to give some privileges to say you know, the penthouse they're going to get the hot water first. Other people not so much. And that's really the underlying principle behind workload management. So there are certain groups and certain people that are running business critical, or mission critical, searches. We want to make sure they get the resources first and then maybe people that are experimenting or kind of kicking the tires. We have a little bit of a gradation of resources. >> So that's essentially programmatic SLAs. I can set those policies, I can change them. >> Absolutely, it's the same level of granular control that say you were on access control. It's the same underlying principle. >> Other things? Go ahead. >> Yeah John just you guys always have some cool, pithy statements. One of the things that jumped out to me in the keynotes, because it made me laugh, was the end of metrics. >> John: Yes. >> You've been talking about data. Data's at the ... the line I heard today was Splunk users are at the crossroads of data so it gives a little insight about what you're doing that's different ways of managing data 'cause every company can interact with the same data. Why is the Splunk user, what is it different, what do they do different, and how is your product different? >> Yeah I mean absolutely. I think the core of what we've always done and Doug talked about it in the keynote yesterday is this idea of this expansive, investigative search. The idea that you're not exactly sure what the right question is so you want to go in, ask a question of the data, which is going to lead you to another question, which is going to lead you to another question, and that's that finding a needle in a pile of needles that Splunk's always great at. And we think of that as more the investigative expansive search. >> Yeah so when I think back I remember talking with companies five years ago when they'd say okay I've got my data scientists and finding which is the right question to ask once I'm swimming in the data can be really tough. Sounds like you're getting answers much faster. It's not necessarily a data scientist, maybe it is. We say BMW on stage. >> Yeah. >> But help us understand why this is just so much simpler and faster. >> Yeah I mean again it's the idea for the IT and security professionals to not necessarily have to know what the right question is or even anticipate the answer, but to find that in an evolving, iterative process. And the idea that there's flexibility, you're in no way penalized, you don't have to go back and re-ingest the data or do anything to say when you're changing exactly what your query is. You're just asking the question which leads to another question, And that's how we think about on the investigative side. From a metric standpoint, we do have additional ... The third big feature that we have in Splunk Enterprise 7.2 is an improved metrics visualization experience. Is the idea of our investigative search which we think we are the best in the industry at. When you're not exactly sure what you're looking for and you're doing a deep dive, but if you know what you're looking for from a monitoring standpoint you're asking the same question again and again and again, over and again. You want be able to have an efficient and easy way to track that if you're just saying I'm looking for CPU utilization or some other metric. >> Just one last follow up on that. I look ... the name of the show is .conf >> Yes. >> Because it talks about the config file. You look at everywhere, people are in the code versus gooey and graphical and visualization. What are you hearing from your user base? How do you balance between the people that want to get in there versus being able to point and click? Or ask a question? >> Yeah this company was built off of the strength of our practitioners and our community, so we always want to make sure that we create a great and powerful experience for those technical users and the people that are in the code and in the configuration files. But you know that's one of the underlying principles behind Splunk Next which was a big announcement part of day one is to bring that power of Splunk to more people. So create the right interface for the right persona and the right people. So the traditional Linux sys admin person who's working in IT or security, they have a certain skill set. So the SPL and those things are native to them. But if you are a business user and you're used to maybe working in Excel or doing pivot tables, you need a visual experience that is more native to the way you work. And the information that's sitting in Splunk is valuable to you we just want to get it to you in the right way. And similar to what we talked about today in the keynote with application developers. The idea of saying well everything that you need is going to be delivered in a payload and json objects makes a lot of sense if you're a modern application developer. If you're a business analyst somewhere that may not make a lot of sense so we want to be able to service all of those personas equally. >> So you've made metrics a first class citizen. >> John: Absolutely. >> Opening it up to more people. I also wanted to ask you about the performance gains. I was talking to somebody and I want to make sure I got these numbers right. It was literally like three orders of magnitude faster. I think the number was 2000 times faster. I don't know if I got that number right, it just sounds ... Implausible. >> That's specifically what we're doing around the data fabric search which we announced in beta on day one. Simply because of the approach to the architecture and the approach to the data ... I mean Splunk is already amazingly fast, amazingly best in class in terms of scale and speed. But you realize that what's fast today because of the pace and growth of data isn't quite so fast two, three, four years down the road. So we're really focused looking well into the future and enabling those types of orders of magnitude growth by completely re imagining and rethinking through what the architecture looks like. >> So talk about that a little bit more. Is that ... I was going to say is that the source of the performance gain? Is it sort of the architecture, is it tighter code, was it a platform do over? >> No I mean it wasn't a platform do over, it's just the idea that in some cases the idea of thinking like I'm federating a search between one index here and one index there, to have a virtualization layer that also taps into compute. Let's say living in a patchy Kafka, taking advantage of those sorts of open source projects and open source technologies to further enable and power the experiences that our customers ultimately want. So we're always looking at what problems our customers are trying to solve. How do we deliver to them through the product and that constant iteration, that constant self evaluation is what drives what we're doing. >> Okay now today was all about the line of business. We've been talking about, I've used the term land and expand about a hundred times today. It's not your term but others have used it in the industry and it's really the template that you're following. You're in deep in sec ops, you're in deep in IT, operations management, and now we're seeing just big data permeate throughout the organization. Splunk is a tool for business users and you're making it easier for them. Talk about Splunk business flow. >> Absolutely, so business flow is the idea that we had ... Again we learned from our customers. We had a couple of customers that were essentially tip of the spear, doing some really interesting things where as you described, let's say the IT department said well we need to pull in this data to check out application performance and those types of things. The same data that's following through is going to give you insight into customer behavior. It's going to give you insight into coupons and promotions and all the things that the business cares about. If you're a product manager, if you're sitting in marketing, if you're sitting in promotions, that's what you want to access and you want to be able to access that in real time. So the challenge is that we're now stepping you with things like business flow is how do you create an interface? How do you create an experience that again matches those folks and how they think about the world? The magic, the value that's sitting in the data is we just have to surface it for the right way for the right people. >> Now the demo, Stu knows I hate demos, but the demo today was awesome. And I really do, I hate demos because most of them are just so boring but this demo was amazing. You took a bunch of log data and a business user ingested it and looked at it and it was just a bunch of data. >> Yeah. >> Like you'd expect and go eh what am I supposed to do with this and then he pushed button and then all of a sudden there was a flow chart and it showed the flow of the customer through the buying pattern. Now maybe that's a simpler use case but it was still very powerful. And then he isolated on where the customer actually made a phone call to the call center because you want to avoid if possible and then he looked at the percentage of drop outs, which was like 90% in that case, versus the percentage of drop outs in a normal flow which was 10%- Oop something's wrong, drilled in, fixed the problem. He showed how he fixed it, oh graphically beautiful. Is it really that easy? >> Yeah I mean I think if you think about what we've done in computing over the last 40 years. If you think about even the most basic word processor, the most basic spreadsheet work, that was done by trained technicians 30-40 years ago. But the democratization of data created this notion of the information worker and we're a decade or so now plus into big data and the idea that oh that's only highly trained professionals and scientists and people that have PHDs. There's always going to be an aspect of the market or an aspect of the use cases that is of course going to be that level of sophistication, but ultimately this is all work for an information worker. If you're an information worker, if you're responsible for driving business results and looking at things, it should be the same level of ease as your traditional sort of office suite. >> So I want to push on that a little if I can. So and just test this, because it looked so amazingly simple. Doug Merritt made the point yesterday that business processes they used to be codified. Codifying business processes is a waste of time because business processes are changing so fast. The business process that you used in the example was a very linear process, admittedly. I'm going to search for a product, maybe read a review, I'm going to put it in my cart, I'm going to buy it. You know, very straightforward. But business processes as we know are unpredictable now. Can that level of simplicity work and the data feed in some kind of unpredictable business process? >> Yeah and again that's our fundamental difference. How we've done it differently than everyone in the market. It's the same thing we did with IT surface intelligence when we launched that back in 2015 because it's not a tops down approach. We're not dictating, taking sort of a central planning approach to say this is what it needs to look like. The data needs to adhere to this structure. The structure comes out of the data and that's what we think. It's a bit of a simplification, but I'm a marketing guy and I can get away with it. But that's where we think we do it differently in a way that allows us to reach all these different users and all these different personas. So it doesn't matter. Again that business process emerges from the data. >> And Stu, that's going to be important when we talk about IOT but jump in here. >> Yeah so I wanted to have you give us a bit of insight on the natural language processing. >> John: Yeah natural language processing. >> You've been playing with things like the Alexa. I've got a Google Home at home, I've got Alexa at home, my family plays with it. Certain things it's okay for but I think about the business environment. The requirements in what you might ask Alexa to ask Splunk seems like that would be challenging. You're got a global audience. You know, languages are tough, accents are tough, syntax is really really challenging. So give us the why and where are we. Is this nascent things? Do you expect customers to really be strongly using this in the near future? >> Absolutely. The notion of natural language search or natural language computing has made huge strides over the last five or six years and again we're leveraging work that's done elsewhere. To Dave's point about demos ... Alexa it looks good on stage. Would we think, and if you're to ask me, we'll see. We'll always learn from the customers and the good thing is I like to be wrong all the time. These are my hypotheses, but my hypothesis is the most actual relevant use of that technology is not going to be speech it's going to be text. It's going to be in Slack or Hipchat where you have a team collaborating on an issue or project and they say I'm looking for this information and they're going to pass that search via text into Splunk and back via Slack in a way that's very transparent. That's where I think the business cases are going to come through and if you were to ask me again, we're starting the betas we're going to learn from our customers. But my assumption is that's going to be much more prevalent within our customer base. >> That's interesting because the quality of that text presumably is going to be much much better, at least today, than what you get with speech. We know well with the transcriptions we do of theCUBE interviews. Okay so that's it. ML and MLP I thought I heard 4.0, right? >> Yeah so we've been pushing really hard on the machine learning tool kit for multiple versions. That team is heavily invested in working with customers to figure out what exactly do they want to do. And as we think about the highly skilled users, our customers that do have data scientists, that do have people that understand the math to go in and say no we need to customize or tweak the algorithm to better fit our business, how do we allow them essentially the bare metal access to the technology. >> We're going to leave dev cloud for Skip if that's okay. I want to talk about industrial IOT. You said something just now that was really important and I want to just take a moment to explain to the audience. What we've seen from IOT, particularly from IT suppliers, is a top down approach. We're going to take our IT framework and put it at the edge. >> Yes. >> And that's not going to work. IOT, industrial IOT, these process engineers, it's going to be a bottoms up approach and it's going to be standard set by OT not IT. >> John: Yes. >> Splunk's advantage is you've got the data. You're sort of agnostic to everything else. Wherever the data is, we're going to have that data so to me your advantage with industrial IOT is you're coming at it from a bottoms up approach as you just described and you should be able to plug into the IOT standards. Now having said that, a lot of data is still analog but that's okay you're pulling machine data. You don't really have tight relationships with the IOT guys but that's okay you got a growing ecosystem. >> We're working on it. >> But talk about industrial IOT and we'll get into some of the challenges. >> Yeah so interestingly we first announced the Industrial Asset Intelligence product at the Hannover Messe show in Germany, which is this massive like 300,000 it's a city, it's amazing. >> I've been, Hannover. One hotel, huge show, 400,000 people. >> Lot of schnitzel (laughs) I was just there. And the interesting thing is it's the first time I'd been at a show really first of all in years where people ... You know if you go to an IT or security show they're like oh we know Splunk, we love Splunk, what's in the next version. It was the first time we were having a lot of people come up to us saying yeah I'm a process engineer in an industrial plant, what's Splunk? Which is a great opportunity. And as you explain the technology to them their mindset is very different in the sense they think of very custom connectors for each piece. They have a very, almost bespoke or matched up notion, of a sense to a piece of equipment. So for an example they'll say oh do you have a connector for and again, I don't have the machine numbers, but like the Siemens 123 machine. And I'll be like well as long as it's textural structural to semi structural data ideally with a time stamp, we can ingest and correlate that. Okay but then what about the Siemens ABC machine? Well the idea that, the notion that ... we don't care where the source is as long as there's a sensor sending the data in a format that we can consume. And if you think back to the beginning of the data stream processor demo that Devani and Eric gave yesterday that showed the history over time, the purple boxes that were built, like we can now ingest data via multiple inputs and via multiple ways into Splunk. And that hopefully enables the IOT ecosystems and the machine manufacturers, but more importantly, the sensor manufacturers because it feels like in my understanding of the market we're still at a point of a lot of folks getting those sensors instrumented. But once it's there and essentially the faucet's turned on, we can pull it all in and we can treat it and ingest it just as easily as we can data from AWS Kineses or Apache Access logs or MySequel logs. >> Yeah and so instrumenting the windmill, to use the metaphor, is not your job. Connectivity to the windmill is not your job, but once those steps have been taken and the business takes those steps because there's a business case, once that's done then the data starts flowing and that's where you come in. >> And there's a tremendous amount of incentive in the industry right now to do that level of instrumentation and connectivity. So it feels like that notion of instrument connect then do the analytics, we're sitting there well positioned once all those things are in place to be one of the top providers for those analytics. >> John I want to ask you something. Stu and I were talking about this at our kickoff and I just want to clarify it. >> Doug Merritt said that he didn't like the term unstructured data. I think that's what he said yesterday, it's just data. My question is how do you guys deal with structured data because there is structured data. Bringing transaction processing data and analytics data together for whatever reason. Whether it's fraud detection, to give the buyer an offer before you lose them, better customer service. How do you handle that kind of structured data that lives in IBM mainframes or whatever. USS mainframes in the case of Carnival. >> Again we want to be able to access data that lives everywhere. And so we've been working with partners for years to pull data off mainframes. Again, the traditional in outs aren't necessarily there but there are incentives in the market. We work with our ecosystem to pull that data to give it to us in a format that makes sense. We've long been able to connect to traditional relational databases so I think when people think of structured data they think about oh it's sitting in a relational database somewhere in Oracle or MySequel or SQL Server. Again, we can connect to that data and that data is important to enhance things particularly for the business user. Because if the log says okay whatever product ID 12345, but the business user needs to know what product ID 12345 is and has a lookup table. Pull it in and now all of a sudden you're creating information that's meaningful to you. But structure again, there's fluidity there. Coming from my background a Json object is structured. You can the same way Theresa Vu in the demo today unfurled in the dev cloud what a Json object looks like. There's structure there. You have key value pairs. There's structure to key value pairs. So all of those things, that's why I think to Doug's point, there's fluidity there. It is definitely a continuum and we want to be able to add value and play at all ends of that continuum. >> And the key is you guys your philosophy is to curate that data in the moment when you need it and then put whatever schema you want at that time. >> Absolutely. Going back to this bottoms up approach and how we approach it differently from basically everyone else in the industry. You pull it in, we take the data as is, we're not transforming or changing or breaking the data or trying to put it into a structure anywhere. But when you ask it a question we will apply a structure to give you the answer. If that data changes when you ask that question again, it's okay it doesn't break the question. That's the magic. >> Sounds like magic. 16,000 customers will tell you that it actually works. So John thanks so much for coming to theCUBE it was great to see you again. >> Thanks so much for having me. >> You're welcome. Alright keep it right there everybody. Stu and I will be back. You're watching theCUBE from Splunk conf18 #splunkconf18. We'll be right back. (electronic drums)
SUMMARY :
brought to you by Splunk. He's the vice president of product marketing at Splunk. and we're excited about what we're bringing to market. Okay well let's start with yesterday's announcements. _ What are the critical aspects of 7.2, and move some of that cool and cold storage off to blob. So that's simplicity and it's less costly, right? Move the storage off to somewhere else and when you need it It just enables more graceful and elastic expansiveness. It's just the ability to say the great thing about Splunk is So that's essentially programmatic SLAs. Absolutely, it's the same level of granular control that Other things? One of the things that jumped out to me in the keynotes, Why is the Splunk user, what is it different, and Doug talked about it in the keynote yesterday is ask once I'm swimming in the data can be really tough. But help us understand why this is just so much And the idea that there's flexibility, you're in no way I look ... the name of the show is You look at everywhere, people are in the code versus So the SPL and those things are native to them. I also wanted to ask you about the performance gains. Simply because of the approach to the architecture and Is it sort of the architecture, is it tighter code, it's just the idea that in some cases the idea of and it's really the template that you're following. So the challenge is that we're now stepping you with things but the demo today was awesome. made a phone call to the call center because it should be the same level of ease as your traditional The business process that you used in the example It's the same thing we did with IT surface intelligence And Stu, that's going to be important when we talk about Yeah so I wanted to have you give us a bit of insight The requirements in what you might ask Alexa to ask Splunk It's going to be in Slack or Hipchat where you have a team That's interesting because the quality of that text bare metal access to the technology. We're going to take our IT framework and put it at the edge. And that's not going to work. Wherever the data is, we're going to have that data some of the challenges. Industrial Asset Intelligence product at the I've been, Hannover. And that hopefully enables the IOT ecosystems and the Yeah and so instrumenting the windmill, once all those things are in place to be one of the top John I want to ask you something. Doug Merritt said that he didn't like the term but the business user needs to know what product ID 12345 is curate that data in the moment when you need it to give you the answer. it was great to see you again. Stu and I will be back.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Doug Merritt | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
John | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Orlando | LOCATION | 0.99+ |
John Rooney | PERSON | 0.99+ |
90% | QUANTITY | 0.99+ |
Jon Rooney | PERSON | 0.99+ |
Germany | LOCATION | 0.99+ |
2015 | DATE | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Doug | PERSON | 0.99+ |
Excel | TITLE | 0.99+ |
Splunk | ORGANIZATION | 0.99+ |
10% | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
Orlando, Florida | LOCATION | 0.99+ |
yesterday | DATE | 0.99+ |
Stu | PERSON | 0.99+ |
Theresa Vu | PERSON | 0.99+ |
2000 times | QUANTITY | 0.99+ |
BMW | ORGANIZATION | 0.99+ |
400,000 people | QUANTITY | 0.99+ |
each piece | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
Hannover | LOCATION | 0.99+ |
Eric | PERSON | 0.99+ |
three | QUANTITY | 0.99+ |
Devani | PERSON | 0.99+ |
one index | QUANTITY | 0.99+ |
four years | QUANTITY | 0.99+ |
16,000 customers | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
300,000 | QUANTITY | 0.98+ |
first time | QUANTITY | 0.98+ |
one | QUANTITY | 0.98+ |
One hotel | QUANTITY | 0.97+ |
Siemens | ORGANIZATION | 0.97+ |
SQL Server | TITLE | 0.97+ |
30-40 years ago | DATE | 0.96+ |
five years ago | DATE | 0.96+ |
both | QUANTITY | 0.96+ |
One | QUANTITY | 0.95+ |
Linux | TITLE | 0.95+ |
Hannover Messe | EVENT | 0.95+ |
one hot water heater | QUANTITY | 0.94+ |
first | QUANTITY | 0.94+ |
Splunk | TITLE | 0.94+ |
Kafka | TITLE | 0.94+ |
Alexa | TITLE | 0.92+ |
three orders | QUANTITY | 0.92+ |
Oracle | ORGANIZATION | 0.92+ |
day one | QUANTITY | 0.91+ |
.conf | OTHER | 0.87+ |
#splunkconf18 | EVENT | 0.86+ |
MySequel | TITLE | 0.86+ |
third big feature | QUANTITY | 0.85+ |
Alex Scarsini, Edgewater Markets | Blockchainweek NYC 2018
>> Announcer: From New York, it's The Cube covering Blockchain Week. Now, here's John Furrier. >> Hello everybody, welcome back. I'm John Furrier, host of The Cube here in New York City for Blockchain Week, New York, also part of the consensus 2018 event, wrapping up day three. We've been rocking and rolling. All the action, cryptos here from business models, financing, technology change and a lot of demos. It's been great. My next guest is Alex Scarsini who's the president of Edgewater Markets. Great to see you, thanks for coming on. >> Thank you, thank you very much. We're excited. >> So we were chatting about a lot of the capital markets and then the go-to markets for these companies, and I got to say the feedback here at the show is the demos are kind of suck-y 'cause everyone's working on the backend technologies. So it's the evolution, but you're starting to see the technology having a real impact. >> Of course. >> What are you working on? Take a minute to explain what you guys are doing, and then we can chat about the general market. >> Sure, sure. So Edgewater Markets, we are developing what we think will be the preeminent platform for the institutional market purely institutional market that'll enable sophisticated investors to be able to buy and sell all the digital currencies or at least in the first stage of our rollout, the 10 or 15 most selected currencies in the same manner that they transact their currency business today, which is efficiently, at low cost, in a low latency environment, and with a A-Z turn around of trade processes from the initial buy or sell to the confirmation process. We've done this before at Edgewater Markets. We do it in the currency business. We have a one-stop shop platform where our clients come to us and efficiently access global equity for the currencies they want. In the market that we have today, it's virtually impossible for the institutional segment to get in and get involved in a scalable manner. There's just too much dislocation globally in terms of exchanges, in terms of collateral that needs to be posted or not, in terms of accessing rates in a low latency environment. None of this exists. >> So is it their problem in that there's too much time to do work? Is it mechanisms that aren't in place? What's the real frustration that you guys are solving? I mean, mention dislocation, be specific. Is it the time it takes? No systems in place? >> Well, imagine that, imagine that you are a large institutional trader and you wanted to buy 1000 units of bitcoin. It's a million dollar log or eight million five hundred dollar log. You'd have to go and check prices on 20 different exchanges where you can buy three cheaply, where you can buy the next four cheaply. By the time you've looked and figured all this out, the price has moved. >> Yeah, exactly. >> It's impossible. Moreover, our clients want to buy 1000 logs and they may very well want to sell them out in 30 seconds. They don't want delivery of the coin, they don't want to deal with cold storage, warm storage. They want to speculate on the movement of these digital currencies the same way they do in Eurodollar, et cetera, so they need to be able to buy their interest in one place efficiently and at the best price. >> It's a great model, so much value there. How's it work and how's it coming together? So, you got to go set up, what, all the market-making deals? So you have to set up the connections? What are you guys bringing to the table? How does the platform work behind the scenes? >> Well the good part about all this is we've done it before. We've been in business for ten years. We've set up offices globally in London, Singapore, New York, Chicago. We're in Mexico City, as well. We have servers in each one of these locations, so we're already a very low latency provider of liquidity, and we already have a like-product for the FX side. We have, obviously, a smart order routing system. We have a central limit order book. We have a pricing engine. We have algos. We've already developed a lot of the processes that we will need for this new product. The most important part of the equation, for us, is we have 300 active institutional clients that are waiting for this product. >> Yeah, they're dying. >> That's why we have a tremendous advantage. >> So what changes are you making for cryptos? So you've got the great leverage from your previous experience, check, awesome. What's the cryptos tweak to your model? What's the key? >> So the market is yet to solve for the custodian part of the equation. In today's world, institutions, and I'm talking about the household names in the macrospace or the high frequency space, I would say 99% of the institutional space deals in the name of their PB, prime broker. Goldman, Morgan Stanley, et cetera, as do we. Now these exchanges are what we call the liquidity providers. You would have to go up and literally set up an account with each one of these exchanges so you can access that liquidity. It's completely inefficient. So what we aim to do, in the absence of a solution in the next year, what we plan to do for our rollout is to open those accounts ourselves and have that collateral with each one of these exchanges, obviously we'll get leverage, there's going to be some cross pollination of products between exchanges at some point. The way you have it today between a London exchange in the equities world and a New York stock exchange where they cross pollinate some of their liquidity pools, you will see that in exchanges throughout the world, and you'll also see some consolidation in that space. So we're going to put up the collateral, we're going to deal with the exchanges, we're going to make sure that we do the post-trade processes on behalf of our client. Our client comes to us, buys 100 units and sells them 30 seconds later, and he's either made money or he's lost money, and that just gets netted out at the end of the day or at the end of the week depending on the agreements that we have with our institutional clients. >> I look at all this, some of the transactions it takes, such a long time to get stuff done because it's just kludgy, it's really a mess. It's exciting that you guys are doing this. >> Well, but there's a need for it. The market is growing tremendously fast. I mean, it's evolved just in the last year, we've evolved from some concepts that were in the preliminary stage to a real demand for a product for an institutional client base that is dying for new product in this environment, meaning the currency markets are very quiet. The bond markets, although they're picking up or percolating right now, are very quiet. This is an area that gives the institutional traders and speculators a chance to arbitrage, to produce alpha, and to do it efficiently. They need a product, and we're there for it. >> Alex, how do people get involved? Obviously they're lining up, waiting for the mousetrap to be built 'cause it's a better mousetrap, obviously, than what's out there. What's next for you guys? How do people get involved? They just call you up at Edgewater Markets? Is there a front end website? How do they contact you guys? >> Well, yeah, certainly we have a website. We're in the process of putting together a product we think will be ready in the next three months for a beta roll out. We've got all hands on deck building it out. We have a handful of clients that have agreed to beta test it for us, so we do think we're ahead of the curve. I've seen a lot of other companies that are trying to do what we do, and we always believe that in the absence of a real clientele that demands the product, it's tough to build what you don't know you'll be asked to build eventually. Our clients are looking at our products, giving us live feedback today as we speak, and these aren't small institutional clients. These are your household names in the macrospace. >> Yeah, they need it. >> The big boys. And so we think we'll have something great in the next few months. >> Great Alex, thanks for coming onto The Cube, really appreciate it. Good luck tonight and continue the events here, and great job. We need that. >> Thank you. Thanks for having me. >> Liquidity's critical marketplaces are being developed. You've got two-sided marketplaces, you got cryptocurrency. This is a new, exciting product at many levels. Financial obviously here with Alex, technology, business model, all covered on The Cube here. I'm John Furrier, thanks for watching. More coverage here in New York City after this short break. (upbeat electronic music)
SUMMARY :
Announcer: From New York, it's The Cube also part of the consensus 2018 event, Thank you, thank you very much. So it's the evolution, but you're starting to see Take a minute to explain what you guys are doing, in the same manner that they transact What's the real frustration that you guys are solving? By the time you've looked and figured all this out, in one place efficiently and at the best price. How does the platform work behind the scenes? We've already developed a lot of the processes a tremendous advantage. What's the cryptos tweak to your model? and that just gets netted out at the end of the day It's exciting that you guys are doing this. This is an area that gives the institutional traders How do they contact you guys? of a real clientele that demands the product, in the next few months. and great job. Thanks for having me. More coverage here in New York City after this short break.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Alex Scarsini | PERSON | 0.99+ |
10 | QUANTITY | 0.99+ |
London | LOCATION | 0.99+ |
Mexico City | LOCATION | 0.99+ |
New York City | LOCATION | 0.99+ |
Morgan Stanley | ORGANIZATION | 0.99+ |
1000 units | QUANTITY | 0.99+ |
1000 logs | QUANTITY | 0.99+ |
eight million | QUANTITY | 0.99+ |
Goldman | ORGANIZATION | 0.99+ |
99% | QUANTITY | 0.99+ |
New York | LOCATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
ten years | QUANTITY | 0.99+ |
Edgewater Markets | ORGANIZATION | 0.99+ |
100 units | QUANTITY | 0.99+ |
20 different exchanges | QUANTITY | 0.99+ |
Alex | PERSON | 0.99+ |
30 seconds | QUANTITY | 0.99+ |
next year | DATE | 0.99+ |
Blockchain Week | EVENT | 0.99+ |
three | QUANTITY | 0.99+ |
last year | DATE | 0.99+ |
today | DATE | 0.99+ |
tonight | DATE | 0.99+ |
Singapore | LOCATION | 0.98+ |
NYC | LOCATION | 0.98+ |
two-sided | QUANTITY | 0.98+ |
day three | QUANTITY | 0.97+ |
first stage | QUANTITY | 0.97+ |
The Cube | ORGANIZATION | 0.96+ |
300 active institutional clients | QUANTITY | 0.95+ |
Chicago | LOCATION | 0.95+ |
one | QUANTITY | 0.94+ |
each one | QUANTITY | 0.94+ |
30 seconds later | DATE | 0.94+ |
2018 | DATE | 0.91+ |
15 most selected currencies | QUANTITY | 0.91+ |
next three months | DATE | 0.88+ |
one-stop | QUANTITY | 0.84+ |
four | QUANTITY | 0.83+ |
Blockchainweek | EVENT | 0.82+ |
five hundred dollar log | QUANTITY | 0.81+ |
a million dollar | QUANTITY | 0.74+ |
next few months | DATE | 0.73+ |
Eurodollar | OTHER | 0.72+ |
bitcoin | OTHER | 0.57+ |
Christine Yen, Honeycomb io | DevNet Create 2018
>> Announcer: Live from the Computer History Museum in Mountain View, California. It's theCUBE, covering DevNet Create 2018. Brought to you by Cisco. >> Hey, welcome back, everyone. This is theCUBE, live here in Mountain View, California, heart of Silicon Valley for Cisco's DevNet Create. This is their Cloud developer event. It's not the main Cisco DevNet which is more of the Cisco developer, this is much more Cloud Native DevOps. I'm joined with my cohost, Lauren Cooney and our next guest is Christine Yen, who is co-founder and Chief Product Officer of Honeycomb.io. Welcome to theCUBE. >> Thank you. >> Great to have an entrepreneur and also Chief Product Officer because you blend in the entrepreneurial zeal, but also you got to build the product in the Cloud Native world. You guys done a few ventures before. First, take a minute and talk about what you guys do, what the company is built on, what's the mission? What's your vision? >> Absolutely, Honeycomb was built, we are an observability platform to help people find the unknown unknowns. Our whole thesis is that the world is getting more complicated. We have microservices and containers, and instead of having five application servers that we treated like pets in the past, we now have 500 containers running that are more like cattle and where any one of them might die at any given time. And we need our tools to be able to support us to figure out how and why. And when something happens, what happened and why, and how do we resolve it? We look around at the landscape and we feel like this dichotomy out there of, we have logging tools and we have metrics tools. And those really evolved from the fact that in 1995, we had to choose between grep or counters. And as technology evolved, those evolved to distribute grep or RDS. And then we have distribute grep with fancy UIs and well, fancy RDS with UIs. And Honeycomb, we were started a couple years ago. We really feel like what if you didn't have to choose? What if technology supported the power of having all the context there the way that you do with logs while still being able to provide instant analytics the way that you have with metrics? >> So the problem that you're solving is one, antiquated methodologies from old architectures and stacks if you will, to helping people save time, with the arcane tools. Is that the main premise? >> We want people to be able to debug their production systems. >> All right, so, beyond that now, the developer that you're targeting, can you take us through a day in the life of where you are helping them, vis a vis the old way? >> Absolutely, so I'll tell a story of when myself and my co-founder, Charity, were working together at PaaS. PaaS, for those who aren't familiar, used to be RD, a backend form of mobile apps. You can think of someone who just wants to build an iOS app, doesn't want to deal with data storage, user records, things like that. And PaaS started in 2011, got bought by Facebook in 2013, spun down very beginning of 2016. And in 2013, when the acquisition happened, we were supporting somewhere on the order of 60,000 different mobile apps. Each one of them could be totally different workload, totally different usage pattern, but any one of them might be experiencing problems. And again, in this old world, this pre-Honeycomb world, we had our top level metrics. We had latency, response, overall throughput, error rates, and we were very proud of them. We were very proud of these big dashboards on the wall that were green. And they were great, except when you had a customer write in being like, "Hey, PaaS is down." And we look at our dashboard we'd be like, "Nope, it's not down. "It must be network issues." >> John: That's on your end. >> Yeah, that's on your end. >> John: Not a good answer. >> Not a good answer, and especially not if that customer was Disney, right? When you're dealing with these high level metrics, and you're processing tens or hundreds of thousands of requests per second, when Disney comes in, they've got eight requests a second and they're seeing all of them fail. Even though those are really important, eight requests per second, you can't tease that out of your graphs. You can't figure out why they're failing, what's going on, how to fix it. You've got to dispatch an engineer to go add a bunch of if app ID equals Disney, track it down, figure out what's going on there. And it takes time. And when we got to Facebook, we were exposed to a type of tool that essentially inspired Honeycomb as it is today that let us capture all this data, capture a bunch of information about everything that was happening down to these eight requests per second. And when a customer complained, we could immediately isolate, oh, this one app, okay let's zoom in. For this one customer, this tiny customer, let's look at their throughput, error rates, latency. Oh, okay. Something looks funny there, let's break down by endpoint for this customer. And it's this iterative fast, highly granular investigation, that is where all of us are approaching today. With our systems getting more complicated you need to be able to isolate. Okay, I don't care about the 200s, I only care about the 500s, and within the 500s, then what's going on? What's going on with this server, with that set of containers? >> So this is basically an issue of data, unstructured data or have the ability to take this data in at the same time with your eye on the prize of instrumentation. And then having the ability to make that addressable and discoverable in real time, is that kind of? >> Yeah, we've been using the term observability to describe this feeling of, I need to be able to find unknown unknowns. And instrumentation is absolutely the tactic to observability of the strategy. It is how people will be able to get information out of their systems in a way that is relevant to their business. A common thing that we'll hear or people will ask, "Oh, can you ingest my nginx logs?" "Can you ingest my SQL logs?" Often, that's a great place to start, but really where are the problems in an application? Where are your problems in the system? Usually it's the places that are custom that the engineers wrote. And tools need to be able to support, providing information, providing graphs, providing analytics in a way that makes it easy for the folks who wrote the code to track down the problem and address them. >> It's a haystack of needles. >> Yeah, absolutely. >> They're all relevant but you don't know which needle you're going to need. >> Exactly. >> So, let me just get this. So I'm ducking out, just trying to understand 'cause this is super important because this is really the key to large scale Cloud ops, what we're talking about here. From a developer standpoint, and we just had a great guest on, talking about testing features and production which is really the important, people want to do that. And then, but for one person, but in production scale, huge problem, opportunity as well. So, if most people think of like, "Oh, I'll just ingest with Splunk," but that's a different, is that different? I mean, 'cause people think of Splunk and they think of Redshift and Kinesis on Amazon, they go, "Okay." Is that the solution? Are you guys different? Are you a tool? How do I understand you guys' context to those known solutions? >> First of all, explain the difference between ourselves and the Redshifts and big queries of the world, and then I'll talk about Splunk. We really view those tools as primarily things built for data scientists. They're in the big data realm, but they are very concerned with being 100% correct. They're concerned with fitting into big data tools and they often have an unfortunate delay in getting data in and making it acquirable. Honeycomb is 100% built for engineers. Engineers of people, the folks who are going to be on the hook for, "Hey, there's downtime, what's going on?" And in-- >> So once business benefits, more data warehouse like. >> Yeah. And what that means is that for Honeycomb, everything is real time. It's real time. We believe in recent data. If you're looking to get query data from a year ago we're not really the thing, but instead of waiting 20 minutes for a query over a huge volume of data, you wait 10 seconds, or it's 3:00 AM and you need to figure out what's happening right now, you can go from query to query, to query, to query, as you come up with hypotheses, validate them or invalidate them, and continue on your investigation path. So that's... >> That makes sense. >> Yeah. >> So data wrangling, doing queries, business intelligence, insights as a service, that's all that? >> Yeah. We almost, we played with and tossed the tagline BI for systems because we want that BI mentality of what's going on, let me investigate. But for the folks who need answers now, an approximate answer now is miles better than a perfect one-- >> And you can't keep large customers waiting, right? At the end of the day, you can't keep the large customers waiting. >> Well, it's also so complicated. The edge is very robust and diverse now. I mean, no-js is a lot of IO going on for instance. So let's just take an example. I had developer talking the other day with me about no-js. It's like, oh, someone's complaining but they're using Firefox. It's like, okay, different memory configuration. So the developer had to debug because the complaints were coming in. Everyone else was fine, but the one guy is complaining because he's on Firefox. Well, how many tabs does he have open? What's the memory look like? So like, this a weird thing, I mean, that's just a weird example, but that's just the kinds of diverse things that developers have to get on. And then where do they start? I mean. >> Absolutely. So, there's something we ran into or we saw our developers run into all the time at PaaS, right? These are mobile developers. They have to worry about not only which version of the app it is, they have to worry about which version of the app, using which version of RSDK on which version of the operating system, where any kind of strange combination of these could result in some terrible user experience. And these are things that don't really work well if you're relying on pre-aggregated 10 series system, like the evolution of the RDS, I mentioned. And for folks who are trying to address this, something like Splunk, these logging tools, frankly, a lot of these tools are built on storage engines that are intended for full text search. They're unstructured text, you're grepping over them, and then you're build indices and structure on top of that. >> There's some lag involved too in that. >> There's so much lag involved. And there's almost this negative feedback loop built in where if you want to add more data, if on each log line you want to start tracking browser user agent, you're going to incur not only extra storage costs, you're going to incur extra read time costs because you're reading that more data, even if you're don't even care about that on those queries. And you're probably incurring cost on the right time to maintain these indices. Honeycomb, we're a column store through and through. We do not care about your unstructured text logs, we really don't want them. We want you to structure your data-- >> John: Did you guys write your own column store or is that? >> We did write our own column store because ultimately there's nothing off the shelf that gave us the speed that we wanted. We wanted to be able to, Hey, sending us data blogs with 20, 50, 200 keys. But if you're running analysis and all you care about is a simple filter and account, you shouldn't have to pull in all this-- >> To become sort of like Ferrari, if you customize, it's really purpose built, is that what you guys did? >> That is. >> So talk about the dynamic, because now you're dealing with things like, I mean, I had a conversation with someone who's looking at say blockchain, where there's some costs involved, obviously writing to the blockchain. And this is not like a crypto thing it's more of a supply chain thing. They want visibility into latency and things of that nature. Does this sounds like you would fit there as a potential use case? Is that something that you guys thought of at all? >> It could absolutely be. I'm actually not super familiar with the blockchain or blockchain based applications but ultimately Honeycomb is intended for you to be able to answer questions about your system in a way that tends to stymie existing tools. So we see lots of people come to us from strange use cases who just want to be able to instrument, "Hey I have this custom logic. "I want to be able to look at what it's doing." And when a customer complains and my graphs are fine or when my graphs are complaining, being able to go in and figure out why. >> Take a minute to talk about the company you founded. How many employees funding, if you can talk about it. And use case customers you have now. And how do you guys engage? The service, is it, do I download code? Is it SaaS? I mean, you got all this great tech. What's the value proposition? >> I think I'll answer this-- >> John: Company first. >> All right. >> John: Status of the company. >> Sure. Honeycomb is about 25 people, 30 people. We raised a series A in January. We are about two and a half years old and we are very much SaaS of the future. We're very opinionated about a number of things and how we want customers to interact with us. So, we are SaaS only. We do offer a secure proxy option for folks who have PII concerns. We only take structured data. So, at our API, you can use whatever you want to slurp data from your system. But at our API, we want JSON. We do offer a wide variety of integrations, connectors, SDKs, to help you structure that data. But ultimately-- >> Do you provide SDKs to your customers? >> We do. So that if they want to instrument their application, we just have the niceties around like batching and doing things asynchronously so it doesn't block their application. But ultimately, so we try to meet folks where they're at, but it's 2016, it was 2017, 2018-- >> You have a hardened API, API pretty much defines your service from an inbound standpoint. Prices, cost, how does someone engage with you guys? When does someone know to engage? Where's the smoke signals? When is the house on fire? Is it like people are standing around? What's the problem? When does someone know to call you guys up at? >> People know to call us when they're having production problems that they can't solve. When it takes them way too long to go from there's an alert that went off or a customer complaint, to, "Oh, I found the problem, I can address it." We price based on storage. So we are a bunch of engineers, we try to keep the business side as simple as possible for better, for worse. And so, the more data you send us, the more it'll cost. If you want a lot of data, but stored for a short period of time, that will cost less than a lot of data stored for a long period of time. One of the things that we, another one of the approaches that is possibly more common in the big data world and less in the monitoring world is we talk a lot about sampling. Sampling as a way to control those costs. Say you are, Facebook, again, I'll return to that example. Facebook knew that in this world where lots and lots of things can go wrong at any point in time, you need to be able to store the actual context of a given event happening. Some unit of work, you want to keep track of all the pieces of metadata that make that piece of work unique. But at Facebook scale, you can't store every single one of them. So, all right, you start to develop these heuristics. What things are more interesting than others? Errors are probably more interesting than 200 okays. Okay. So we'll keep track of most errors, we'll store 1% of successful requests. Okay. Well, within that, what about errors? Okay. Well, things that time out are maybe more interesting than things that are permissioning errors. And you start to develop this sampling scheme that essentially maps to the interesting ness of the traffic that's flowing through your system. To throw out some numbers, I think-- >> Machine learning is perfect for that too. They can then use the sampling. >> Yeah. There's definitely some learning that can happen to determine what things should be dropped on the ground, what requests are perfectly representative of a large swath of things. And Instagram, used a tool like this inside Facebook. They stored something like 1/10 of a percent or a 1/100 of a percent of their requests. 'Cause simply, that was enough to give them a sketch of what representative traffic, what's going wrong, or what's weird that, and is worth digging into. >> Final question. What's your priorities for the product roadmap? What are you guys focused on now? Get some fresh funding, that's great. So expand the team, hiring probably. Like product, what's the focus on the product? >> Focus on the product is making this mindset of observability accessible to software engineers. Right, we're entering this world where more and more, it's the software engineers deploying their code, pushing things out in containers. And they're going to need to also develop this sense of, "Okay, well, how do I make sure "something's working in production? "How do I make sure something keeps working? "And how do I think about correctness "in this world where it's not just my component, "it's my component talking to these other folks' pieces?" We believe really strongly that the era of this single person in a room keeping everything up, is outdated. It's teams now, it's on call rotations. It's handing off the baton and sharing knowledge. One of the things that we're really trying to build into the product, that we're hoping that this is the year that we can really deliver on this, is this feeling of, I might not be the best debugger on the team or I might not be the best person, best constructor of graphs on the team, and John, you might be. But how can a tool help me as a new person on a team, learn from what you've done? How can a tool help me be like, Oh man, last week when John was on call, he ran into something around my SQL also. History doesn't repeat, but it rhymes. So how can I learn from the sequence of those things-- >> John: Something an expert system. >> Yeah. Like how can we help build experts? How can we raise entire teams to the level of the best debugger? >> And that's the beautiful thing with metadata, metadata is a wonderful thing. 'Cause Jeff Jonas said on the, he was a Cube alumni, entrepreneur, famous data entrepreneur, observation space is super critical for understanding how to make AI work. And that's to your point, having observation data, super important. And of course our observation space is all things. Here at DevNet Create, Christine, thanks for coming on theCUBE, spending the time. >> Thank you. >> Fascinating story, great new venture. Congratulations. >> Christine: Thank you. >> And tackling the world of making developers more productive in real time in production. Really making an impact to coders and sharing and learning. Here in theCUBE, we're doing our share, live coverage here in Mountain View, DevNet Create. We'll be back with more after this short break. (gentle music)
SUMMARY :
Brought to you by Cisco. It's not the main Cisco DevNet in the Cloud Native world. the way that you have with metrics? Is that the main premise? to debug their production systems. on the wall that were green. I only care about the 500s, And then having the ability to make that that the engineers wrote. but you don't know which Is that the solution? and big queries of the world, So once business benefits, or it's 3:00 AM and you need to figure out But for the folks who need answers now, And you can't keep large So the developer had to debug all the time at PaaS, right? on the right time to and all you care about is a Is that something that you is intended for you about the company you founded. and how we want customers So that if they want to call you guys up at? And so, the more data you perfect for that too. that can happen to determine what things focus on the product? that the era of this to the level of the best debugger? And that's the beautiful And tackling the world
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Lauren Cooney | PERSON | 0.99+ |
John | PERSON | 0.99+ |
Jeff Jonas | PERSON | 0.99+ |
Christine | PERSON | 0.99+ |
January | DATE | 0.99+ |
2013 | DATE | 0.99+ |
tens | QUANTITY | 0.99+ |
20 | QUANTITY | 0.99+ |
1995 | DATE | 0.99+ |
Christine Yen | PERSON | 0.99+ |
20 minutes | QUANTITY | 0.99+ |
2011 | DATE | 0.99+ |
Disney | ORGANIZATION | 0.99+ |
10 seconds | QUANTITY | 0.99+ |
Firefox | TITLE | 0.99+ |
1% | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
100% | QUANTITY | 0.99+ |
500 containers | QUANTITY | 0.99+ |
Cisco | ORGANIZATION | 0.99+ |
3:00 AM | DATE | 0.99+ |
30 people | QUANTITY | 0.99+ |
Ferrari | ORGANIZATION | 0.99+ |
iOS | TITLE | 0.99+ |
50 | QUANTITY | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
1/100 | QUANTITY | 0.99+ |
Mountain View, California | LOCATION | 0.99+ |
Honeycomb.io | ORGANIZATION | 0.99+ |
last week | DATE | 0.99+ |
2017 | DATE | 0.99+ |
Honeycomb | ORGANIZATION | 0.99+ |
Mountain View | LOCATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
60,000 different mobile apps | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
First | QUANTITY | 0.99+ |
200 keys | QUANTITY | 0.98+ |
2016 | DATE | 0.98+ |
2018 | DATE | 0.98+ |
Cube | ORGANIZATION | 0.98+ |
DevNet Create | ORGANIZATION | 0.97+ |
SQL | TITLE | 0.97+ |
five application servers | QUANTITY | 0.97+ |
one customer | QUANTITY | 0.97+ |
a year ago | DATE | 0.96+ |
ORGANIZATION | 0.96+ | |
one person | QUANTITY | 0.95+ |
one | QUANTITY | 0.95+ |
about 25 people | QUANTITY | 0.94+ |
JSON | TITLE | 0.94+ |
about two and a half years old | QUANTITY | 0.94+ |
series A | OTHER | 0.93+ |
Each one | QUANTITY | 0.93+ |
one guy | QUANTITY | 0.91+ |
eight requests per second | QUANTITY | 0.9+ |
eight requests a second | QUANTITY | 0.89+ |
less than a lot of data | QUANTITY | 0.89+ |
1/10 of a percent | QUANTITY | 0.89+ |
each log line | QUANTITY | 0.88+ |
one app | QUANTITY | 0.87+ |
Splunk | ORGANIZATION | 0.86+ |
couple years ago | DATE | 0.85+ |
a percent | QUANTITY | 0.85+ |
Zachary Musgrave & Chris Gordon, Yelp | Splunk .conf 2017
>> Narrator: Live from Washington D.C., it's theCUBE. Covering .conf2017. Brought to you by Splunk. >> Well welcome back here on theCUBE. We continue our coverage of .conf2017, we're in Washington D.C. Along with Dave Vellante, I'm John Walls. And Dave, you know what time it is, by the way? Just about? >> I don't know, this is the penultimate interview. >> It's almost five o'clock. >> Okay. >> And that means it's almost happy hour time. So I was thinking where might we go tonight, so-- >> There's an app for that. >> There was, and so I looked. It turns out that the Penny Whiskey Cafe is just two tenths of a mile from here. And you know how I knew that? >> How's the ratings on that? >> We got four. >> Four and half with 52. >> 52 reviews? >> Yeah, I feel good about that. >> Yeah, that's pretty good. That's a substantive base. >> I feel very solid with that one. We'll make it 53 in about a half hour. Of course I found it on Yelp. We have a couple of gentlemen from Yelp with us tonight. I don't have to tell you what Yelp does, it does everything for everybody, right. Zach Musgrave, technical lead, and Chris Gordon, software engineer at Yelp. Gentlemen, thanks for being here. And U can join us, by the way, later on, at the Penny Whiskey if you'd like to. First off, what are you doing here, right, at Splunk? What's Yelp and Splunk, what's that intersection all about? Zach, if you would. >> Sure, well Yelp uses Splunk for all sorts of purposes. Operational, intelligence, business metrics, pretty much any sort of analytics from event driven data that you can really think of, Yelp has found a way, and our engineers have found a way to get that into Splunk and derive business value from it. So Chris and I are actually here, we just gave a breakout session at .conf, talking about how we find strong business value and how we quantify that value and mutate our Splunk cluster to really drive that. >> Okay. >> So, so how do you find value then, I mean, what was? >> It's hard. Chris was one of the people who really, really drove this for us. And when we looked at this, you know I once had an engineer who came up to our team, we maintain Splunk amongst other things, and the engineer said can I ingest 10 terabytes of data a day into Splunk and then keep it forever? And I said, um, please don't. And then we talked a bit more about what that engineer was actually trying to do and why they needed this massive amount of data, and we found a better way that was much more efficient. And then where we didn't need to keep all the data forever. So, by being able to have those conversations and to quantify with the data you're already ingesting into Splunk, being able to quanitfy that and actually show how many people were searching this, how's it being used, what's the depth of the search look like, how far back are they looking in time. You can really optimize your Splunk cluster to get a lot more business value than just naively setting it up and turning it on. >> So you weren't taking a brute force approach, you were smarter about that, but you weren't deduping, you were identifying the data that was not necessary to keep, did I get that right? >> Correct. Yeah, we essentially kind of identified what are highest cost per search logs, which we basically just totaled up how many times each log was searched, and then tried to quantify how much each logs was costing us. And then this ended up being a really good metric for figuring out what we'd want to remove or something that was a candidate for dislodging the data somehow. >> So, you guys gave a talk today. We were talking off camera about pricing, that's not something you guys get involved in, but I would categorize this as sort of how do you get the most out of that asset, called Splunk, right. Is that sort of the >> Exactly. >> theme of your talk, right? >> Yeah. We talk a lot about expected value amongst our team, and in the talk we just gave. And we don't ever think about this as, oh do this so that you can spend less money on Splunk or on your infrastructure that's backing Splunk. Think about is more as we have this right now and we can utilize it more effectively. We can get more value out of what we already have. >> Okay, so, I wonder if we could just talk a little bit about your environment. We know you run on AWS. How does that cloud fit in with Splunk, paint a picture for us, if you would. What does it all look like? >> Yeah, so we have two clusters actually. One is the high value, high quality of service cluster, it's the larger generic, we call it generic prod, and then we have another one, where we kind of have our more verbose, maybe slightly less valuable per log cluster. And this runs on a D2, which is just instant storage. And then the higher performance cluster runs all on a GP2. So it's basically just SSDs. And we also do, we also have four copies of each log and we have two searchable copies of each log, so it's pretty well replicated. >> Dave: Okay, so that's how you protect the data. >> Yeah. >> Is to make copies, in what, in different zones, or? >> Yeah, we have two copies of each log in each availability zone, and then one searchable copy of each log in each availability zone. >> And you guys are cloud natives, all cloud, just out of school and graduate school. So you talked about infrastructure as code. You don't do any of that on-prem stuff, you're not like installing gear. And so it's not part of your lexicon, right? >> No. >> Okay. So I want to do a little editorial thing. Kristen Nicole, our managing editor, sent the note around today saying 101s get the best traffic on the website. So I want to do a little DevOps 101, okay. Even though, it's second nature to you, and a lot of people in our audience know what it is. How do you describe DevOps? Give us the 101 on DevOps. >> Okay so, DevOps is a complicated thing, but and occasionally you see it as like a role on like a job board or something. And that always strikes me as odd, because it's not really a role. Like it's a philosophy moreso. The way that I always see it, is it used to be like pre DevOps, was the software developers make a thing, and then they throw it over the fence, and operations just picks it up. And they're like well what do we do with this, and deploy it, okay, good luck. And so with this result in a sort of an us against them mentality, where the developers aren't incentivized to really make it resilient, or really document it well, and operations and the sys admins are not incentivized to really be flexible and to be really hard charging and move quickly, because they're the ones who are going to be on call for whatever the developers made. DevOps is a we, instead of an us verses them. So for example, product teams have an on-call rotation. Operations and sys admins write code. There are still definitely specializations, but it all comes together in a much more holistic manner. >> Okay, and the ops guys will write code, as opposed to hacking code, messing up your code, throwing it back over the fence, and saying hey your code doesn't work. >> Exactly. >> And then you say well it worked when I gave it to you. And then like you said that sort of finger pointing. >> We are totally done with works on my machine, it's over. No more. >> Okay, and the benefits obviously are higher quality, faster time to market, less food fighting. >> Yup, exactly. In the old model you'd have a new deployment of like a website like maybe once a week or maybe even once a month. Yelp deploys multiple times everyday over and over again. And each one of those is going to include changes from a dozen different engineers. So we need to be agile in that manner, just like with our Splunk cluster. >> I mean you guys are relatively new, four years and two years, perspectively. But these days it's a long time. How would you describe your Splunk journey. Where did it start and where do you want to take it? >> I would say it started, you actually had Kris Wehner on here last year, and he talked a lot about it. He was the VP of engineering at SeatMe. And he kind of got Yelp onto the whole Splunk train. And at that point it was used mostly by SeatMe and everyone at Yelp was like oh this is fantastic, we want to use this. And we started basically migrating it to our VPC. And have generally, we're starting to now get everything going, get all the kinks worked out, and really now we're trying to see where we can provide the most value and make things as easy as possible for our developers to add logs and add searches and get what they need out of it. >> So what kind of use cases are you envisioning, and where are you getting value out of it? >> So we have our operations teams get a lot of value out of it when there's some outage happening. And it's really useful for them to be able to just look at the access logs and see what's going on. And Splunk makes that very easy. And we also get a lot of value out of Yelp's application logs. Splunk has been great for figuring out when something's not right. And allowing us to dig in further. >> So yeah, at the end of the day, as consumers, what does this mean to us, ultimately? Like our searches are faster, searches are more refined, searches are more accurate? What does it mean to me at the end of the day that you're enabling what activity through this technology. >> Dave: Yeah, it'll be more secure? >> Yeah, what does it mean? >> As an end user of Yelp? >> Yes. >> So, I'll give you one example that always sticks out in my mind. So I don't know if you all know this, but you can actually do things like order food via Yelp, you can make appointments via Yelp, even with like a dentist. You can beauty appointments, all sorts of personal services. >> Hair salon came up today actually, when I was looking for a bar. >> Absolutely. That's not supposed to happen. >> Dave: Well that was the Penny Whiskey Cafe. >> You never know, but what ever's next door I don't know. >> Can you get a haircut while you drink? >> Hair salons in the District are pretty impressive. >> I wasn't planning on it, no. But anyway, I'm sorry. >> Anyway, so we work with a lot of external partners to enable all these different integrations, right. So you press start order, and then eventually you see the menu, and then you add some stuff to your cart, and then you have to pay. And so if you haven't given us your credit card information yet, then you have to enter that, and that has to go to a payment processor, the order of course has to go out to the partner who's going to fulfill your order, and so on. So there's this pipeline of many different micro services plus the main Yelp application, plus this partner who's actually fulfilling your order, plus the payment processor, and so on, and so on. And it ends up with this really complicated state machine. So the way that actually works under the hood, to be very simplistic, is there's a unique order identifier that is assigned to you when you start the order. And then that passed through the whole process. So at every step in this process a bunch of events are emitted out of the various parts of the pipeline and into Splunk, where they're then matched to show that your order is progressing. And the order didn't get stuck. Because you know what's really sad is when you order food and it doesn't show up. So we really have to guard against that. >> Yeah, we hate that. >> Yeah, everybody does. So it's really important that we're able to unify this data, from all these different places, Splunk's really great for that, and to be able to then alert on that and page somebody and say hey, something's not quite right here, we have hungry folks. >> So while I have the smartest guys that we've interviewed all week here, you mentioned, >> Please. You mentioned, aw shucks, I know. You mentioned state machine. Are you playing around with functional programming, so called server lists, probably don't like that word either, but what are you doing there? Are you finding sort of new applications in use cases for so called server lists? >> I would say not so much. I don't know, is anyone at Yelp doing that? >> Yeah, there's some Lambda stuff going on. Like core back end is doing that work right now. A lot of our infrastructure is actually build up before the AWS Lambdas were a thing. So we found other ways to do that, and we have this really cool internal platform as a service, it's a docker, and some scheduling stuff on top of that. So a lot of things, like it's really easy to just launch a batch job in there. And it takes away some of the need for the true server lists. >> Well the reason I ask is because people are saying a lot of the state list IoT apps are going to use that sort of Lambda or homegrown stuff. And I'm not sure what the play is for Yelp in Internet of Things. I would imagine there's actually a play there for you guys though, and I'm curious as to the data angle, and maybe where Splunk might fit in. >> I'm certain that we're going to be using Splunk to read data from all of those different components as they're being launched. I know that there's been a couple early forays into the Lambda space that I've seen go by in code reviews and everything. But of course, with Splunk itself we can get data out of those. So as that happens, like we already have all our pipe lining set up. And it'll be pretty easy for them to analyze their self with Splunk. >> What gets you young folks excited these days? What keeps you enthralled and passionate? What do you look for? >> I don't know I think just in general anything that empowers you to get a lot done without having to fight it constantly. And general DevOps tools have been getting really good at that recently. And yeah, I would say anything that empowers you, gives you the feeling that you can do anything really. >> Yeah, all of the infrastructure is code stuff that's going on right now. So one of the pipelines that we use to get data out of Amazon S3, but it passes notifications through this S3 event notifications to Amazon SNS, to Amazon SQS, to our Splunk forwarders. And so that's a very complicated pipeline. And you have to set it all up, it works really well, but here's the cool part. That's all defined in code. And so this means that if you set up a new integration there's a code review. And we have some verification and validation that it's correct. And furthermore, if anything goes wrong with it, we can just hit a button and it recreates itself. That's what gets me happy. When tools get in my way that's not so good. >> Well and it just leaves more time for higher value activities and that's exciting. the transformation in infrastructure over the last five years has just been mind boggling. So, thanks you guys. >> It does. It does give me a lot of pleasure when something can go catastrophically wrong, and then just like, oh wait, it's self healing, all it can take is give three plays fine. And we're all dandy. >> Well to Dave's point, while I was off camera I did a search on the two smartest guys in the room. And it said one is six feet away the other one is seven feet away, so Yelp works, I mean it really does. But thanks for the time. It's been interesting. Next generation, right? So far over us. >> Yeah, I know. It's kind of depressing, but I love it. (laughing) >> Very good, thanks guys. >> Thank you so much. >> Back with more, here on theCUBE at .conf2017. We are live, Washington D.C. >> Dave: I've kind of had it with millennial. (upbeat music)
SUMMARY :
Brought to you by Splunk. And Dave, you know what time it is, by the way? And that means it's almost happy hour time. And you know how I knew that? Yeah, that's pretty good. I don't have to tell you what Yelp does, from event driven data that you can really think of, and to quantify with the data And then this ended up being a really good metric as sort of how do you get the most out of that asset, and in the talk we just gave. We know you run on AWS. and then we have another one, Yeah, we have two copies of each log And you guys are cloud natives, all cloud, and a lot of people in our audience know what it is. and operations and the sys admins Okay, and the ops guys will write code, And then you say We are totally done with works on my machine, it's over. Okay, and the benefits obviously are And each one of those is going to include changes How would you describe your Splunk journey. And he kind of got Yelp onto the whole Splunk train. And we also get a lot of value What does it mean to me at the end of the day So I don't know if you all know this, Hair salon came up today actually, That's not supposed to happen. but what ever's next door I don't know. Hair salons in the District I wasn't planning on it, and then you add some stuff to your cart, and to be able to then alert on that but what are you doing there? I don't know, is anyone at Yelp doing that? And it takes away some of the need and I'm curious as to the data angle, And it'll be pretty easy for them to analyze anything that empowers you to get a lot done And so this means that if you set up Well and it just leaves more time and then just like, oh wait, And it said one is six feet away the other one It's kind of depressing, but I love it. Back with more, here on theCUBE at .conf2017. Dave: I've kind of had it with millennial.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Chris | PERSON | 0.99+ |
Zach Musgrave | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Chris Gordon | PERSON | 0.99+ |
Yelp | ORGANIZATION | 0.99+ |
Kristen Nicole | PERSON | 0.99+ |
John Walls | PERSON | 0.99+ |
SeatMe | ORGANIZATION | 0.99+ |
six feet | QUANTITY | 0.99+ |
four | QUANTITY | 0.99+ |
seven feet | QUANTITY | 0.99+ |
Kris Wehner | PERSON | 0.99+ |
Four | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
Washington D.C. | LOCATION | 0.99+ |
Zach | PERSON | 0.99+ |
two copies | QUANTITY | 0.99+ |
last year | DATE | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
two smartest guys | QUANTITY | 0.99+ |
once a week | QUANTITY | 0.99+ |
four years | QUANTITY | 0.99+ |
each log | QUANTITY | 0.99+ |
53 | QUANTITY | 0.99+ |
once a month | QUANTITY | 0.99+ |
Splunk | ORGANIZATION | 0.99+ |
one | QUANTITY | 0.99+ |
two clusters | QUANTITY | 0.99+ |
Zachary Musgrave | PERSON | 0.99+ |
Lambda | TITLE | 0.99+ |
each logs | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
52 reviews | QUANTITY | 0.99+ |
52 | QUANTITY | 0.99+ |
tonight | DATE | 0.99+ |
second nature | QUANTITY | 0.99+ |
four copies | QUANTITY | 0.99+ |
Amazon | ORGANIZATION | 0.98+ |
DevOps | TITLE | 0.98+ |
Penny Whiskey Cafe | ORGANIZATION | 0.98+ |
Splunk | PERSON | 0.98+ |
First | QUANTITY | 0.97+ |
Lambdas | TITLE | 0.97+ |
DevOps 101 | TITLE | 0.97+ |
about a half hour | QUANTITY | 0.97+ |
each one | QUANTITY | 0.96+ |
one example | QUANTITY | 0.96+ |
each availability zone | QUANTITY | 0.95+ |
two years | QUANTITY | 0.94+ |
Day 3 Open | Red Hat Summit 2017
>> (upbeat music) Live from Boston Massachusetts. It's theCube! Covering Red Hat Summit 2017. Brought to you by Red Hat. >> It is day three of the Red Hat Summit, here in Boston Massachusetts. I'm Rebecca Knight. Along with Stu Miniman. We are wrapping up this conference Stu. We just had the final keynote of the morning. Before the cameras were rolling, you were teasing me a little bit that you have more scoop on the AWS deal. I'm interested to hear what you learned. >> (Stu) Yeah, Rebecca. First of all, may the fourth be with you. >> (Rebecca) Well, thank you. Of course, yes. And also with you. >> (Stu) Always. >> Yeah. (giggles) >> (Stu) So, day three of the keynote. They started out with a little bit of fun. They gave out some "May The Fourth Be With You" t-shirts. They had a little Star Wars duel that I was Periscoping this morning. So, love their geeking out. I've got my Millennium Falcon cuff links on. >> (Rebecca) You're into it. >> I saw a bunch of guys wearing t-shirts >> (Rebecca) Princess Leia was walking around! >> Princess Leia was walking around. There were storm troopers there. >> (Rebecca) Which is a little sad to see, but yes. >> (Stu) Uh, yeah. Carrie Fisher. >> Yes. >> Absolutely, but the Amazon stuff. Sure, I think this is the biggest news coming out of the show. I've said this a number of times. And we're still kind of teasing out exactly what it is. Cause, partially really this is still being built out. There's not going to be shipping until later this year. So things like how pricing works. We're still going to get there. But there's some people that were like "Oh wait!' "Open shift can be in AWS, that's great!" "But then I can do AWS services on premises." Well, what that doesn't mean, of course is that I don't have everything that Amazon does packaged up into a nice little container. We understand how computer coding works. And even with open-source and how we can make things server-less. And it's not like I can take everything that everybody says and shove it in my data center. It's just not feasible. What that means though, is it is the same applications that I can run. It's running in OpenShift. And really, there's the hooks and the API's to make sure that I can leverage services that are used in AWS. Of course, from my standpoint I'm like "OK!" So, tell me a little bit about how what latency there's going to be between those services. But it will be well understood as we build these what it's going to be use for. Certain use cases. We already talked to Optim. I was really excited about how they could do this for their environment. So, it's something we expect to be talking about throughout the rest of the year. And by the time we get to AWS Reinvent the week after Thanksgiving, I expect we'll have a lot more detail. So, looking forward to that. >> (Rebecca) And it will be rolled out too. So we'll have a really good sense of how it's working in the marketplace. >> (Stu) Absolutely. >> So other thoughts on the key note. I mean, one of the things that really struck me was talking about open-source. The history of open-source. It started because of a need to license existing technologies in a cheaper way. But then, really, the point that was made is that open-source taught tech how to collaborate. And then tech taught the world how to collaborate. Because it really was the model for what we're seeing with crowdsourcing solutions to problems facing education, climate change, the developing world. So I think that that is really something that Red Hat has done really well. In terms of highlighting how open-source is attacking many of the worlds most pressing problems. >> (Stu) Yeah, Rebecca I agree. We talked with Jim Whitehurst and watched him in the keynotes in previous days. And talked about communities and innovation and how that works. And in a lot of tech conferences it's like "Okay, what are the business outcomes?" And here it's, "Well, how are we helping the greater good?" "How are we helping education?" It was great to see kids that are coding and doing some cool things. And they're like, "Oh yeah, I've done Java and all these other things." And the Red Hat guys were like, "Hey >> (Rebecca) We're hiring. Yeah. (giggles) >> can we go hire this seventh grader?" Had the open-source hardware initiative that they were talking about. And how they can do that. Everything from healthcare to get a device that used to be $10,000 to be able to put together the genome. Is I can buy it on Amazon for What was it? Like six seven hundred dollars and put it together myself. So, open-source and hardware are something we've been keeping an eye on. We've been at the Open Compute Project event. Which Facebook launched. But, these other initiatives. They had.... It was funny, she said like, "There's the internet of things." And they have the thing called "The Thing" that you can tie into other pieces. There was another one that weaved this into fabric. And we can sensor and do that. We know healthcare, of course. Lot's of open-source initiatives. So, lots of places where open-source communities and projects are helping proliferate and make greater good and make the world a greater place. Flattening the world in many cases too. So, it was exciting to see. >> And the woman from the Open-Source Association. She made this great point. And she wasn't trying to be flip. But she said one of our questions is: Are you emotionally ready to be part of this community? And I thought that that was so interesting because it is such a different perspective. Particularly from the product side. Where, "This is my IP. This is our idea. This is our lifeblood. And this is how we're going to make money." But this idea of, No. You need to be willing to share. You need to be willing to be copied. And this is about how we build ideas and build the next great things. >> (Stu) Yeah, if you look at the history of the internet, there was always. Right, is this something I have to share information? Or do we build collaboration? You know, back to the old bulletin board days. Through the homebrew computing clubs. Some of the great progress that we've made in technology and then technology enabling beyond have been because we can work in a group. We can work... Build on what everyone else has done. And that's always how science is done. And open-source is just trying to take us to the next level. >> Right. Right. Right. And in terms of one of the last... One of the last things that they featured in the keynote was what's going on at the MIT media lab. Changing the face of agriculture. And how they are coding climate. And how they are coding plant nutrition. And really this is just going to have such a big change in how we consume food and where food is grown. The nutrients we derive from fruit. I was really blown away by the fact that the average apple we eat in the grocery store has been around for 14 months. Ew, ew! (laughs) So, I mean, I'm just exciting what they're doing. >> Yeah, absolutely right. If we can help make sure people get clean water. Make sure people have availability of food. Shorten those cycles. >> (Rebecca) Right, right. Exactly. >> The amount of information, data. The whole Farm to Table Initiative. A lot of times data is involved in that. >> (Rebecca) Yeah. It's not necessarily just the stuff that you know, grown on the roof next door. Or in the farm a block away. I looked at a local food chain that's everywhere is like Chipotle. You know? >> (Rebecca) Right. >> They use data to be able to work with local farmers. Get what they can. Try to help change some of the culture pieces to bring that in. And then they ended up the keynote talking more about innovation award winners. You and I have had the chance to interview a bunch of them. It's a program I really like. And talking to some of the Red Hatters there actually was some focus to work with... Talk to governments. Talk to a lot of internationals. Because when they started the program a few years ago. It started out very U.S.-centric. So, they said "Yeah." It was a little bit coincidence that this year it's all international. Except for RackSpace. But, we should be blind when we think about who has great ideas and good innovation. And at this conference, I bumped into a lot of people internationally. Talked to a few people coming back from the Red Sox game. And it was like, "How was it?" And they were like, "Well, I got a hotdog and I understood this. But that whole ball and thing flying around, I don't get it." And things like that. >> So, they're learning about code but also baseball. So this is >> (Stu) Yeah, what's your take on the global community that you've seen at the show this week? >> (Rebecca) Well, as you've said, there are representatives from 70 countries here. So this really does feel like the United Nations of open-source. I think what is fascinating is that we're here in the states. And so we think about these hotbeds of technological innovation. We're here in Boston. Of course there's Silicon Valley. Then there are North Carolina, where Red Hat's based. Atlanta, Austin, Seattle, of course. So all these places where we see so much innovation and technological progress taking place here in the states. And so, it can be easy to forget that there are also pockets all over Europe. All over South America. In Africa, doing cool things with technology. And I think that that is also ... When we get back to one of the sub themes of this conference... I mean, it's not a sub theme. It is the theme. About how we work today. How we share ideas. How we collaborate. And how we manage and inspire people to do their best work. I think that that is what I'd like to dig into a little today. If we can. And see how it is different in these various countries. >> Yeah, and this show, what I like is when its 13th year of the show, it started out going to a few locations. Now it's very stable. Next year, they'll be back in San Francisco. The year after, they'll be back here in Boston. They've go the new Boston office opening up within walking distance of where we are. Here GE is opening up their big building. I just heard there's lots of startups when I've been walking around the area. Every time I come down to the Sea Port District. It's like, "Wow, look at all the tech." It's like, Log Me In is right down the road. There's this hot little storage company called Wasabi. That's like two blocks away. Really excited but, one last thing back on the international piece. Next week's OpenStack Summit. I'll be here, doing theCube. And some of the feedback I've been getting this week It's like, "Look, the misperception on an OpenStack." One of the reasons why people are like, "Oh, the project's floundering. And it's not doing great, is because the two big use case. One, the telecommunication space. Which is a small segment of the global population. And two, it's gaining a lot of traction in Europe and in Asia. Whereas, in North America public cloud has kind of pushed it aside a little bit. So, unfortunately the global tech press tends to be very much, "Oh wait, if it's seventy-five percent adoption in North America, that's what we expect. If its seventy-five percent overseas, it's not happening. So (giggles) it's kind of interesting. >> (Rebecca) Right. And that myopia is really a problem because these are the trends that are shaping our future. >> (Stu) Yeah, yeah. >> So today, I'm also going to be talking to the Women In Tech winners. That very exciting. One of the women was talking about how she got her idea. Or really, her idea became more formulated, more crystallized, at the Grace Hopper Conference. We, of course, have a great partnership with the Grace Hopper Conference. So, I'm excited to talk to her more about that today too. >> (Stu) Yeah, good lineup. We have few more partners. Another customer EasiER AG who did the keynote yesterday. Looking forward to digging in. Kind of wrapping up all of this. And Rebecca it's been fun doing it with you this week. >> And I'm with you. And may the force... May the fourth be with you. >> And with you. >> (giggles) Thank you, we'll have more today later. From the Red Hat Summit. Here in Boston, I'm Rebecca Knight for Stu Miniman. (upbeat music)
SUMMARY :
Brought to you by Red Hat. We just had the final keynote of the morning. may the fourth be with you. And also with you. They had a little Star Wars duel that I was Periscoping Princess Leia was walking around. (Stu) Uh, yeah. And by the time we get to AWS Reinvent (Rebecca) And it will be rolled out too. is attacking many of the worlds most pressing problems. And the Red Hat guys were like, "Hey (Rebecca) We're hiring. And we can sensor and do that. And the woman from the Open-Source Association. Some of the great progress that we've made in technology And in terms of one of the last... If we can help (Rebecca) Right, right. The amount of information, data. It's not necessarily just the stuff that You and I have had the chance to interview a bunch of them. So this is And so, it can be easy to forget And some of the feedback I've been getting this week And that myopia is really a problem One of the women was talking about how she And Rebecca it's been fun doing it with you this week. And may the force... From the Red Hat Summit.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Rebecca | PERSON | 0.99+ |
Jim Whitehurst | PERSON | 0.99+ |
Rebecca Knight | PERSON | 0.99+ |
Boston | LOCATION | 0.99+ |
Chipotle | ORGANIZATION | 0.99+ |
Europe | LOCATION | 0.99+ |
Asia | LOCATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
North Carolina | LOCATION | 0.99+ |
$10,000 | QUANTITY | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
GE | ORGANIZATION | 0.99+ |
Atlanta | LOCATION | 0.99+ |
ORGANIZATION | 0.99+ | |
AWS | ORGANIZATION | 0.99+ |
Seattle | LOCATION | 0.99+ |
Austin | LOCATION | 0.99+ |
Africa | LOCATION | 0.99+ |
Wasabi | ORGANIZATION | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
Carrie Fisher | PERSON | 0.99+ |
Boston Massachusetts | LOCATION | 0.99+ |
San Francisco | LOCATION | 0.99+ |
Next year | DATE | 0.99+ |
North America | LOCATION | 0.99+ |
South America | LOCATION | 0.99+ |
Red Sox | ORGANIZATION | 0.99+ |
seventy-five percent | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
Next week | DATE | 0.99+ |
yesterday | DATE | 0.99+ |
70 countries | QUANTITY | 0.99+ |
13th year | QUANTITY | 0.99+ |
Java | TITLE | 0.99+ |
OpenShift | TITLE | 0.99+ |
this week | DATE | 0.99+ |
today | DATE | 0.99+ |
six seven hundred dollars | QUANTITY | 0.98+ |
Grace Hopper Conference | EVENT | 0.98+ |
two | QUANTITY | 0.98+ |
Red Hat Summit | EVENT | 0.98+ |
Stu | PERSON | 0.98+ |
two blocks | QUANTITY | 0.98+ |
OpenStack Summit | EVENT | 0.98+ |
one | QUANTITY | 0.98+ |
Sea Port District | LOCATION | 0.98+ |
United Nations | ORGANIZATION | 0.98+ |
this year | DATE | 0.97+ |
later this year | DATE | 0.97+ |
fourth | QUANTITY | 0.97+ |
Star Wars | TITLE | 0.97+ |
Red Hat Summit 2017 | EVENT | 0.97+ |
May The Fourth Be With You | TITLE | 0.96+ |
Princess Leia | PERSON | 0.96+ |