Chris Jones QA Session **DO NOT PUBLISH**
(upbeat music) >> Okay, welcome back everyone. I'm John Furrier here in theCUBE, in Palo Alto for "CUBE Conversation" with Chris Jones, Director of Product Management at Platform9. I've got a series of questions, had a great conversation earlier. Chris, I have a couple questions for you, what do you think? >> Let's do it, John. >> Okay, how does Platform9 Solution, you- can it be used on any infrastructure anywhere, cloud, edge, on-premise? >> It can, that's the beauty of our control plane, right? It was born in the cloud, and we primarily deliver that SaaS, which allows it to work in your data center, on bare metal, on VMs, or with public cloud infrastructure. We now give you the ability to take that control plane, install it in your data center, and then use it with anything, or even in air gap. And that includes capabilities with bare metal orchestration as well. >> Second question. How does Platform9 ensure maximum uptime, and proactive issue resolution? >> Oh, that's a good question. So if you come to Platform nine we're going to talk about always on assurance. What is driving that is a system of three components around self-healing, monitoring, and proactive assistance. So our software will heal broken things on nodes, right? If something stops running that should be running, it will attempt to restart that. We also have monitoring that's deployed with everything. So you build a cluster in AWS, well, we put open source monitoring agents, that are actually Prometheus, on every single node. That means it's resilient, right? So if you lose a node, you don't lose monitoring. But that data importantly comes back to our control plane, and that's the control plane that you can put in your data center as well. That data is what alerts us, and you as a user, anytime of the day that something's going wrong. Let's say etcd latency, good example, etcd is going slow. We'll find out, we might not be able to take restorative action immediately, but we're definitely going to reach out and say,, "You have a problem, let's get ahead of this and let's prevent that from becoming a bigger problem." And that's what we're delivering. When we say always on assurance, we're talking about self-healing, we're talking about remote monitoring, we're talking about being proactive with our customers, not waiting for the phone call or the support desk ticket saying, "Oh we think something's not working." Or worse, the customer has an outage. >> Awesome. Thanks for sharing. Can you explain the process for implementing Platform9 within a company's existing infrastructure. >> Are we doing air gap, or on-prem or SaaS approached? SaaS approach I think is by far the easiest, right? We can build a dedicated Platform9 control plane instance in a manner of minutes, for any customer. So when we do a proof of concept or onboarding, we just literally put in an email address, put in the name you want for your fully qualified domain name, and your instance is up. From that point onwards, the user can just log in, and using our CLI, talk to any number of, say, virtual machines, or physical servers in their environment for, you know, doing this in a data center or colo, and say, "I want these to be my Kubernetes control plane nodes. Here's the five of them. Here's the VIP for the load balancing, the API server and here are all of my compute nodes." And that CLI will work with the SaaS control plane, and go and build the cluster. That's as simple as it, CentOS, Ubuntu, just plain old operating system. Our software takes care of all the prerequisites, installing all the pieces, putting down MetalLB, CoreDNS, Metrics Server, Kubernetes dashboard, etcd backups. You built some servers. That's essentially what you've done, and the rest is being handled by Platform9. It's as simple as that. >> Great, thanks for that. What are the two traditional paths for companies considering the cloud native journey? The two paths. >> The traditional paths. I think that's your engineering team running so fast that before you even realize that you've got, you know, 10 EKS clusters. Or, hey, we can do this. You know, I've got the I can build it mentality. Let's go DIY completely open source Kubernetes on our infrastructure, and we're going to piecemeal build it all up together. They're, I think the pathways that people traditionally look at this journey, as opposed to having that third alternative saying can I just consume it on my infrastructure, be it cloud or on-premise or at the edge. >> Third is the new way, you guys do that. >> That's been our focus since the company was, you know, brought together back in the open OpenStack days. >> Awesome, what's the makeup of your customer base? Is there a certain pattern to the size or environments that you guys work with? Is there a pattern or consistency to your customer base? >> It's a spread, right? We've got large enterprises like Juniper, and we go all the way down to people with 20, 30, 50 nodes in total. We've got people in banking and finance, we've got things all the way through to telecommunications and storage infrastructure. >> What's your favorite feature of Platform9? >> My favorite feature? You know, if I ask, should I say this as a pre-sales engineer, let me show you a favorite thing. My immediate response is, I should never do this. (John laughs) To me it's just being able to define my cluster and say, go. And in five minutes I have that environment, I can see everything that's running, right? It's all unified, it's one spot, right? I'm a cluster admin. I said I wanted three control plane, 25 workers. Here's the infrastructure, it creates it, and once it's built, I can see everything that's running, right? All the applications that are there. One UI, I don't have to go click around. I'm not trying to solve things or download things. It's the fact that it's unified and just delivered in one hit. >> What is the one thing that people should know about Platform9 that they might not know about it? >> I think it's that we help developers and engineers as much as we can help our operations teams. I think, for a long time we've sort of targeted that user and said, hey, we, we really help you. It's like, but why are they doing this? Why are they building any infrastructure or any cloud platform? Well, it's to run applications and services, to help their customers, but how do they get there? There's people building and writing those things, and we're helping them, right? For the last two years, we've been really focused on making it simple, and I think that's an important thing to know. >> Chris, thanks so much, appreciate it. >> Yeah, thank you, John. >> Okay, that's theCUBE Q&A session here with Platform9. I'm John Furrier, thanks for watching. (light music)
SUMMARY :
Chris, I have a couple questions It can, that's the beauty and proactive issue resolution? and that's the control Can you explain the process and go and build the cluster. What are the two traditional paths be it cloud or on-premise or at the edge. the company was, you know, and we go all the way down It's the fact that it's unified For the last two years, Okay, that's theCUBE Q&A
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Chris | PERSON | 0.99+ |
Chris Jones | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
John | PERSON | 0.99+ |
25 workers | QUANTITY | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
five minutes | QUANTITY | 0.99+ |
five | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Platform9 | ORGANIZATION | 0.99+ |
Platform9 | TITLE | 0.99+ |
Juniper | ORGANIZATION | 0.99+ |
Third | QUANTITY | 0.99+ |
CentOS | TITLE | 0.99+ |
Second question | QUANTITY | 0.99+ |
one spot | QUANTITY | 0.99+ |
two paths | QUANTITY | 0.98+ |
Ubuntu | TITLE | 0.97+ |
one hit | QUANTITY | 0.97+ |
20 | QUANTITY | 0.97+ |
10 EKS | QUANTITY | 0.96+ |
One UI | QUANTITY | 0.96+ |
third alternative | QUANTITY | 0.95+ |
Prometheus | TITLE | 0.94+ |
couple questions | QUANTITY | 0.93+ |
50 | QUANTITY | 0.92+ |
two traditional paths | QUANTITY | 0.9+ |
one thing | QUANTITY | 0.89+ |
30 | QUANTITY | 0.86+ |
single node | QUANTITY | 0.85+ |
Kubernetes | TITLE | 0.85+ |
Platform nine | TITLE | 0.82+ |
last two years | DATE | 0.8+ |
CoreDNS | TITLE | 0.78+ |
OpenStack | TITLE | 0.74+ |
three components | QUANTITY | 0.71+ |
three control plane | QUANTITY | 0.7+ |
theCUBE | ORGANIZATION | 0.5+ |
CLI | TITLE | 0.48+ |
CUBE | EVENT | 0.32+ |
Kevin Miller and Ed Walsh | AWS re:Invent 2022 - Global Startup Program
hi everybody welcome back to re invent 2022. this is thecube's exclusive coverage we're here at the satellite set it's up on the fifth floor of the Venetian Conference Center and this is part of the global startup program the AWS startup showcase series that we've been running all through last year and and into this year with AWS and featuring some of its its Global Partners Ed wallson series the CEO of chaos search many times Cube Alum and Kevin Miller there's also a cube Alum vice president GM of S3 at AWS guys good to see you again yeah great to see you Dave hi Kevin this is we call this our Super Bowl so this must be like your I don't know uh World Cup it's a pretty big event yeah it's the World Cup for sure yeah so a lot of S3 talk you know I mean that's what got us all started in 2006 so absolutely what's new in S3 yeah it's been a great show we've had a number of really interesting launches over the last few weeks and a few at the show as well so you know we've been really focused on helping customers that are running Mass scale data Lakes including you know whether it's structured or unstructured data we actually announced just a few just an hour ago I think it was a new capability to give customers cross-account access points for sharing data securely with other parts of the organization and that's something that we'd heard from customers is as they are growing and have more data sets and they're looking to to get more out of their data they are increasingly looking to enable multiple teams across their businesses to access those data sets securely and that's what we provide with cross-count access points we also launched yesterday our multi-region access point failover capabilities and so again this is where customers have data sets and they're using multiple regions for certain critical workloads they're now able to to use that to fail to control the failover between different regions in AWS and then one other launch I would just highlight is some improvements we made to storage lens which is our really a very novel and you need capability to help customers really understand what storage they have where who's accessing it when it's being accessed and we added a bunch of new metrics storage lens has been pretty exciting for a lot of customers in fact we looked at the data and saw that customers who have adopted storage lens typically within six months they saved more than six times what they had invested in turning storage lens on and certainly in this environment right now we have a lot of customers who are it's pretty top of mind they're looking for ways to optimize their their costs in the cloud and take some of those savings and be able to reinvest them in new innovation so pretty exciting with the storage lens launch I think what's interesting about S3 is that you know pre-cloud Object Store was this kind of a niche right and then of course you guys announced you know S3 in 2006 as I said and okay great you know cheap and deep storage simple get put now the conversations about how to enable value from from data absolutely analytics and it's just a whole new world and Ed you've talked many times I love the term yeah we built chaos search on the on the shoulders of giants right and so the under underlying that is S3 but the value that you can build on top of that has been key and I don't think we've talked about his shoulders and Giants but we've talked about how we literally you know we have a big Vision right so hard to kind of solve the challenge to analytics at scale we really focus on the you know the you know Big Data coming environment get analytics so we talk about the on the shoulders Giants obviously Isaac Newton's you know metaphor of I learned from everything before and we layer on top so really when you talk about all the things come from S3 like I just smile because like we picked it up naturally we went all in an S3 and this is where I think you're going Dave but everyone is so let's just cut the chase like so any of the data platforms you're using S3 is what you're building but we did it a little bit differently so at first people using a cold storage like you said and then they ETL it up into a different platforms for analytics of different sorts now people are using it closer they're doing caching layers and cashing out and they're that's where but that's where the attributes of a scale or reliability are what we did is we actually make S3 a database so literally we have no persistence outside that three and that kind of comes in so it's working really well with clients because most of the thing is we pick up all these attributes of scale reliability and it shows up in the clients environments and so when you launch all these new scalable things we just see it like our clients constantly comment like one of our biggest customers fintech in uh Europe they go to Black Friday again black Friday's not one days and they lose scale from what is it 58 terabytes a day and they're going up to 187 terabytes a day and we don't Flinch they say how do you do that well we built our platform on S3 as long as you can stream it to S3 so they're saying I can't overrun S3 and it's a natural play so it's it's really nice that but we take out those attributes but same thing that's why we're able to you know help clients get you know really you know Equifax is a good example maybe they're able to consolidate 12 their divisions on one platform we couldn't have done that without the scale and the performance of what you can get S3 but also they saved 90 I'm able to do that but that's really because the only persistence is S3 and what you guys are delivering but and then we really for focus on shoulders Giants we're doing on top of that innovating on top of your platforms and bringing that out so things like you know we have a unique data representation that makes it easy to ingest this data because it's kind of coming at you four v's of big data we allow you to do that make it performant on s3h so now you're doing hot analytics on S3 as if it's just a native database in memory but there's no memory SSC caching and then multi-model once you get it there don't move it leverage it in place so you know elasticsearch access you know Cabana grafana access or SQL access with your tools so we're seeing that constantly but we always talk about on the shoulders of giants but even this week I get comments from our customers like how did you do that and most of it is because we built on top of what you guys provided so it's really working out pretty well and you know we talk a lot about digital transformation of course we had the pleasure sitting down with Adam solipski prior John Furrier flew to Seattle sits down his annual one-on-one with the AWS CEO which is kind of cool yeah it was it's good it's like study for the test you know and uh and so but but one of the interesting things he said was you know we're one of our challenges going forward is is how do we go Beyond digital transformation into business transformation like okay well that's that's interesting I was talking to a customer today AWS customer and obviously others because they're 100 year old company and they're basically their business was they call them like the Uber for for servicing appliances when your Appliance breaks you got to get a person to serve it a service if it's out of warranty you know these guys do that so they got to basically have a you know a network of technicians yeah and they gotta deal with the customers no phone right so they had a completely you know that was a business transformation right they're becoming you know everybody says they're coming a software company but they're building it of course yeah right on the cloud so wonder if you guys could each talk about what's what you're seeing in terms of changing not only in the sort of I.T and the digital transformation but also the business transformation yeah I know I I 100 agree that I think business transformation is probably that one of the top themes I'm hearing from customers of all sizes right now even in this environment I think customers are looking for what can I do to drive top line or you know improve bottom line or just improve my customer experience and really you know sort of have that effect where I'm helping customers get more done and you know it is it is very tricky because to do that successfully the customers that are doing that successfully I think are really getting into the lines of businesses and figuring out you know it's probably a different skill set possibly a different culture different norms and practices and process and so it's it's a lot more than just a like you said a lot more than just the technology involved but when it you know we sort of liquidate it down into the data that's where absolutely we see that as a critical function for lines of businesses to become more comfortable first off knowing what data sets they have what data they they could access but possibly aren't today and then starting to tap into those data sources and then as as that progresses figuring out how to share and collaborate with data sets across a company to you know to correlate across those data sets and and drive more insights and then as all that's being done of course it's important to measure the results and be able to really see is this what what effect is this having and proving that effect and certainly I've seen plenty of customers be able to show you know this is a percentage increase in top or bottom line and uh so that pattern is playing out a lot and actually a lot of how we think about where we're going with S3 is related to how do we make it easier for customers to to do everything that I just described to have to understand what data they have to make it accessible and you know it's great to have such a great ecosystem of partners that are then building on top of that and innovating to help customers connect really directly with the businesses that they're running and driving those insights well and customers are hours today one of the things I loved that Adam said he said where Amazon is strategically very very patient but tactically we're really impatient and the customers out there like how are you going to help me increase Revenue how are you going to help me cut costs you know we were talking about how off off camera how you know software can actually help do that yeah it's deflationary I love the quote right so software's deflationary as costs come up how do you go drive it also free up the team and you nail it it's like okay everyone wants to save money but they're not putting off these projects in fact the digital transformation or the business it's actually moving forward but they're getting a little bit bigger but everyone's looking for creative ways to look at their architecture and it becomes larger larger we talked about a couple of those examples but like even like uh things like observability they want to give this tool set this data to all the developers all their sres same data to all the security team and then to do that they need to find a way an architect should do that scale and save money simultaneously so we see constantly people who are pairing us up with some of these larger firms like uh or like keep your data dog keep your Splunk use us to reduce the cost that one and one is actually cheaper than what you have but then they use it either to save money we're saving 50 to 80 hard dollars but more importantly to free up your team from the toil and then they they turn around and make that budget neutral and then allowed to get the same tools to more people across the org because they're sometimes constrained of getting the access to everyone explain that a little bit more let's say I got a Splunk or data dog I'm sifting through you know logs how exactly do you help so it's pretty simple I'll use dad dog example so let's say using data dog preservability so it's just your developers your sres managing environments all these platforms are really good at being a monitoring alerting type of tool what they're not necessarily great at is keeping the data for longer periods like the log data the bigger data that's where we're strong what you see is like a data dog let's say you're using it for a minister for to keep 30 days of logs which is not enough like let's say you're running environment you're finding that performance issue you kind of want to look to last quarter in last month in or maybe last Black Friday so 30 days is not enough but will charge you two eighty two dollars and eighty cents a gigabyte don't focus on just 280 and then if you just turn the knob and keep seven days but keep two years of data on us which is on S3 it goes down to 22 cents plus our list price of 80 cents goes to a dollar two compared to 280. so here's the thing what they're able to do is just turn a knob get more data we do an integration so you can go right from data dog or grafana directly into our platform so the user doesn't see it but they save money A lot of times they don't just save the money now they use that to go fund and get data dog to a lot more people make sense so it's a creativity they're looking at it and they're looking at tools we see the same thing with a grafana if you look at the whole grafana play which is hey you can't put it in one place but put Prometheus for metrics or traces we fit well with logs but they're using that to bring down their costs because a lot of this data just really bogs down these applications the alerting monitoring are good at small data they're not good at the big data which is what we're really good at and then the one and one is actually less than you paid for the one so it and it works pretty well so things are really unpredictable right now in the economy you know during the pandemic we've sort of lockdown and then the stock market went crazy we're like okay it's going to end it's going to end and then it looked like it was going to end and then it you know but last year it reinvented just just in that sweet spot before Omicron so we we tucked it in which which was awesome right it was a great great event we really really missed one physical reinvent you know which was very rare so that's cool but I've called it the slingshot economy it feels like you know you're driving down the highway and you got to hit the brakes and then all of a sudden you're going okay we're through it Oh no you're gonna hit the brakes again yeah so it's very very hard to predict and I was listening to jassy this morning he was talking about yeah consumers they're still spending but what they're doing is they're they're shopping for more features they might be you know buying a TV that's less expensive you know more value for the money so okay so hopefully the consumer spending will get us out of this but you don't really know you know and I don't yeah you know we don't seem to have the algorithms we've never been through something like this before so what are you guys seeing in terms of customer Behavior given that uncertainty well one thing I would highlight that I think particularly going back to what we were just talking about as far as business and digital transformation I think some customers are still appreciating the fact that where you know yesterday you may have had to to buy some Capital put out some capital and commit to something for a large upfront expenditure is that you know today the value of being able to experiment and scale up and then most importantly scale down and dynamically based on is the experiment working out am I seeing real value from it and doing that on a time scale of a day or a week or a few months that is so important right now because again it gets to I am looking for a ways to innovate and to drive Top Line growth but I I can't commit to a multi-year sort of uh set of costs to to do that so and I think plenty of customers are finding that even a few months of experimentation gives them some really valuable insight as far as is this going to be successful or not and so I think that again just of course with S3 and storage from day one we've been elastic pay for what you use if you're not using the storage you don't get charged for it and I think that particularly right now having the applications and the rest of the ecosystem around the storage and the data be able to scale up and scale down is is just ever more important and when people see that like typically they're looking to do more with it so if they find you usually find these little Department projects but they see a way to actually move faster and save money I think it is a mix of those two they're looking to expand it which can be a nightmare for sales Cycles because they take longer but people are looking well why don't you leverage this and go across division so we do see people trying to leverage it because they're still I don't think digital transformation is slowing down but a lot more to be honest a lot more approvals at this point for everything it is you know Adam and another great quote in his in his keynote he said if you want to save money the Cloud's a place to do it absolutely and I read an article recently and I was looking through and I said this is the first time you know AWS has ever seen a downturn because the cloud was too early back then I'm like you weren't paying attention in 2008 because that was the first major inflection point for cloud adoption where CFO said okay stop the capex we're going to Opex and you saw the cloud take off and then 2010 started this you know amazing cycle that we really haven't seen anything like it where they were doubling down in Investments and they were real hardcore investment it wasn't like 1998 99 was all just going out the door for no clear reason yeah so that Foundation is now in place and I think it makes a lot of sense and it could be here for for a while where people are saying Hey I want to optimize and I'm going to do that on the cloud yeah no I mean I've obviously I certainly agree with Adam's quote I think really that's been in aws's DNA from from day one right is that ability to scale costs with with the actual consumption and paying for what you use and I think that you know certainly moments like now are ones that can really motivate change in an organization in a way that might not have been as palatable when it just it didn't feel like it was as necessary yeah all right we got to go give you a last word uh I think it's been a great event I love all your announcements I think this is wonderful uh it's been a great show I love uh in fact how many people are here at reinvent north of 50 000. yeah I mean I feel like it was it's as big if not bigger than 2019. people have said ah 2019 was a record when you count out all the professors I don't know it feels it feels as big if not bigger so there's great energy yeah it's quite amazing and uh and we're thrilled to be part of it guys thanks for coming on thecube again really appreciate it face to face all right thank you for watching this is Dave vellante for the cube your leader in Enterprise and emerging Tech coverage we'll be right back foreign
SUMMARY :
across a company to you know to
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Ed Walsh | PERSON | 0.99+ |
Kevin Miller | PERSON | 0.99+ |
two years | QUANTITY | 0.99+ |
2006 | DATE | 0.99+ |
2008 | DATE | 0.99+ |
seven days | QUANTITY | 0.99+ |
Adam | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
30 days | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
50 | QUANTITY | 0.99+ |
Adam solipski | PERSON | 0.99+ |
Dave vellante | PERSON | 0.99+ |
two | QUANTITY | 0.99+ |
eighty cents | QUANTITY | 0.99+ |
Europe | LOCATION | 0.99+ |
22 cents | QUANTITY | 0.99+ |
Kevin | PERSON | 0.99+ |
80 cents | QUANTITY | 0.99+ |
Seattle | LOCATION | 0.99+ |
12 | QUANTITY | 0.99+ |
2010 | DATE | 0.99+ |
Isaac Newton | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Super Bowl | EVENT | 0.99+ |
a day | QUANTITY | 0.99+ |
Venetian Conference Center | LOCATION | 0.99+ |
fifth floor | QUANTITY | 0.99+ |
Uber | ORGANIZATION | 0.99+ |
World Cup | EVENT | 0.99+ |
last year | DATE | 0.99+ |
last quarter | DATE | 0.99+ |
yesterday | DATE | 0.99+ |
S3 | TITLE | 0.99+ |
last month | DATE | 0.99+ |
more than six times | QUANTITY | 0.99+ |
2019 | DATE | 0.98+ |
Prometheus | TITLE | 0.98+ |
six months | QUANTITY | 0.98+ |
280 | QUANTITY | 0.98+ |
pandemic | EVENT | 0.98+ |
Black Friday | EVENT | 0.97+ |
an hour ago | DATE | 0.97+ |
today | DATE | 0.97+ |
58 terabytes a day | QUANTITY | 0.97+ |
100 year old | QUANTITY | 0.97+ |
this morning | DATE | 0.97+ |
a week | QUANTITY | 0.97+ |
Ed wallson | PERSON | 0.97+ |
three | QUANTITY | 0.96+ |
Equifax | ORGANIZATION | 0.96+ |
jassy | PERSON | 0.96+ |
one platform | QUANTITY | 0.96+ |
this year | DATE | 0.96+ |
grafana | TITLE | 0.96+ |
one days | QUANTITY | 0.95+ |
first time | QUANTITY | 0.95+ |
one | QUANTITY | 0.95+ |
black Friday | EVENT | 0.93+ |
this week | DATE | 0.92+ |
first major inflection | QUANTITY | 0.91+ |
one place | QUANTITY | 0.91+ |
SQL | TITLE | 0.9+ |
last | DATE | 0.89+ |
Store | TITLE | 0.89+ |
Austin Parker, Lightstep | AWS re:Invent 2022
(lively music) >> Good afternoon cloud community and welcome back to beautiful Las Vegas, Nevada. We are here at AWS re:Invent, day four of our wall to wall coverage. It is day four in the afternoon and we are holding strong. I'm Savannah Peterson, joined by my fabulous co-host Paul Gillen. Paul, how you doing? >> I'm doing well, fine Savannah. You? >> You look great. >> We're in the home stretch here. >> Yeah, (laughs) we are. >> You still look fresh as a daisy. I don't know how you do it. >> (laughs) You're too kind. You're too kind, but I'm vain enough to take that compliment. I'm very excited about the conversation that we're going to have up next. We get to get a little DevRel and we got a little swagger on the stage. Welcome, Austin. How you doing? >> Hey, great to be here. Thanks for having me. >> Savannah: Yeah, it's our pleasure. How's the show been for you so far? >> Busy, exciting. Feels a lot like, you know it used to be right? >> Yeah, I know. A little reminiscent of the before times. >> Well, before times. >> Before we dig into the technical stuff, you're the most intriguingly dressed person we've had on the show this week. >> Austin: I feel extremely underdressed. >> Well, and we were talking about developer fancy. Talk to me a little bit about your approach to fashion. Wasn't expecting to lead with this, but I like this but I like this actually. >> No, it's actually good with my PR. You're going to love it. My approach, here's the thing, I give free advice all the time about developer relations, about things that work, have worked, and don't work in community and all that stuff. I love talking about that. Someone came up to me and said, "Where do you get your fashion tips from? What's the secret Discord server that I need to go on?" I'm like, "I will never tell." >> Oh, okay. >> This is an actual trait secret. >> Top secret. Wow! Talk about. >> If someone else starts wearing the hat, then everyone's going to be like, "There's so many white guys." Look, I'm a white guy with a beard that works in technology. >> Savannah: I've never met one of those. >> Exactly, there's none of them at all. So, you have to do something to kind stand out from the crowd a little bit. >> I love it, and it's a talk trigger. We're talking about it now. Production team loved it. It's fantastic. >> It's great. >> So your DevRel for Lightstep, in case the audience isn't familiar tell us about Lightstep. >> So Lightstep is a cloud native observability platform built at planet scale, and it powers observability at some places you've heard of like Spotify, GitHub, right? We're designed to really help developers that are working in the cloud with Kubernetes, with these huge distributed systems, understand application performance and being able to find problems, fix problems. We're also part of the ServiceNow family and as we all know ServiceNow is on a mission to help the world of work work better by powering digital transformation around IT and customer experiences for their many, many, many global 2000 customers. We love them very much. >> You know, it's a big love fest here. A lot of people have talked about the collaboration, so many companies working together. You mentioned unified observability. What is unified observability? >> So if you think about a tradition, or if you've heard about this traditional idea of observability where you have three pillars, right? You have metrics, and you have logs, and you have traces. All those three things are different data sources. They're picked up by different tools. They're analyzed by different people for different purposes. What we believe and what we're working to accomplish right now is to take all that and if you think those pillars, flip 'em on their side and think of them as streams of data. If we can take those streams and integrate them together and let you treat traces and metrics and logs not as these kind of inviolate experiences where you're kind of paging between things and going between tab A to tab B to tab C, and give you a standard way to query this, a standard way to display this, and letting you kind of find the most relevant data, then it really unlocks a lot of power for like developers and SREs to spend less time like managing tools. You know, figuring out where to build their query or what dashboard to check, more just being able to like kind of ask a question, get an answer. When you have an incident or an outage that's the most important thing, right? How quickly can you get those answers that you need so that you can restore system health? >> You don't want to be looking in multiple spots to figure out what's going on. >> Absolutely. I mean, some people hear unified observability and they go to like tool consolidation, right? That's something I hear from a lot of our users and a lot of people in re:Invent. I'll talk to SREs, they're like, "Yeah, we've got like six or seven different metrics products alone, just on services that they cover." It is important to kind of consolidate that but we're really taking it a step lower. We're looking at the data layer and trying to say, "Okay, if the data is all consistent and vendor neutral then that gives you flexibility not only from a tool consolidation perspective but also you know, a consistency, reliability. You could have a single way to deploy your observability out regardless of what cloud you're on, regardless if you're using Kubernetes or Fargate or whatever else. or even just Bare Metal or EC2 Bare Metal, right? There's been so much historically in this space. There's been a lot of silos and we think that unify diversability means that we kind of break down those silos, right? The way that we're doing it primarily is through a project called OpenTelemetry which you might have heard of. You want to talk about that in a minute? . >> Savannah: Yeah, let's talk about it right now. Why don't you tell us about it? Keep going, you're great. You're on a roll. >> I am. >> Savannah: We'll just hang out over here. >> It's day four. I'm going to ask the questions and answer the questions. (Savannah laughs) >> Yes, you're right. >> I do yeah. >> Open Tele- >> OpenTelemetry . >> Explain what OpenTelemetry is first. >> OpenTelemetry is a CNCF project, Cloud Native Computing Foundation. The goal is to make telemetry data, high quality telemetry data, a builtin feature of cloud native software right? So right now if you wanted to get logging data out, depending on your application stack, depending on your application run time, depending on language, depending on your deployment environment. You might have a lot... You have to make a lot of choices, right? About like, what am I going to use? >> Savannah: So many different choices, and the players are changing all the time. >> Exactly, and a lot of times what people will do is they'll go and they'll say like, "We have to use this commercial solution because they have a proprietary agent that can do a lot of this for us." You know? And if you look at all those proprietary agents, what you find very quickly is it's very commodified right? There's no real difference in what they're doing at a code level and what's stopped the industry from really adopting a standard way to create this logs and metrics and traces, is simply just the fact that there was no standard. And so, OpenTelemetry is that standard, right? We've got dozens of companies many of them like very, many of them here right? Competitors all the same, working together to build this open standard and implementation of telemetry data for cloud native software and really any software right? Like we support over 12 languages. We support Kubernetes, Amazon. AWS is a huge contributor actually and we're doing some really exciting stuff with them on their Amazon distribution of OpenTelemetry. So it's been extremely interesting to see it over the past like couple years go from like, "Hey, here's this like new thing that we're doing over here," to really it's a generalized acceptance that this is the way of the future. This is what we should have been doing all along. >> Yeah. >> My opinion is there is a perception out there that observability is kind of a commodity now that all the players have the same set of tools, same set of 15 or 17 or whatever tools, and that there's very little distinction in functionality. Would you agree with that? >> I don't know if I would characterize it that way entirely. I do think that there's a lot of duplicated effort that happens and part of the reason is because of this telemetry data problem, right? Because you have to wind up... You know, there's this idea of table stakes monitoring that we talk about right? Table stakes monitoring is the stuff that you're having to do every single day to kind of make sure your system is healthy to be able to... When there's an alert, gets triggered, to see why it got triggered and to go fix it, right? Because everyone has the kind of work on that table stake stuff and then build all these integrations, there's very little time for innovation on top of that right? Because you're spending all your time just like working on keeping up with technology. >> Savannah: Doing the boring stuff to make sure the wheels don't fall off, basically. >> Austin: Right? What I think the real advantage of OpenTelemetry is that it really, from like a vendor perspective, like it unblocks us from having to kind of do all this repetitive commodified work. It lets us help move that out to the community level so that... Instead of having to kind of build, your Kubernetes integration for example, you can just have like, "Hey, OpenTelemetry is integrated into Kubernetes and you just have this data now." If you are a commercial product, or if you're even someone that's interested in fixing a, scratching a particular itch about observability. It's like, "I have this specific way that I'm doing Kubernetes and I need something to help me really analyze that data. Well, I've got the data now I can just go create a project. I can create an analysis tool." I think that's what you'll see over time as OpenTelemetry promulgates out into the ecosystem is more people building interesting analysis features, people using things like machine learning to analyze this large amount, large and consistent amount of OpenTelemetry data. It's going to be a big shakeup I think, but it has the potential to really unlock a lot of value for our customers. >> Well, so you're, you're a developer relations guy. What are developers asking for right now out of their observability platforms? >> Austin: That's a great question. I think there's two things. The first is that they want it to just work. It's actually the biggest thing, right? There's so many kind of... This goes back to the tool proliferation, right? People have too much data in too many different places, and getting that data out can still be really challenging. And so, the biggest thing they want is just like, "I want something that I can... I want a lot of these questions I have to ask, answered already and OpenTelemetry is going towards it." Keep in mind it's the project's only three years old, so we obviously have room to grow but there are people running it in production and it works really well for them but there's more that we can do. The second thing is, and this isn't what really is interesting to me, is it's less what they're asking for and more what they're not asking for. Because a lot of the stuff that you see people, saying around, "Oh, we need this like very specific sort of lower level telemetry data, or we need this kind of universal thing." People really just want to be able to get questions or get questions answered, right? They want tools that kind of have these workflows where you don't have to be an expert because a lot of times this tooling gets locked behind sort of is gate kept almost in a organization where there are teams that's like, "We're responsible for this and we're going to set it up and manage it for you, and we won't let you do things outside of it because that would mess up- >> Savannah: Here's your sandbox and- >> Right, this is your sandbox you can play in and a lot of times that's really useful and very tuned for the problems that you saw yesterday, but people are looking at like what are the problems I'm going to get tomorrow? We're deploying more rapidly. We have more and more intentional change happening in the system. Like it's not enough to have this reactive sort of approach where our SRE teams are kind of like or this observability team is building a platform for us. Developers want to be able to get in and have these kind of guided workflows really that say like, "Hey, here's where you're starting at. Let's get you to an answer. Let's help you find the needle in the haystack as it were, without you having to become a master of six different or seven different tools." >> Savannah: Right, and it shouldn't be that complicated. >> It shouldn't be. I mean we've certainly... We've been working on this problem for many years now, starting with a lot of our team that started at Google and helped build Google's planet scale monitoring systems. So we have a lot of experience in the field. It's actually one... An interesting story that our founder or now general manager tells BHS, Ben Sigelman, and he told me this story once and it's like... He had built this really cool thing called Dapper that was a tracing system at Google, and people weren't using it. Because they were like, "This is really cool, but I don't know how to... but it's not relevant to me." And he's like, the one thing that we did to get to increase usage 20 times over was we just put a link. So we went to the place that people were already looking for that data and we added a link that says, "Hey, go over here and look at this." It's those simple connections being able to kind of draw people from like point A to point B, take them from familiar workflows into unfamiliar ones. You know, that's how we think about these problems right? How is this becoming a daily part of someone's usage? How is this helping them solve problems faster and really improve their their life? >> Savannah: Yeah, exactly. It comes down to quality of life. >> Warner made the case this morning that computer architecture should be inherently event-driven and that we are moving toward a world where the person matters less than what the software does, right? The software is triggering events. Does this complicate observability or simplify it? >> Austin: I think that at the end of the day, it's about getting the... Observability to me in a lot of ways is about modeling your system, right? It's about you as a developer being able to say this is what I expect the system to do and I don't think the actual application architecture really matters that much, right? Because it's about you. You are building a system, right? It can be event driven, can be support request response, can be whatever it is. You have to be able to say, "This is what I expect to... For these given inputs, this is the expected output." Now maybe there's a lot of stuff that happens in the middle that you don't really care about. And then, I talk to people here and everyone's talking about serverless right? Everyone... You can see there's obviously some amazing statistics about how many people are using Lambda, and it's very exciting. There's a lot of stuff that you shouldn't have to care about as a developer, but you should care about those inputs and outputs. You will need to have that kind of intermediate information and understand like, what was the exact path that I took through this invented system? What was the actual resources that were being used? Because even if you trust that all this magic behind the scenes is just going to work forever, sometimes it's still really useful to have that sort of lower level abstraction, to say like, "Well, this is what actually happened so that I can figure out when I deployed a new change, did I make performance better or worse?" Or being able to kind of segregate your data out and say like... Doing AB testing, right? Doing canary releases, doing all of these things that you hear about as best practices or well architected applications. Observability is at the core of all that. You need observability to kind of do any of, ask any of those higher level interesting questions. >> Savannah: We are here at ReInvent. Tell us a little bit more about the partnership with AWS. >> So I would have to actually probably refer you to someone at Service Now on that. I know that we are a partner. We collaborate with them on various things. But really at Lightstep, we're very focused on kind of the open source part of this. So we work with AWS through the OpenTelemetry project, on things like the AWS distribution for OpenTelemetry which is really... It's OpenTelemetry, again is really designed to be like a neutral standard but we know that there are going to be integrators and implementers that need to package up and bundle it in a certain way to make it easy for their end users to consume it. So that's what Amazon has done with ADOT which is the shortening for it. So it's available in several different ways. You can use it as like an SDK and drop it into your application. There's Lambda layers. If you want to get Lambda observability, you just add this extension in and then suddenly you're getting OpenTelemetry data on the other side. So it's really cool. It's been a really exciting to kind of work with people on the AWS side over the past several years. >> Savannah: It's exciting, >> I've personally seen just a lot of change. I was talking to a PM earlier this week... It's like, "Hey, two years ago I came and talked to you about OpenTelemetry and here we are today. You're still talking about OpenTelemetry." And they're like, "What changes?" Our customers have started coming to us asking for OpenTelemetry and we see the same thing now. >> Savannah: Timing is right. >> Timing is right, but we see the same thing... Even talking to ServiceNow customers who are... These very big enterprises, banks, finance, healthcare, whatever, telcos, it used to be... You'd have to go to them and say like, "Let me tell you about distributed tracing. Let me tell you about OpenTelemetry. Let me tell you about observability." Now they're coming in and saying, "Yeah, so we're standard." If you think about Kubernetes and how Kubernetes, a lot of enterprises have spent the past five-six years standardizing, and Kubernetes is a way to deploy applications or manage containerized applications. They're doing the same journey now with OpenTelemetry where they're saying, "This is what we're betting on and we want partners we want people to help us go along that way." >> I love it, and they work hand in hand in all CNCF projects as well that you're talking about. >> Austin: Right, so we're integrated into Kubernetes. You can find OpenTelemetry and things like kept in which is application standards. And over time, it'll just like promulgate out from there. So it's really exciting times. >> A bunch of CNCF projects in this area right? Prometheus. >> Prometheus, yeah. Yeah, so we inter-operate with Prometheus as well. So if you have Prometheus metrics, then OpenTelemetry can read those. It's a... OpenTelemetry metrics are like a super set of Prometheus. We've been working with the Prometheus community for quite a while to make sure that there's really good compatibility because so many people use Prometheus you know? >> Yeah. All right, so last question. New tradition for us here on theCUBE. We're looking for your 32nd hot take, Instagram reel, biggest theme, biggest buzz for those not here on the show floor. >> Oh gosh. >> Savannah: It could be for you too. It could be whatever for... >> I think the two things that are really striking to me is one serverless. Like I see... I thought people were talking about servers a lot and they were talking about it more than ever. Two, I really think it is observability right? Like we've gone from observability being kind of a niche. >> Savannah: Not that you're biased. >> Huh? >> Savannah: Not that you're biased. >> Not that I'm biased. It used to be a niche. I'd have to go niche thing where I would go and explain what this is to people and nowpeople are coming up. It's like, "Yeah, yeah, we're using OpenTelemetry." It's very cool. I've been involved with OpenTelemetry since the jump, since it was started really. It's been very exciting to see and gratifying to see like how much adoption we've gotten even in a short amount of time. >> Yeah, absolutely. It's a pretty... Yeah, it's been a lot. That was great. Perfect soundbite for us. >> Austin: Thanks, I love soundbites. >> Savannah: Yeah. Awesome. We love your hat and your soundbites equally. Thank you so much for being on the show with us today. >> Thank you for having me. >> Savannah: Hey, anytime, anytime. Will we see you in Amsterdam, speaking of KubeCon? Awesome, we'll be there. >> There's some real exciting OpenTelemetry stuff coming up for KubeCon. >> Well, we'll have to get you back on theCUBE. (talking simultaneously) Love that for us. Thank you all for tuning in two hour wall to wall coverage here, day four at AWS re:Invent in fabulous Las Vegas, Nevada, with Paul Gillin. I'm Savannah Peterson and you're watching theCUBE, the leader in high tech coverage. (lively music)
SUMMARY :
and we are holding strong. I'm doing well, fine Savannah. I don't know how you do it. and we got a little swagger on the stage. Hey, great to be here. How's the show been for you so far? Feels a lot like, you A little reminiscent of the before times. on the show this week. Well, and we were talking server that I need to go on?" Talk about. then everyone's going to be like, something to kind stand out and it's a talk trigger. in case the audience isn't familiar and being able to find about the collaboration, and going between tab A to tab B to tab C, in multiple spots to and they go to like tool Why don't you tell us about it? Savannah: We'll just and answer the questions. The goal is to make telemetry data, and the players are changing all the time. Exactly, and a lot of and that there's very little and part of the reason is because of this boring stuff to make sure but it has the potential to really unlock What are developers asking for right now and we won't let you for the problems that you saw yesterday, Savannah: Right, and it And he's like, the one thing that we did It comes down to quality of life. and that we are moving toward a world is just going to work forever, about the partnership with AWS. that need to package up and talked to you about OpenTelemetry and Kubernetes is a way and they work hand in hand and things like kept in which A bunch of CNCF projects So if you have Prometheus metrics, We're looking for your 32nd hot take, Savannah: It could be for you too. that are really striking to me and gratifying to see like It's a pretty... on the show with us today. Will we see you in Amsterdam, OpenTelemetry stuff coming up I'm Savannah Peterson and
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Peter Burris | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Michael Dell | PERSON | 0.99+ |
Rebecca Knight | PERSON | 0.99+ |
Michael | PERSON | 0.99+ |
Comcast | ORGANIZATION | 0.99+ |
Elizabeth | PERSON | 0.99+ |
Paul Gillan | PERSON | 0.99+ |
Jeff Clark | PERSON | 0.99+ |
Paul Gillin | PERSON | 0.99+ |
Nokia | ORGANIZATION | 0.99+ |
Savannah | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Richard | PERSON | 0.99+ |
Micheal | PERSON | 0.99+ |
Carolyn Rodz | PERSON | 0.99+ |
Dave Vallante | PERSON | 0.99+ |
Verizon | ORGANIZATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Eric Seidman | PERSON | 0.99+ |
Paul | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
ORGANIZATION | 0.99+ | |
Keith | PERSON | 0.99+ |
Chris McNabb | PERSON | 0.99+ |
Joe | PERSON | 0.99+ |
Carolyn | PERSON | 0.99+ |
Qualcomm | ORGANIZATION | 0.99+ |
Alice | PERSON | 0.99+ |
2006 | DATE | 0.99+ |
John | PERSON | 0.99+ |
Netflix | ORGANIZATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
congress | ORGANIZATION | 0.99+ |
Ericsson | ORGANIZATION | 0.99+ |
AT&T | ORGANIZATION | 0.99+ |
Elizabeth Gore | PERSON | 0.99+ |
Paul Gillen | PERSON | 0.99+ |
Madhu Kutty | PERSON | 0.99+ |
1999 | DATE | 0.99+ |
Michael Conlan | PERSON | 0.99+ |
2013 | DATE | 0.99+ |
Michael Candolim | PERSON | 0.99+ |
Pat | PERSON | 0.99+ |
Yvonne Wassenaar | PERSON | 0.99+ |
Mark Krzysko | PERSON | 0.99+ |
Boston | LOCATION | 0.99+ |
Pat Gelsinger | PERSON | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
Willie Lu | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Yvonne | PERSON | 0.99+ |
Hertz | ORGANIZATION | 0.99+ |
Andy | PERSON | 0.99+ |
2012 | DATE | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Richard Hartmann, Grafana Labs | KubeCon + CloudNativeCon NA 2022
>>Good afternoon everyone, and welcome back to the Cube. I am Savannah Peterson here, coming to you from Detroit, Michigan. We're at Cuban Day three. Such a series of exciting interviews. We've done over 30, but this conversation is gonna be extra special, don't you think, John? >>Yeah, this is gonna be a good one. Griffon Labs is here with us. We're getting the conversation of what's going on in the industry management, watching the Kubernetes clusters. This is large scale conversations this week. It's gonna be a good one. >>Yeah. Yeah. I'm very excited. He's also got a fantastic Twitter handle, twitchy. H Please welcome Richie Hartman, who is the director of community here at Griffon. Richie, thank you so much for joining us. Thanks >>For having me. >>How's the show been for you? >>Busy. I, I mean, I, I, >>In >>A word, I have a ton of talks at at like maintain a thing and like the covering board searches at the TLC panel. I run forme day. So it's, it's been busy. It, yeah. Monday, I didn't have to run anything. That was quite nice. But there >>You, you have your hands in a lot. I'm not even gonna cover it. Looking at your bio, there's, there's so many different things that you're working on. I know that Grafana specifically had some announcements this week. Yeah, >>Yeah, yeah. We had quite a few, like the, the two largest ones is a, we now have a field Kubernetes integration on Grafana Cloud. So our, our approach is generally extremely open source first. So we try to push stuff into the exporters, like into the open source exporters, into mixes into things which are out there as open source for anyone to use. But that's little bit like a tool set, not a ready made solution. So when we talk integrations, we actually talk about things where you get this like one click experience, You log into your Grafana cloud, you click, I have a Kubernetes, which probably most of us have, and things just work like you in just the data. You have to write dashboards, you have to write alerts, you have to write everything to just get started with extremely opinionated dashboards, SLOs, alerts, again, all those things made by experts, so anyone can use them. And you don't have to reinvent the view for every single user. So that's the one. The other is, >>It's a big deal. >>Oh yeah, it is. Yeah. It is. It, we, we has, its heavily in integrations course. While, I mean, I don't have to convince anyone that perme is a DD factor standard in everything. Cloudnative. But again, it's, it's, it's sometimes a little bit hard to handle or a little bit not easy to get into. So, so smoothing this, this, this path onto onboarding yourself onto this stack and onto those types of solutions. Yes. Is what a lot of people need. Course, if you, if you look at the statistics from coupon, and we just heard this in the governing board session yesterday. Yeah. Like 60% of the people here are first time attendees. So there's a lot of people who just come into this thing and who need, like, this is your path. This is where you should be going. Or at least if you want to go, go there. This is how to get there. >>Here's your runway for takeoff. Yes. Yeah. I think that's a really good point. And I love that you, you had those numbers. I was curious. I, I had seen on Twitter, speaking of Twitter, I had seen, I had seen that, that there were a lot of people here coming for the first time. You're a community guy. Are we at an inflection point where this community is about to continue to scale? >>That's a very good question. Which I can't really answer. So I mean, >>Obviously I bet you're gonna try. >>I covid changed a few things. Yeah. Probably most people, >>A couple things. I mean, you know, casually, it's like such a gentle way of putting that, that was >>Beautiful. I'm gonna say yes, just to explode. All these new ERs are gonna learn Prometheus. They're gonna roll in with a open, open metrics, open telemetry. I love it, >>You know, But, but at the same time, like Cuban is, is ramping back up. But if you look at the, if you look at the registration numbers between Valencia Andro, it was more or less the same. Interesting. Which, so it didn't go onto this, onto this flu trajectory, which it was on like, up to, up to 2019. I expect this to take up again. But also with the economic situation, everything, I, I don't think >>It's, I think the jury's still out on hybrid. I think there's a lot, lot more hybrid. Let's see how the projects are gonna go. That's what I think it's gonna be the tell sign. How many people are in participating? How are the project's advancing? Some of the momentum, >>I mean, from the project level, Most of this is online anyway. Of course. That's how open source, right. I've been working for >>Ages. That's >>Cause you don't have any trouble budget or, or any office or, It's >>Always been that way. >>Yeah, precisely. So the projects are arguably spearheading this, this development and the, the online numbers. I I, I have some numbers in my head, but I'm, I'm not a hundred percent certain to, but they're higher for this time in Detroit than in volunteer as far somewhere. Cool. So that is growing and it's grown in parallel, which also is great. Cause it's much more accessible, much more inclusive. You don't have to have a budget of at least, let's say, I don't know, two to five k to, to fly over the pond and, and attend this thing. You can just do it from your home. So that is, that's a lot more inclusive. And I expect this to, to basically be a second more or less orthogonal growth, growth path. But the best thing about coupon is the hallway track. I'm just meeting people, talking to people and that kind of thing is not really possible with, >>It's, it's great to see people >>In person. No, and it makes such a difference. I mean, yeah. Even and interviewing people in person too. I mean, it does a, it's, it's, and, and this, this whole, I mean cncf, this whole community, every company here is community first. It's how these projects come to be. I think it's awesome. I feel like you got something you're saying to say, Johnny. >>Yeah. And I love some of the advancements. Rich Richie, we talked last time about, you know, open telemetry, open metrics. You're involved in dashboards. Yeah. One of the themes here is ease of use, simplicity, developer productivity. Where do you see the ease of use going from a project standpoint? For me, as you mentions everywhere, it's pretty much, it is, it's almost all corners of the world. Yep. And new people coming in. How, how are you making it easier? What's going on? Give us the update on that. >>So we also, funnily enough at precisely this topic in the TC panel just a few hours ago, about ease of use and about how to, how to make things easier to, to handle how developers currently, like if they just want to get into the cloud native seen, they have like, like we, we did some neck and math, like maybe 10 tools at least, which you have to be somewhat proficient in to just get started, which is honestly horrendous. Yeah. Course. Like with a server, I just had my survey install my thing and it runs, maybe I need a database, but that's roughly it. And this needs to change again. Like it's, it's nice that everything is, is un unraveled. And you have, you, you, you, you don't have those service boundaries which you had before. You can do all the horizontal scaling, you can do all the automatic scaling, all those things that they're super nice. But at the same time, this complexity, which used to be nicely compartmentalized, was deliberately broken up. And so it's becoming a lot harder to, to, like, we, we need to find new ways to compartmentalize this complexity back to, to human understandable levels again, in particular, as we keep onboarding new and new and new, new people, of course it's just not good use of anyone's time to, to just like learn the basics again and again and again. This is something which should be just compartmentalized and automated away. We're >>The three, We were talking to Matt Klein earlier and he was talking about as projects become mature and all over the place and have reach and and usage, you gotta work on the boring stuff. Yes. And when it's boring, that means you have success. Yes. But then you gotta work on the plumbing. What are some of the things that you guys are working on? Because people are relying on the product. >>Oh yeah. So for with my premises head on, the highlight feature is exponential or native or spars. Histograms. There's like three different names for one single concept. If you know Prometheus, you ha you currently have hard bucket boundaries where I say my latency is lower equal two seconds, one second, a hundred milliseconds, what have you. And I can put stuff into those histogram buckets accordingly to those predefined levels, which is extremely efficient, but like on the, on the code level. But it's not very nice for the humans course you need to understand your system before you're able to, to, to choose good cutoff points. And if you, if you, if you add new ones, that's completely fine. But if you want to actually change them, course you, you figured out that you made a fundamental mistake, you're going to have a break in the continue continuity of your observability data. And you cannot undo this in, into the past. So this is just gone native histograms. On the other hand, allow me to, to, okay, I'm not going to get get into the math, but basically you define a single formula, which there comes a good default. If you have good reasons, then you can change it. But if you don't, just don't talk, >>The people are in the math, Hit him up on Twitter. Twitter, h you'll get you that math. >>So the, >>The thing is people want the math, believe me. >>Oh >>Yeah. I mean we don't have time, but hit him up. Yeah. >>There's ProCon in two weeks in Munich and there will be whole talk about like the, the dirty details of all of the stuff. But the, the high level answer is it just does what people would expect it to do. And with very little overhead, you become, you get highly, highly or high resolution histograms, which is really important for a lot of use cases. But this is not just Prometheus with my open metrics head on the 2.0 feature, like the breaking highlight feature of Open Metrics 2.0 will be you guested precisely the same with my open telemetry head on. Low and behold the same underlying technology is being put or has been put into open telemetry. And we've worked for month and month and month and even longer between all different projects to, to assert that we have one single standard which is actually compatible with each other course. One of the worst things which you can have in the cloud ecosystem is if you have soly different things and they break in subtly wrong ways, like it's much better to just not work than to break in a way, which is just a little bit wrong. Of course you won't figure this out until it's too late. So we spent, like with all three hats, we spent insane amounts of time on making this happen and, and making this nice. >>Savannah, one of the things we have so much going on at Cube Con. I mean just you're unpacking like probably another day of cube. We can't go four days, but open time. >>I know, I know. I'm the same >>Open telemetry >>Challenge acceptance open. >>Sorry, we're gonna stay here. All the, They >>Shut the lights off on us last night. >>They literally gonna pull the plug on us. Yeah, yeah, yeah, yeah. They've done that before. It's not the first time we go until they kick us out. We love, love doing this. But Open telemetry is got a lot of news too. So that's, We haven't really talked much about that. >>We haven't at >>All. So there's a lot of stuff going on that, I won't call it boring. That's like code word's. That's cube talk for, for it's working. Yeah. So it's not bad, but there's a lot of stuff going on. Like open telemetry, open metrics, This is the stuff that matters cuz when you go in large scale, that's key. It's just what, missing all the, all the stuff. >>No, >>What are we missing? What are people missing? What's going on in the show that you think that's not actually being reported on? I mean it's a lot of high web assembly for instance got a lot >>Of high. Oh yeah, I was gonna say, I'm glad you're asking this because you, you've already mentioned about seven different hats that you wear. I can only imagine how many hats are actually in your hat cabinet. But you, you are someone with your, with your fingers in a lot of different things. So you can kind of give us a state of the union. Yeah. So go ahead. Let's talk about >>It. So I think you already hit a few good points. Ease of use is definitely one of them. And, and improving the developer experience and not having this like a value of pain. Yeah. That is one of the really big ones. It's going to be interesting cause it is boring. It is janitorial and it needs a different type of persona. A lot of, or maybe not most, but a large fraction of developers like the shiny stuff. And we could see this in Prometheus where like initially the people who contributed this the most where like those restless people who need to fix that one thing, this is impossible, are going to do it. Which changed over the years where the people who now contribute the most are off the janitorial. Like keep things boring, keep things running, still have substantial changes. But but not like more on the maintenance level. >>Yeah. The maintainers. I was just gonna bring that >>Up. Yeah. On the, on the keep things boring while still pushing 'em forward. Yeah. And the thing about ease of use is a lot of this is boring. A lot of this is strategy. A lot of this is toil. A lot of this takes lots of research also in areas where developers are not really good at, like UX for example, and ui like most software developers are really bad at those cause they just think differently from normal humans, I guess. >>So that's an interesting observation that you just made. I we could unpack that on a whole nother show as well. >>So the, the thing is this is going to be interesting for the open source scene course. This needs deliberate investment by companies who assign people to those projects and say, okay, fix that one thing or make it easier to use what have you. That is a lot easier with, with first party products and projects from companies cuz they can invest directly into the thing and they see much more of a value prop. It's, it's kind of normal by now to, to allow developers or even assigned developers onto open source projects. That's not so much the case for the tpms, for the architects, for the UX and your I people like for the documentation people that there's not as much awareness of that this is also driving value for everyone. Yes. And also there's not much as much. >>Yeah, that's a great point. This whole workflow production system of open source, which has grown and keeps growing and we'll keep growing. These be funded. And one of the things we were talking earlier in another session about is about the recession potentially we're hitting and the global issues, macroeconomics that might force some of these projects or companies not to get VC >>Funding. It's such a theme at the show. So, >>So to me, I said it's just not about VC funding. There's other funding mechanisms that's community oriented. There's companies participating, there's other meccas. Richie, if you could have your wishlist of how things could progress an open source, what would you want to see happen in terms of how it's, how things are funded, how things are executed. Cuz developers are going to run businesses. Cuz ultimately if you follow digital transformation to completion, it and developers aren't a department serving the business. They are the business. And that's coming fast. You know, what has to happen in your opinion, if you had the wish magic wand, what would you, what would you snap your fingers to make happen? >>If I had a magic wand that's very different from, from what is achievable. But let, let's >>Go with, Okay, go with the magic wand first. Cause we'll, we'll, we'll we'll riff on that. So >>I'm here for dreams. Yeah, yeah, >>Yeah. I mean I, I've been in open source for more than two, two decades, but now, and most of the open source is being driven forward by people who are not being paid for those. So for example, Gana is the first time I'm actually paid by a company to do my com community work. It's always been on the side. Of course I believe in it and I like doing it. I'm also not bad at it. And so I just kept doing it. But it was like at night on the weekends and everything. And to be honest, it's still at night and in the weekends, but the majority of it is during paid company time, which is awesome. Yeah. Most of the people who have driven this space forward are not in this position. They're doing it at night, they're doing it on the weekends. They're doing it out of dedication to a cause. Yeah. >>The commitment is insane. >>Yeah. At the same time you have companies mostly hyperscalers and either they have really big cloud offerings or they have really big advertisement business or both. And they're extracting a huge amount of value, which has been created in large part elsewhere. Like yes, they employ a ton of developers, but a lot of the technologies they built on and the shoulders of the giants they stand upon it are really poorly paid. And there are some efforts to like, I think the core foundation like which redistribute a little bit of money and such. But if I had my magic wand, everyone who is an open source and actually drives things forwards, get, I don't know, 20% of the value which they create just magically somehow. Yeah. >>Or, or other companies don't extract as much value and, and redistribute more like put more full-time engineers onto projects or whichever, like that would be the ideal state where the people who actually make the thing out of dedication are not more or less left on the sideline. Of course they're too dedicated to just say, Okay, I'm, I'm not doing this anymore. You figure this stuff out and let things tremble and falter. So I mean, it's like with nurses and such who, who just like, they, they know they have something which is important and they keep doing it. Of course they believe in it. >>I think this, I think this is an opportunity to start messaging this narrative because yeah, absolutely. Now we're at an inflection point where there's a big community, there is a shared responsibility in my opinion, to not spread the wealth, but make sure that it's equally balanced and, and the, and I think there's a way to do that. I don't know how yet, but I see that more than ever, it's not just come in, raid the kingdom, steal all the jewels, monetize it, and throw some token token money around. >>Well, in the burnout. Yeah, I mean I, the other thing that I'm thinking about too is it's, you know, it's, it's the, it's the financial aspect of this. It's the cognitive load. And I'm curious actually, when I ask you this question, how do you avoid burnout? You do a million different things and we're, you know, I'm sure the open source community that passion the >>Coach. Yeah. So it's just write code, >>It's, oh, my, my, my software engineering days are firmly over. I'm, I'm, I'm like, I'm the cat herer and the janitor and like this type of thing. I, I don't really write code anymore. >>It's how do you avoid burnout? >>So a i I didn't curse ahead burnout a few years ago. I was not nice, but that was still when I had like a full day job and that day job was super intense and on top I did all the things. Part of being honest, a lot of the people who do this are really dedicated and are really bad at setting boundaries between work >>And process. That's why I bring it up. Yeah. Literally why I bring it up. Yeah. >>I I I'm firmly in that area and I'm, I'm, I don't claim I have this fully figured out yet. It's also even more risky to some extent per like, it's, it's good if you're paid for this and you can do it during your work time. But on the other hand, if it's so nice and like if your hobby and your job are almost completely intersectional, it >>Becomes really, the lines are blurry. >>Yeah. And then yeah, like have work from home. You, you don't even commute anything or anymore. You just sit down at your computer and you just have fun doing your stuff and all of a sudden it's deep at night and you're still like, I want to keep going. >>Sounds like God, something cute. I >>Know. I was gonna say, I was like, passion is something we all have in common here on this. >>That's the key. That is the key point There is a, the, the passion project becomes the job. But now the contribution is interesting because now yeah, this ecosystem is, is has a commercial aspect. Again, this is the, this is the balance between commercialization and keeping that organic production system that's called open source. I mean, it's so fascinating and this is amazing. I want to continue that conversation. It's >>Awesome. Yeah. Yeah. This is, this is great. Richard, this entire conversation has been excellent. Thank you so much for joining us. How can people find you? I mean, I give em your Twitter handle, but if they wanna find out more about Grafana Prometheus and the 1700 things you do >>For grafana grafana.com, for Prometheus, promeus.io for my own stuff, GitHub slash richie age slash talks. Of course I track all my talks in there and like, I don't, I currently don't have a personal website cause I stop bothering, but my, like that repository is, is very, you find what I do over, like for example, the recording link will be uploaded to this GitHub. >>Yeah. Great. Follow. You also run a lot of events and a lot of community activity. Congratulations for you. Also, I talked about this last time, the largest IRC network on earth. You ran, built a data center from scratch. What happened? You done >>That? >>Haven't done a, he even built a cloud hyperscale compete with Amazon. That's the next one. Why don't you put that on the >>Plate? We'll be sure to feature whatever Richie does next year on the cube. >>I'm game. Yeah. >>Fantastic. On that note, Richie, again, thank you so much for being here, John, always a pleasure. Thank you. And thank you for tuning in to us here live from Detroit, Michigan on the cube. My name is Savannah Peterson and here's to hoping that you find balance in your life this weekend.
SUMMARY :
We've done over 30, but this conversation is gonna be extra special, don't you think, We're getting the conversation of what's going on in the industry management, Richie, thank you so much for joining us. I mean, I, I, I run forme day. You, you have your hands in a lot. You have to write dashboards, you have to write alerts, you have to write everything to just get started with Like 60% of the people here are first time attendees. And I love that you, you had those numbers. So I mean, I covid changed a few things. I mean, you know, casually, it's like such a gentle way of putting that, I love it, I expect this to take up again. Some of the momentum, I mean, from the project level, Most of this is online anyway. So the projects are arguably spearheading this, I feel like you got something you're saying to say, Johnny. it's almost all corners of the world. You can do all the horizontal scaling, you can do all the automatic scaling, all those things that they're super nice. What are some of the things that you But it's not very nice for the humans course you need The people are in the math, Hit him up on Twitter. Yeah. One of the worst things which you can have in the cloud ecosystem is if you have soly different things and Savannah, one of the things we have so much going on at Cube Con. I'm the same All the, They It's not the first time we go until they Like open telemetry, open metrics, This is the stuff that matters cuz when you go in large scale, So you can kind of give us a state of the union. And, and improving the developer experience and not having this like a I was just gonna bring that the thing about ease of use is a lot of this is boring. So that's an interesting observation that you just made. So the, the thing is this is going to be interesting for the open source scene course. And one of the things we were talking earlier in So, Richie, if you could have your wishlist of how things could But let, let's So Yeah, yeah, Gana is the first time I'm actually paid by a company to do my com community work. shoulders of the giants they stand upon it are really poorly paid. are not more or less left on the sideline. I think this, I think this is an opportunity to start messaging this narrative because yeah, Yeah, I mean I, the other thing that I'm thinking about too is it's, you know, I'm, I'm like, I'm the cat herer and the janitor and like this type of thing. a lot of the people who do this are really dedicated and are really Yeah. I I I'm firmly in that area and I'm, I'm, I don't claim I have this fully You, you don't even commute anything or anymore. I That is the key point There is a, the, the passion project becomes the job. things you do like that repository is, is very, you find what I do over, like for example, the recording link will be uploaded Also, I talked about this last time, the largest IRC network on earth. That's the next one. We'll be sure to feature whatever Richie does next year on the cube. Yeah. My name is Savannah Peterson and here's to hoping that you find balance in your life this weekend.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Richie Hartman | PERSON | 0.99+ |
Richie | PERSON | 0.99+ |
Matt Klein | PERSON | 0.99+ |
Savannah Peterson | PERSON | 0.99+ |
Richard Hartmann | PERSON | 0.99+ |
Richard | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
John | PERSON | 0.99+ |
Grafana Labs | ORGANIZATION | 0.99+ |
Prometheus | TITLE | 0.99+ |
Rich Richie | PERSON | 0.99+ |
60% | QUANTITY | 0.99+ |
Griffon Labs | ORGANIZATION | 0.99+ |
two seconds | QUANTITY | 0.99+ |
one second | QUANTITY | 0.99+ |
Munich | LOCATION | 0.99+ |
20% | QUANTITY | 0.99+ |
10 tools | QUANTITY | 0.99+ |
Detroit | LOCATION | 0.99+ |
Monday | DATE | 0.99+ |
Detroit, Michigan | LOCATION | 0.99+ |
Grafana | ORGANIZATION | 0.99+ |
yesterday | DATE | 0.99+ |
Grafana Prometheus | TITLE | 0.99+ |
three | QUANTITY | 0.99+ |
five k | QUANTITY | 0.99+ |
first time | QUANTITY | 0.99+ |
two | QUANTITY | 0.98+ |
next year | DATE | 0.98+ |
both | QUANTITY | 0.98+ |
one | QUANTITY | 0.98+ |
this week | DATE | 0.98+ |
two decades | QUANTITY | 0.98+ |
one single concept | QUANTITY | 0.98+ |
GitHub | ORGANIZATION | 0.98+ |
2019 | DATE | 0.98+ |
Grafana cloud | TITLE | 0.98+ |
One | QUANTITY | 0.97+ |
last night | DATE | 0.97+ |
Savannah | PERSON | 0.97+ |
ORGANIZATION | 0.96+ | |
earth | LOCATION | 0.96+ |
four days | QUANTITY | 0.96+ |
over 30 | QUANTITY | 0.95+ |
Johnny | PERSON | 0.95+ |
one click | QUANTITY | 0.95+ |
Grafana Cloud | TITLE | 0.95+ |
CloudNativeCon | EVENT | 0.94+ |
few hours ago | DATE | 0.93+ |
2.0 | OTHER | 0.93+ |
Griffon | ORGANIZATION | 0.93+ |
hundred percent | QUANTITY | 0.92+ |
two weeks | QUANTITY | 0.92+ |
one thing | QUANTITY | 0.91+ |
grafana grafana.com | OTHER | 0.9+ |
more than two | QUANTITY | 0.89+ |
three different names | QUANTITY | 0.88+ |
two largest | QUANTITY | 0.88+ |
promeus.io | OTHER | 0.86+ |
a hundred milliseconds | QUANTITY | 0.86+ |
few years ago | DATE | 0.86+ |
single formula | QUANTITY | 0.85+ |
first | QUANTITY | 0.83+ |
Con. | EVENT | 0.83+ |
IRC | ORGANIZATION | 0.82+ |
Kubernetes | TITLE | 0.81+ |
seven different hats | QUANTITY | 0.8+ |
one single standard | QUANTITY | 0.79+ |
Valencia Andro | ORGANIZATION | 0.79+ |
NA 2022 | EVENT | 0.77+ |
Open Metrics 2.0 | OTHER | 0.74+ |
KubeCon + | EVENT | 0.7+ |
Ian Smith, Chronosphere | KubeCon + CloudNativeCon NA 2022`
(upbeat music) >> Good Friday morning everyone from Motor City, Lisa Martin here with John Furrier. This is our third day, theCUBE's third day of coverage of KubeCon + CloudNativeCon 22' North America. John, we've had some amazing conversations the last three days. We've had some good conversations about observability. We're going to take that one step further and look beyond its three pillars. >> Yeah, this is going to be a great segment. Looking forward to this. This is about in depth conversation on observability. The guest is technical and it's on the front lines with customers. Looking forward to this segment. Should be great. >> Yeah. Ian Smith is here, the field CTO at Chronosphere. Ian, welcome to theCUBE. Great to have you. >> Thank you so much. It's great to be here. >> All right. Talk about the traditional three pillars, approach, and observability. What are some of the challenges with that, and how does Chronosphere solve those? >> Sure. So hopefully everyone knows people think of the three pillars as logs, metrics and traces. What do you do with that? There's no action there. It's just data, right? You collect this data, you go put it somewhere, but it's not actually talking about any sort of outcomes. And I think that's really the heart of the issue, is you're not achieving anything. You're just collecting a whole bunch of data. Where do you put it? What are you... What can you do with it? Those are the fundamental questions. And so one of the things that we're focused on at Chronosphere is, well, what are those outcomes? What is the real value of that? And for example, thinking about phases of observability. When you have an incident or you're trying to investigate something through observability, you probably want to know what's going on. You want to triage any problems you detect. And then finally, you want to understand the cause of those and be able to take longer term steps to address them. >> What do customers do when they start thinking about it? Because observability has that promise. Hey, you know, get the data, we'll throw AI at it. >> Ian: Yeah. >> And that'll solve the problem. When they get over their skis, when do they realize that they're really not tackling it properly, or the ones that are taking the right approach? What's the revelation? What's your take on that? You're in the front lines. What's going on with the customer? The good and the bad. What's the scene look like? >> Yeah, so I think the bad is, you know, you end up buying a lot of things or implementing even in open source or self building, and it's very disconnected. You're not... You don't have a workflow, you don't have a path to success. If you ask different teams, like how do you address these particular problems? They're going to give you a bunch of different answers. And then if you ask about what their success rate is, it's probably very uneven. Another key indicator of problems is that, well, do you always need particular senior engineers in your instance or to help answer particular performance problems? And it's a massive anti pattern, right? You have your senior engineers who are probably need to be focused on innovation and competitive differentiation, but then they become the bottleneck. And you have this massive sort of wedge of maybe less experienced engineers, but no less valuable in the overall company perspective, who aren't effective at being able to address these problems because the tooling isn't right, the workflows are incorrect. >> So the senior engineers are getting pulled in to kind of fix and troubleshoot or observe what the observability data did or didn't deliver. >> Correct. Yeah. And you know, the promise of observability, a lot of people talk about unknown unknowns and there's a lot of, you know, crafting complex queries and all this other things. It's a very romantic sort of deep dive approach. But realistically, you need to make it very accessible. If you're relying on complex query languages and the required knowledge about the architecture and everything every other team is doing, that knowledge is going to be super concentrated in just a couple of heads. And those heads shouldn't be woken up every time at 3:00 AM. They shouldn't be on every instant call. But oftentimes they are the sort of linchpin to addressing, oh, as a business we need to be up 99.99% of the time. So how do we accomplish that? Well, we're going to end up burning those people. >> Lisa: Yeah. >> But also it leads to a great dissatisfaction in the bulk of the engineers who are, you know, just trying to build and operate the services. >> So talk... You mentioned that some of the problems with the traditional three pillars are, it's not outcome based, it leads to silo approaches. What is Chronosphere's definition and can you walk us through those three phases and how that really gives you that competitive edge in the market? >> Yeah, so the three phases being know, triage and understand. So just knowing about a problem, and you can relate this very specifically to capabilities, but it's not capabilities first, not feature function first. So know, I need to be able to alert on things. So I do need to collect data that gives me those signals. But particularly as you know, the industry starts moving towards as slows. You start getting more business relevant data. Everyone knows about alert storms. And as you mentioned, you know, there's this great white hope of AI and machine learning, but AI machine learning is putting a trust in sort of a black box, or the more likely reality is that really statistical model. And you have to go and spend a very significant amount time programming it for sort of not great outcomes. So know, okay, I want to know that I have a problem, I want to maybe understand the symptoms of that particular problem. And then triage, okay, maybe I have a lot of things going wrong at the same time, but I need to be very precise about my resources. I need to be able to understand the scope and importance. Maybe I have five major SLOs being violated right now. Which one is the greatest business impact? Which symptoms are impacting my most valuable customers? And then from there, not getting into the situation, which is very common where, okay, well we have every... Your customer facing engineering team, they have to be on the call. So we have 15 customer facing web services. They all have to be on that call. Triage is that really important aspect of really mitigating the cost to the organization because everyone goes, oh, well I achieved my MTTR and my experience from a variety of vendors is that most organizations, unless you're essentially failing as a business, you achieve your SLA, you know, three nines, four nines, whatever it is. But the cost of doing that becomes incredibly extreme. >> This is huge point. I want to dig into that if you don't mind, 'cause you know, we've been all seeing the cost of ownership miles in it all, the cost of doing business, cost of the shark fan, the iceberg, what's under the water, all those metaphors. >> Ian: Yeah. >> When you look at what you're talking about here, there are actually, actually real hardcore costs that might be under the water, so to speak, like labor, senior engineering time, 'cause Cloud Native engineers are coding in the pipelines. A lot of impact. Can you quantify and just share an example or illustrate where the costs are? 'Cause this is something that's kind of not obvious. >> Ian: Yeah. >> On the hard costs. It's not like a dollar amount, but time resource breach, wrong triage, gap in the data. What are some of the costs? >> Yeah, and I think they're actually far more important than the hard costs of infrastructure and licensing. And of course there are many organizations out there using open source observability components together. And they go, Oh it's free. No licensing costs. But you think again about those outcomes. Okay, I have these 15 teams and okay, I have X number of incidents a month, if I pull a representative from every single one of those teams on. And it turns out that, you know, as we get down in further phases, we need to be able to understand and remediate the issue. But actually only two teams required of that. There's 13 individuals who do not need to be on the call. Okay, yes, I met my SLA and MTTR, but if I am from a competitive standpoint, I'm comparing myself to a very similar organization that only need to impact those two engineers versus the 15 that I had over here. Who is going to be the most competitive? Who's going to be most differentiated? And it's not just in terms of number of lines of code, but leading to burnout of your engineers and the churn of that VPs of engineering, particularly in today's economy, the hardest thing to do is acquire engineers and retain them. So why do you want to burn them unnecessarily on when you can say, okay, well I can achieve the same or better result if I think more clearly about my observability, but reduce the number of people involved, reduce the number of, you know, senior engineers involved, and ultimately have those resources more focused on innovation. >> You know, one thing I want, at least want get in there, but one thing that's come up a lot this year, more than I've ever seen before, we've heard about the skill gaps, obviously, but burnout is huge. >> Ian: Yes. >> That's coming up more and more. This is a real... This actually doesn't help the skills gap either. >> Ian: Correct. >> Because you got skills gap, that's a cost potentially. >> Ian: Yeah. >> And then you got burnout. >> Ian: Yeah. >> People just kind of sitting on their hands or just walking away. >> Yeah. So one of the things that we're doing with Chronosphere is, you know, while we do deal with the, you know, the pillar data, but we're thinking about it more, what can you achieve with that? Right? So, and aligning with the know, triage and understand. And so you think about things like alerts, you know, dashboards, you be able to start triaging your symptoms. But really importantly, how do we bring the capabilities of things like distributed tracing where they can actually impact this? And it's not just in the context of, well, what can we do in this one incident? So there may be scenarios where you, absolutely do need those power users or those really sophisticated engineers. But from a product challenge perspective, what I'm personally really excited about is how do you capture that insight and those capabilities and then feed that back in from a product perspective so it's accessible. So you know, everyone talks about unknown unknowns in observability and then everyone sort of is a little dismissive of monitoring, but monitoring that thing, that democratizes access and the decision making capacity. So if you say I once worked at an organization and there were three engineers in the whole company who could generate the list of customers who were impacted by a particular incident. And I was in post sales at the time. So anytime there was a major incident, need to go generate that list. Those three engineers were on every single incident until one of them got frustrated and built a tool. But he built it entirely on his own. But can you think from an observability perspective, can you build a thing that it makes all those kinds of capabilities accessible to the first point where you take that alert, you know, which customers are affected or whatever other context was useful last time, but took an hour, two hours to achieve. And so that's what really makes a dramatic difference over time, is it's not about the day one experience, but how does the product evolve with the requirements and the workflow- >> And Cloud Native engineers, they're coding so they can actually be reactive. That's interesting, a platform and a tool. >> Ian: Yes. >> And platform engineering is the hottest topic at this event. And this year, I would say with Cloud Native hearing a lot more. I mean, I think that comes from the fact that SREs not really SRE, I think it's more a platform engineer. >> Ian: Yes. >> Not everyone's an... Not company has an SRE or SRE environment. But platform engineering is becoming that new layer that enables the developers. >> Ian: Correct. >> This is what you're talking about. >> Yeah. And there's lots of different labels for it, but I think organizations that really think about it well they're thinking about things like those teams, that developer efficiency, developer productivity. Because again, it's about the outcomes. It's not, oh, we just need to keep the site reliable. Yes, you can do that, but as we talked about, there are many different ways that you can burn unnecessary resources. But if you focus on developer efficiency and productivity, there's retainment, there's that competitive differentiation. >> Let's uplevel those business outcomes. Obviously you talked about in three phases, know, triage and understand. You've got great alignment with the Cloud Native engineers, the end users. Imagine that you're facilitating company's ability to reduce churn, attract more talent, retain talent. But what are some of the business outcomes? Like to the customer experience to the brand? >> Ian: Sure. >> Talk about it in some of those contexts. >> Yeah. One of the things that not a lot of organizations think about is, what is the reliability of my observability solution? It's like, well, that's not what I'm focused on. I'm focused on the reliability of my own website. Okay, let's take the, common open source pattern. I'm going to deploy my observability solution next to my core site infrastructure. Okay, I now have a platform problem because DNS stopped working in cloud provider of my choice. It's also affecting my observability solution. So at the moment that I need- >> And the tool chain and everything else. >> Yeah. At the moment that I need it the most to understand what's going on and to be able to know triage and understand that fails me at the same time. It's like, so reliability has this very big impact. So being able to make sure that my solution's reliable so that when I need it the most, and I can affect reliability of my own solution, my own SLA. That's a really key aspect of it. One of the things though that we, look at is it's not just about the outcomes and the value, it's ROI, right? It's what are you investing to put into that? So we've talked a little bit about the engineering cost, there's the infrastructure cost, but there's also a massive data explosion, particularly with Cloud Native. >> Yes. Give us... Alright, put that into real world examples. A customer that you think really articulates the value of what Chronosphere is delivering and why you're different in the market. >> Yeah, so DoorDash is a great customer example. They're here at KubeCon talking about their experience with Chronosphere and you know, the Cloud Native technologies, Prometheus and those other components align with Chronosphere. But being able to undergo, you know, a transformation, they're a Cloud Native organization, but going a transformation from StatsD to very heavy microservices, very heavy Kubernetes and orchestration. And doing that with your massive explosion, particularly during the last couple of years, obviously that's had a very positive impact on their business. But being able to do that in a cost effective way, right? One of the dirty little secrets about observability in particular is your business growth might be, let's say 50%, 60%, your infrastructure spend in the cloud providers is maybe going to be another 10, 15% on top of that. But then you have the intersection of, well my engineers need more data to diagnose things. The business needs more data to understand what's going on. Plus we've had this massive explosion of containers and everything like that. So oftentimes your business growth is going to be more than doubled with your observability data growth and SaaS solutions and even your on-premises solutions. What's the main cost driver? It's the volume of data that you're processing and storing. And so Chronosphere one of the key things that we do, because we're focused on organizational pain for larger scale organizations, is well, how do we extract the maximum volume of the data you're generating without having to store all of that data and then present it not just from a cost perspective, but also from a performance perspective. >> Yes. >> John: Yeah. >> And so feeding all into developer productivity and also lowering that investment so that your return can stand out more clearly and more valuably when you are assessing that TCO. >> Better insights and outcomes drives developer productivity for sure. That also has top theme here at KubeCon this year. It always is, but this is more than ever 'cause of the velocity. My question for you, given that you're the field chief technology officer for Chronosphere and you have a unique position, you've got a great experience in the industry, been involved in some really big companies and cutting edge. What's the competitive landscape? 'Cause the customers sometimes are confused by all the pitches they're getting from other vendors. Some are bolting on observability. Some have created like I would say, a shim layer or horizontally scalable platform or platform engineering approach. It's a data problem. Okay. This is a data architecture challenge. You mentioned that many times. What's the difference between a pretender and a player in this space? What's the winning architecture look like? What's a, I won't say phony or fake solution, but ones that customers should be aware of? Because my opinion, if you have a gap in the data or you configure it wrong, like a bolt on and say DNS crashes you're dead in the water. >> Ian: Yeah. >> What's the right approach from a customer standpoint? How do they squint through all the noise to figure out what's the right approach? >> Yeah, so I mean, I think one of the ways, and I've worked with customers in a pre-sales capacity for a very long time I know all the tricks of guiding you through. I think it needs to be very clear that customers should not be guided by the vendor. You don't talk to one vendor and they decide, Oh, I'm going to evaluate based off this. We need to particularly get away from feature based evaluations. Features are very important, but they're all have to be aligned around outcomes. And then you have to clearly understand, where am I today? What do I do today? And what is going to be the transformation that I have to go through to take advantage of these features? They can get very entrancing to say, Oh, there's a list of 25 features that this solution has that no one else has, but how am I going to get value out of that? >> I mean, distributed tracing is a distributed word. Distributed is the key word. This is a system architecture. The holistic big picture comes in. How do they figure that out? Knowing what they're transforming into? How does it fit in? >> Ian: Yeah. >> What's the right approach? >> Too often I say distributed tracing, particularly, you know, bought, because again, look at the shiny features look at the the premise and the MTTR expectations, all these other things. And then it's off to the side. We go through the traditional usage of metrics very often, very log heavy approaches, maybe even some legacy APM. And then it's sort of at last resort. And out of all the tools, I think distributed tracing is the worst in the problem we talked about earlier where the most sophisticated engineers, the ones who are being longest tenured, are the only ones who end up using it. So adoption is really, really poor. So again, what do we do today? Well, we alert, we probably want to understand our symptoms, but then what is the key problem? Oh, we spend a lot of time digging into the where the problem exists in my architecture, we talked about, you know, getting every engineer in at the same time, but how do we reduce the number of engineers involved? How do we make it so that, well, this looks like a great day one experience, but what is my day 30 experience like? Day 90. How is the product get more valuable? How do I get my most senior engineers out of this, not just on day one, but as we progress through it? >> You got to operationalize it. That's the key. >> Yeah, Correct. >> Summarize this as we wrap here. When you're in customer conversations, what is the key factor behind Chronosphere's success? If you can boil it down to that key nugget, what is it? >> I think the key nugget is that we're not just fixated on sort of like technical features and functions and frankly gimmicks of like, Oh, what could you possibly do with these three pillars of data? It's more about what can we do to solve organizational pain at the high level? You know, things like what is the cost of these solutions? But then also on the individual level, it's like, what exactly is an engineer trying to do? And how is their quality of life affected by this kind of tooling? And it's something I'm very passionate about. >> Sounds like it. Well, the quality of life's important, right? For everybody, for the business, and ultimately ends up affecting the overall customer experience. So great job, Ian, thank you so much for joining John and me talking about what you guys are doing beyond the three pillars of observability at Chronosphere. We appreciate your insights. >> Thank you so much. >> John: All right. >> All right. For John Furrier and our guest, I'm Lisa Martin. You're watching theCUBE live Friday morning from KubeCon + CloudNativeCon 22' from Detroit. Our next guest joins theCUBE momentarily, so stick around. (upbeat music)
SUMMARY :
the last three days. it's on the front lines Ian Smith is here, the It's great to be here. What are some of the challenges with that, the cause of those and be able to take Hey, you know, get the And that'll solve the problem. They're going to give you a So the senior engineers and the required knowledge in the bulk of the and how that really gives you the cost to the organization cost of the shark fan, are coding in the pipelines. What are some of the costs? reduce the number of, you know, but burnout is huge. the skills gap either. Because you got skills gap, People just kind of And it's not just in the context of, And Cloud Native engineers, is the hottest topic that enables the developers. Because again, it's about the outcomes. the Cloud Native engineers, Talk about it in One of the things that not the most to understand what's the value of what One of the dirty little when you are assessing that TCO. 'cause of the velocity. And then you have to clearly understand, Distributed is the key word. And out of all the tools, That's the key. If you can boil it down the cost of these solutions? beyond the three pillars For John Furrier and our
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Ian | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
John | PERSON | 0.99+ |
Ian Smith | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
two hours | QUANTITY | 0.99+ |
15 teams | QUANTITY | 0.99+ |
John Furrier | PERSON | 0.99+ |
13 individuals | QUANTITY | 0.99+ |
25 features | QUANTITY | 0.99+ |
50% | QUANTITY | 0.99+ |
three engineers | QUANTITY | 0.99+ |
three engineers | QUANTITY | 0.99+ |
60% | QUANTITY | 0.99+ |
two teams | QUANTITY | 0.99+ |
an hour | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
third day | QUANTITY | 0.99+ |
15 | QUANTITY | 0.99+ |
10, 15% | QUANTITY | 0.99+ |
Detroit | LOCATION | 0.99+ |
two engineers | QUANTITY | 0.99+ |
3:00 AM | DATE | 0.99+ |
KubeCon | EVENT | 0.99+ |
15 customer | QUANTITY | 0.99+ |
Friday morning | DATE | 0.99+ |
first point | QUANTITY | 0.99+ |
KubeCon | ORGANIZATION | 0.99+ |
Cloud Native | ORGANIZATION | 0.99+ |
three phases | QUANTITY | 0.98+ |
one | QUANTITY | 0.98+ |
three pillars | QUANTITY | 0.98+ |
DoorDash | ORGANIZATION | 0.98+ |
One | QUANTITY | 0.97+ |
this year | DATE | 0.97+ |
theCUBE | ORGANIZATION | 0.96+ |
three nines | QUANTITY | 0.95+ |
three pillars | QUANTITY | 0.94+ |
day one | QUANTITY | 0.94+ |
one step | QUANTITY | 0.93+ |
Chronosphere | TITLE | 0.92+ |
one incident | QUANTITY | 0.92+ |
North America | LOCATION | 0.92+ |
CloudNativeCon | EVENT | 0.91+ |
Prometheus | TITLE | 0.91+ |
99.99% | QUANTITY | 0.9+ |
first | QUANTITY | 0.89+ |
one thing | QUANTITY | 0.89+ |
four nines | QUANTITY | 0.86+ |
last couple of years | DATE | 0.85+ |
one vendor | QUANTITY | 0.85+ |
Chronosphere | ORGANIZATION | 0.84+ |
Day 90 | QUANTITY | 0.84+ |
Cloud Native | TITLE | 0.83+ |
Martin Mao & Jeff Cobb, Chronosphere | KubeCon + CloudNativeCon NA 2022
>>Good afternoon everyone, and welcome back to Cuan where my cohost John Farer and I are broadcasting live, along with Lisa Martin from Cuan Detroit, Michigan. We are joined this afternoon by two very interesting gentlemen who also happen to be legends on the cube. John, how long have you known the next few? They've, >>They've made their mark on the cube with Jerry Chen from Greylock was one of our most attended cube guests. He's a VC partner at Greylock and an investor and this company that just launched their new cloud observability platform should be a great segment. >>Well, I'm excited. I are. Are you excited? Should I string this out just a little bit longer? No, I won't. I won't do that to you. Please welcome Martin and Jeff from Chronosphere Martin. Jeff, thank you so much for being >>Here. Thank you for having us. Thank you. >>I noticed right away that you have raised a mammoth series C. Yeah. 200 million if I'm not mistaken. >>That is correct. >>Where's the company at? >>Yeah, so we raised that series C a year ago. In fact, we were just talking about it a year ago at Cub Con. Since then, at the time we're about 80 employees or so. Since then, we've tripled the headcount, so we're over 200 people. Casual, triple casual, triple of the headcount. Yeah. Luckily it was the support of business, which is also tripled in the last year. So we're very lucky from that perspective as well. And a couple of other things we're pretty proud of last year. We've had a hundred percent customer retention, which is always a great thing to have as a SaaS platform there. >>Real metric if you've had a hundred percent. I'm >>Kidding. It's a good metric to, to put out there if you had a hundred percent. I would say for sure. It's an A for sure and exactly welcome to meet >>Anyone else who's had a hundred percent >>Customer attention here at coupon this week and 90% of our customers are using more of the service and, and you know, therefore paying more for the service as well. So those are great science for us and I think it shows that we're clearly doing something right on the product side. I would say. And >>Last and last time you're on the cube. We're talking about about the right data. Not so much a lot of data, if I remember correctly. Yeah, a hundred percent. And that was a unique approach. Yeah, it's a data world on relative observability. And you guys just launched a new release of your platform, cloud native platform. What's new in the platform? Can you share an update on what you guys release? >>Yeah, well we did and, and you, you bring up a great point. You know, like it's not just in observably but overall data is exploding. Alright, so three things there. It's like, hey, can your platform even handle the explosion of data? Can it control it over time and make sure that as your business grows, the data doesn't continue explode at the same time. And then for the end users, can they make sense of all this data? Cuz what's the point of having it if the end users can't make sense of it? So actually our product announcement this time is a pretty big refresh of, of a lot of features in our, in our platform. And it actually tackles all three of these particular components. And I'll let Jeff, our head of product, Doug, >>You, you run product, you get the keys to the kingdom, I do product roadmap. People saying, Hey this, take this out. You're under a lot of pressure. What makes the platform platform a great observability product? >>So the keystone of what we do that's different is helping you control the data, right? As we're talking about there's an infinite amount of data. These systems are getting more and more and more complicated. A lot of what we do is help you understand the utility of the telemetry so that you can optimize for keeping and storing and paying for the data that's actually helpful as opposed to the stuff that isn't. >>What's the benefit now with observability, with all the noise out in the marketplace, there's been a shift over the past couple years. Cloud native at scale, you're seeing a lot more automation, almost a set to support the growth for more application development. We had a Docker CEO on earlier today, he said there are more applications being deployed in the past year than in the history of open source. So more and more apps are being deployed, more data's being generated. What's the key to observability right now that's gonna separate the winners from the losers? >>Yeah, I think, you know, not only are there more applications being deployed, but there are smaller and small applications being deployed mostly on containers these days more than if they, hence this conference gets larger and larger every year. Right? So, you know, I think the key is a can your system handle this data explosion is, is the first thing. Not only can it handle the data explosion, but you know, APM solutions have been around for a very long time and those were really introspecting into an application. Whereas these days what's more important is, well how is your application interfacing with every other application in your distributed architecture there, right? So the use case is slightly different there. And then to what Jeff was saying is like once the data is there, not only making use of what is actually useful to you, but then having the end user make sense of it. >>Because we, we, we always think about the technology changes. We forget that the end users are different now we used to have IT operations team operating everything and the developers would write the application, just throw it over the wall. These days the developers have to actually operate this thing in production. So the end users of these systems are very different as well. And you can imagine these are folks, your average developer as maybe not operated things for many years in production before. So they need to, that they need to pick up a new skill set, they need to use new tooling in order to, to do that. So yeah, it's, it's, >>And you got the developer persona, you got a developer that's building products for builders and developers that are building products to be consumed. So they're not, they're not really infrastructure builders, they're just app developers. >>Exactly. Exactly. That's right. And that's what a lot of the new functionality that we're introducing here at the show is all about is helping developers who build software by day and are on call by night, actually get in context. There's so much data chances of when that, when one of those pages goes off and your number comes up, that the problem happens to be in the part of the system that you know a lot about are pretty low, chances are you're gonna get bothered about something else. So we've built a feature, we call it collections that's about putting you in the right context and connecting you into the piece of the system where the problem is to orient you and to get you started. So instead of waiting through, through hundreds of millions of things, you're waiting through the stuff that's in the immediate neighborhood of where the >>Problem is. Yeah. To your point about data, you can't let it go unchecked. That's right. You gotta gotta understand that. And we were talking about containers again with, again with docker, you know, nuance point, but oh, scan your container. But not everyone's scanning the containers security nightmare, right? I mean, >>Well I think one of the things that I, I loved in reading the notes in preparation for you coming up is you've actually created cloud native observability with the goal of eliminating engineering burnout. And what you're talking about there is actually the cognitive burden of when things happen. Yeah, for sure. We we're, you know, we're not just designing for when everything goes right, You need to be prepared for when everything goes wrong and that poor lonely individual in the middle of the night has, it's >>A tough job. >>Has to navigate that >>And, and observability is just one thing you gotta mean like security is another thing. So, so many more things have been piled on top of the developer in addition to actually creating the application. Right? It is. There is a lot. And you know, observably is one of those key things you need to do your job. So as much as, as much as we can make that easier, that's a better bit. Like there are so many things being piled on right now. >>That's the holy grail right there. Because they don't want to be doing exactly >>The work. Exactly. They're not observability experts. >>Exactly. And automating that in. So where do you guys weigh in on the automation wave? Everything's automation. Yeah. Is that kind of a hand waving or what's going on? What's the reality? What's actually happening? >>Yeah, I think automation I think is key. You hear a lot of ai ml ops there. I, I don't know if I really believe in that or having a machine self heal itself or anything like that. But I think automation is key because there are a lot of repeatable tasks in a lot of what you're doing. So once you detect that something goes wrong, generally if you've seen it before, you know what the fix is. So I think automation plays a key on the sense that once it's detected again the second time, the third time, okay, I know what I did the previous time, let, let's make sure we can do that again. So automation I think is key. I think it helps a lot with the burnout. I dunno if I'd go as far as the >>Same burnout's a big deal. >>Well there's an example again in the, in the stuff we're releasing this week, a new feature we call query accelerator. That's a form of automation. Problem is you got all this data, mountain of data, put you in the right context so you're at least in the right neighborhood, but now you need to query it. You gotta get the data to actually inform the specific problem you're trying to solve. And the burden on the developer in that situation is really high. You have to know what you're looking for and you have to know how to efficiently ask for it. So you're not waiting for a long time and >>We >>Built a feature, you tell us what you want, we will figure out how to get it for you efficiently. That's the kind of automation that we're focused on. That's actually a good service. How can we, it >>Sounds >>Blissful. How can we accelerate and optimize what you were gonna do anyway, rather than trying to read your mind or predict the future. >>Yes, >>Savannah, some community forward. Yeah, I, I'm, so I'm curious, you, you clearly lead with a lot of empathy, both of you and, and putting your, well you probably have experience with this as well, but putting your mind or putting yourself in the mind to the developer are, what's that like for you from a product development standpoint? Are you doing a lot of community engagement? Are you talking to developers to try and anticipate what they're gonna be needing next in terms of, of your offering? Or how has that work >>For you? Oh, for sure. So, so I run product, I have a lot of product managers who work for me. Somebody that I used to work with, she was accusing me, but what she called, she called me an anthropologist of a product manager. I >>Get these kind of you, the very good design school vibes from you both of you, which >>Is, and the reason why she said the way you do this, you go and you live with them in order to figure out what a day in their life is really like, what the job is really like, what's easy, what's hard. And that's what we try to aim at and try to optimize for. So that's very much the way that we do all of >>Our work. And that's really also highlights the fact that we're in a market that requires acute realtime data from the customer. Cause it's, and it's all new data. Well >>Yeah, it's all changing. The tools change every day. I mean if we're not watching how, and >>So to your point, you need it in real time as well. The whole point of moving to cloud native is you have a reliable product or service there. And like if you need to wait a few minutes to even know that something's wrong, like you've already lost at that point, you've already lost a ton of customers, potentially. You've already lost a ton of business. You know, to your point about the, the community earlier, one other thing we're trying to do is also give back to the community a little bit. So actually two days ago we just announced the open source of a tool that we've been using in our product for a very long time. But of course our product is, is a paid product, right? But actually open source a part of that tool thus that the broader community can benefit as well. And that tool which, which tool is that? It's, it's called Prom lens. And it's actually the Prometheus project is the open sourced metrics project that everybody uses. So this is a query builder that helps developers understand how to create queries in a much more efficient way. We've had in our product for a long time, but we're like, let's give that back to the community so that the broader community of developers out there can have a much easier time creating these queries as well. What's >>Been the feedback? >>We only now it's two days ago so I'm not, I'm not exactly sure. I imagine >>It's great. They're probably playing with it right now. >>Exactly. Exactly. Exactly. For sure. I imagine. Great. >>Yeah, you guys mentioned burnout before and we heard this a lot now you mentioned in terms of data we've been hearing and reporting about Insta security world, which is also data specific observability ties right into security. Yep. How does a company figure out, first of all, burnout's a big problem. It's more and more data coming. It's like, it's like doesn't stop and the breaches are coming too. How does a company know when they need that their observability strategy is broken? Is there sig signs of you know, burnout? Is there signs of breaches? I mean, what are some of the tell signs that if I'm a CSO I go, you know what, maybe I should check out promisee. When do, when do you guys match in and go we're a perfect fit to solve that problem? >>Yeah, I, I would say, you know, because we're focused on the observability side, less so on the security side, some of those signals are like how many incidents do you have? How many outages do you have? What's the occurrence of these things and how long does it take to recover from from from these particular incidents? How >>Upsetting are we finding customers? >>Upsetting are >>Customer. Exactly. >>And and one trend was seeing >>Not churn happening. Exactly. >>And one trend we're seeing in the industry is that 68% of companies are saying that they're having more incidents over time. Right. And if you have more incidents, you can imagine more engineers are being paid, are being woken up and they're being put under more stress. And one thing you said that very interesting is, you know, I think generally in the observability world, you ideally actually don't want to figure out the problem when it goes wrong. Ideally what you want to do these days is figure out how do I remediate this and get the business back to a running state as quickly as I can. And then when the business isn't burning, let me go and figure out what the underlying root cause is. So the strategy there is changed as well from the APM days where like I don't want to figure out the problem in real time. I wanna make sure my business and my service is running as it should be. And then separately from that, once it is then I wanna go >>Under understand that assume it's gonna happen, be ready to close that isolate >>The >>Fire. Exactly. Exactly. And, and you know, you can imagine, you know the whole movement towards C I C D, like generally when you don't touch a system, nothing goes wrong. You deploy change, first thing you do is not figure out why you change break thing. Get that back like underplay that change roll that change back, get your business back to a estate and then take the time where you're not under pressure, you're not gonna be burnt out to figure out what was it about my change that that broke everything. So, yeah. Got >>It. >>Well it's not surprising that you've added some new exciting customers to the roster. We have. We have. You want to tell the audience who they might >>Be? Yes. It's been a few big names in the last year we're pretty excited about. One is Snapchat, I think everybody knows, knows that application And one is Robin Hood. So you know, you can imagine very large, I'll say tech forward companies that have completed their migrations to, to cloud native or a wallet on their way to Cloudnative and, and we like helping those customers for sure. We also like helping a lot of startups out there cause they start off in the cloud native world. Like if you're gonna build a business today, you're gonna use Kubernetes from day one. Right? But we're actually interestingly seeing more and more of is traditional enterprises who are just early, pretty early on in their cloudnative migration then now starting to adopt cloud native at scale and now they're running to the same problems. As well >>Said, the Gartner data last year was something like 85% of companies had not made that transformation. Right. So, and that, I mean that's looking at larger scale companies, obviously >>A hundred, you're >>Right on the pulse. They >>Have finished it, but a lot of them are starting it now. So we're seeing pilot >>Projects, testing and cadence. And I imagine it's a bit of a different pace when you're working with some of those transforming companies versus those startups that are, are just getting rolling. I >>Love and you know, you have a lot of legacy use case you have to, like, if you're a startup, you can imagine there's no baggage, there's no legacy. You're just starting brand new, right? If you're a large enterprise, you have to really think about, okay, well how do I get my active business moved over? But yeah. >>Yeah. And how do you guys see the whole cloud native scale moving with the hyper scales? Like aws? You've got a lot of multi-cloud conversation. We call it super cloud in our narrative, but there's now this new, we're gonna get some of common services being identified. We're seeing a, we're seeing a lot more people recognize and with Kubernetes that hey, you know what, you could get some common services maybe across clouds with SOS doing storage. We got Min iOS doing some storage. Yeah. Cloud flare, I mean starting to see a lot more non-hyper scale systems. >>Yeah, I mean I, and I think that's the pattern there and I think it, it's, especially for enterprise at the top end, right? You see a, a lot of companies are trying to de-risk by saying, Hey, I, I don't want to bet maybe on one cloud provider, I sort of need to hedge my bets a little bit. And Kubernetes is a great tool to go do that. You can imagine some of these other tools you mentioned is a great way to do that. Observability is another great way to do that. Or the cloud providers have their observability or monitoring tooling, but it's really optimized just for that cloud provider, just for those services there. So if you're really trying to run either your custom applications or a multi-cloud approach, you really can't use one cloud providers solution to go solve that problem. Do you >>Guys see yourselves with that unifying >>Layer? We, we, we are a little bit as that lay because it's agnostic to each of the cloud providers. And the other thing is we actually like to understand where our customers run and then try to run their observability stack on a different cloud provider. Cuz we use the cloud ourselves. We're not running our own data centers of course, but it's an interesting thing where everybody has a common dependency on the cloud provider. So when us e one ofs hate to call them out, but when us E one ofs goes down, imagine half the internet goes down, right? And that's the time that you actually need observability. Right? Seriously. And every other tooling there. So we try to find out where do you run and then we try to actually run you elsewhere. But yeah, >>I like that. And nobody wants to see the ugly bits anyway. Exactly. And we all know who when we're all using someone when everything >>Exactly. Exactly, exactly. >>People off the internet. So it's very, I, I really love that. Martin, Jeff, thank you so much for being here with us. Thank you. What's next? What, how do people find out, how do they get one of the jobs since three Xing your >>Employee growth? We're hiring a lot. I think the best thing is to go check out our website chronosphere.io. You'll find out a lot about our, our, our careers, our job openings, the culture we're trying to build here. Find out a lot about the product as well. If you do have an observability problem, like that's the best place to go to find out about that as well. Right. >>Fantastic. Well if you want to join a quarter billion, a quarter of a billion dollar rocket ship over here and certainly a unicorn, get in touch with Martin and Jeff. John, thank you so much for joining me for this very special edition and thank all of you for tuning in to the Cube live here from Motor City. My name's Savannah Peterson and we'll see you in a little bit. >>Robert Herbeck. People obviously know you from Shark Tanks, but the Herbeck group has been really laser focused on cyber security. So I actually helped to bring my.
SUMMARY :
John, how long have you known the next few? He's a VC partner at Greylock and an investor and this company that just launched their new cloud Jeff, thank you so much for being Thank you. I noticed right away that you have raised a mammoth series C. And a couple of other things we're pretty proud of last year. Real metric if you've had a hundred percent. It's a good metric to, to put out there if you had a hundred percent. and you know, therefore paying more for the service as well. And you guys just launched a new release of your platform, cloud native platform. So actually our product announcement this time is a pretty big refresh of, You, you run product, you get the keys to the kingdom, I do product roadmap. So the keystone of what we do that's different is helping you control the What's the key to observability right now that's gonna separate the winners from the losers? Not only can it handle the data explosion, but you know, APM solutions have been around for And you can imagine these are folks, And you got the developer persona, you got a developer that's building the part of the system that you know a lot about are pretty low, chances are you're gonna get bothered about And we were talking about containers again with, again with docker, you know, nuance point, We we're, you know, we're not just designing for when everything goes right, You need to be prepared for when everything And you know, observably is one of those key things you need to do your job. That's the holy grail right there. Exactly. So where do you guys weigh in on the automation wave? So once you detect that something goes wrong, generally if you've seen it before, you know what the fix is. You gotta get the data to actually inform the specific problem you're trying to solve. Built a feature, you tell us what you want, we will figure out how to get it for you efficiently. How can we accelerate and optimize what you were gonna do anyway, empathy, both of you and, and putting your, well you probably have experience with this as well, of a product manager. Is, and the reason why she said the way you do this, you go and you live with them in order to And that's really also highlights the fact that we're in a market that requires acute realtime I mean if we're not watching how, and And like if you need to wait a few minutes to even know that something's wrong, like you've already lost at that point, I imagine They're probably playing with it right now. I imagine. I mean, what are some of the tell signs that if I'm a CSO I go, you know what, Exactly. Exactly. And if you have more incidents, you can imagine more engineers are being paid, are being woken up and they're being put And, and you know, you can imagine, you know the whole movement towards C I C D, You want to tell the audience who they might So you know, you can imagine very large, Said, the Gartner data last year was something like 85% of companies had not made that transformation. Right on the pulse. So we're seeing pilot And I imagine it's a bit Love and you know, you have a lot of legacy use case you have to, like, if you're a startup, you can imagine there's no baggage, We're seeing a, we're seeing a lot more people recognize and with Kubernetes that hey, you know what, tools you mentioned is a great way to do that. And that's the time that you actually need observability. And we all know who when we're all using someone when Exactly. Martin, Jeff, thank you so much for being here with If you do have an observability problem, like that's the best place to go to find out about of you for tuning in to the Cube live here from Motor City. People obviously know you from Shark Tanks, but the Herbeck group has been really
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Jeff | PERSON | 0.99+ |
Martin | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Jerry Chen | PERSON | 0.99+ |
Jeff Cobb | PERSON | 0.99+ |
Doug | PERSON | 0.99+ |
John | PERSON | 0.99+ |
John Farer | PERSON | 0.99+ |
Savannah Peterson | PERSON | 0.99+ |
Robert Herbeck | PERSON | 0.99+ |
third time | QUANTITY | 0.99+ |
last year | DATE | 0.99+ |
Cuan | LOCATION | 0.99+ |
85% | QUANTITY | 0.99+ |
68% | QUANTITY | 0.99+ |
Martin Mao | PERSON | 0.99+ |
second time | QUANTITY | 0.99+ |
90% | QUANTITY | 0.99+ |
Gartner | ORGANIZATION | 0.99+ |
first | QUANTITY | 0.99+ |
a year ago | DATE | 0.99+ |
200 million | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
hundred percent | QUANTITY | 0.99+ |
two days ago | DATE | 0.99+ |
KubeCon | EVENT | 0.99+ |
Shark Tanks | ORGANIZATION | 0.99+ |
CloudNativeCon | EVENT | 0.98+ |
one | QUANTITY | 0.98+ |
Snapchat | ORGANIZATION | 0.98+ |
hundreds of millions | QUANTITY | 0.98+ |
Docker | ORGANIZATION | 0.98+ |
over 200 people | QUANTITY | 0.98+ |
this week | DATE | 0.97+ |
one trend | QUANTITY | 0.97+ |
Greylock | ORGANIZATION | 0.97+ |
three things | QUANTITY | 0.97+ |
first thing | QUANTITY | 0.97+ |
three | QUANTITY | 0.97+ |
One | QUANTITY | 0.95+ |
past year | DATE | 0.95+ |
about 80 employees | QUANTITY | 0.93+ |
two very interesting gentlemen | QUANTITY | 0.93+ |
Savannah | PERSON | 0.93+ |
Kubernetes | TITLE | 0.92+ |
this afternoon | DATE | 0.92+ |
one thing | QUANTITY | 0.91+ |
Prometheus | TITLE | 0.9+ |
today | DATE | 0.9+ |
each | QUANTITY | 0.9+ |
chronosphere.io | OTHER | 0.89+ |
Robin Hood | PERSON | 0.88+ |
Motor City | LOCATION | 0.87+ |
earlier today | DATE | 0.86+ |
Cuan Detroit, Michigan | LOCATION | 0.85+ |
a quarter of a billion dollar | QUANTITY | 0.82+ |
Chronosphere | ORGANIZATION | 0.81+ |
ton | QUANTITY | 0.81+ |
iOS | TITLE | 0.81+ |
aws | ORGANIZATION | 0.8+ |
half | QUANTITY | 0.78+ |
past couple years | DATE | 0.75+ |
NA 2022 | EVENT | 0.75+ |
Prom lens | OTHER | 0.72+ |
hundred | QUANTITY | 0.68+ |
day one | QUANTITY | 0.65+ |
quarter billion | QUANTITY | 0.63+ |
wave | EVENT | 0.63+ |
KubeCon Preview with Madhura Maskasky
(upbeat music) >> Hello, everyone. Welcome to theCUBE here, in Palo Alto, California for a Cube Conversation. I'm John Furrier, host of theCUBE. This is a KubeCon preview conversation. We got a great guest here, in studio, Madhura Maskasky, Co-Founder and VP of Product, Head of Product at Platform9. Madhura, great to see you. Thank you for coming in and sharing this conversation about, this cube conversation about KubeCon, a Kubecon conversation. >> Thanks for having me. >> A light nice play on words there, a little word play, but the fun thing about theCUBE is, we were there at the beginning when OpenStack was kind of on its transition, Kubernetes was just starting. I remember talking to Lou Tucker back in, I think Seattle or some event and Craig McLuckie was still working at Google at the time. And Google was debating on putting the paper out and so much has happened. Being present at creation, you guys have been there too with Platform9. Present at creation of the Kubernetes wave was not obvious only a few insiders kind of got the big picture. We were one of 'em. We saw this as a big wave. Docker containers at that time was a unicorn funded company. Now they've went back to their roots a few years ago. I think four years ago, they went back and recapped and now they're all pure open source. Since then Docker containers and containers have really powered the Kubernetes wave. Combined with the amazing work of the CNCF and KubeCon which we've been covering every year. You saw the maturation, you saw the wave, the early days, end user projects being contributed. Like Envoy's been a huge success. And then the white spaces filling in on the map, you got observability, you've got run time, you got all the things, still some white spaces in there but it's really been great to watch this growth. So I have to ask you, what do you expect this year? You guys have some cutting edge technology. You got Arlo announced and a lot's going on Kubernetes this year. It's going mainstream. You're starting to see the traditional enterprises embrace and some are scaling faster than others, manage services, plethora of choices. What do you expect this year at KubeCon North America in Detroit? >> Yeah, so I think you summarize kind of that life cycle or lifeline of Kubernetes pretty well. I think I remember the times when, just at the very beginning of Kubernetes, after it was released we were sitting I think with box, box dot com and they were describing to us why they are early adopters of Kubernetes. And we were just sitting down taking notes trying to understand this new project and what value it adds, right? And then flash forward to today where there are Dilbert strips written about Kubernetes. That's how popular it has become. So, I think as that has happened, I think one of the things that's also happened is the enterprises that adopted it relatively early are running it at a massive scale or looking to run it at massive scale. And so I think at scale cloud-native is going to be the most important theme. At scale governance, at scale manageability are going to be top of the mind. And the third factor, I think that's going to be top of the mind is cost control at scale. >> Yeah, and one of the things that we've seen is that the incubated projects a lot more being incubated now and you got the combination of end user and company contributed open source. You guys are contributing RLO >> RLO. >> and open source. >> Yeah. >> That's been part of your game plan there. So you guys are no stranger open source. How do you see this year's momentum? Is it more white space being filled? What's new coming out of the block? What do you think is going to come out of this year? What's rising in terms of traction? What do you see emerging as more notable that might not have been there last year? >> Yeah, so I think it's all about filling that white space, some level of consolidation, et cetera. That's usually the trend in the cloud-native space. And I think it's going to continue to be on that and it's going to be tooling that lets users simplify their lives. Now that Kubernetes is part of your day to day. And so it is observability, et cetera, have always been top of the mind, but I think starting this year, et cetera it's going to be at the next level. Which is gone other times of just running your Prometheus at individual cluster level, just to take that as an example. Now you need a solution- >> Yep. >> that operates at this massive scale across different distributions and your edge locations. So, it's taking those same problems but taking them to that next order of management. >> I'm looking at my notes here and I see orchestration and service mesh, which Envoy does. And you're seeing other solutions come out as well like Linkerd and whatnot. Some are more popular than others. What areas do you see are most needed? If you could go in there and be program chair for a day and you've got a day job as VP of product at Platform9. So you kind of have to have that future view of the roadmap and looking back at where you've come, what would you want to prioritize if you could bring your VP of product skills to the open source and saying, hey, can I point out some needs here? What would you say? >> Yeah, I think just the more tooling that lets people make sense and reduce some of the chaos that this prowling ecosystem of cloud-native creates. Which is tooling, that is not adding more tooling that covers white space is great, but introducing abilities that let you better manage what you have today is probably absolutely top of the mind. And I think that's really not covered today in terms of tools that are around. >> You know, I've been watching the top five incubated projects in CNCF, Argo cracked the top five. I think they got close to 12,000 GitHub stars. They have a conference now, ArgoCon here in California. What is that about? >> Yeah. >> Why is that so popular? I mean, I know it's kind of about obviously workflows and dealing with good pipeline, but why is that so popular right now? >> I think it's very interesting and I think Argo's journey and it's just climbed up in terms of its Github stars for example. And I think it's because as these scale factors that we talk about on one end number of nodes and clusters growing, and on the other end number of sites you're managing grows. I think that CD or continuous deployment of applications it used to kind of be something that you want to get to, it's that north star, but most enterprises wouldn't quite be there. They would either think that they're not ready and it's not needed enough to get there. But now when you're operating at that level of scale and to still maintain consistency without sky rocketing your costs, in terms of ops people, CD almost becomes a necessity. You need some kind of manageable, predictable way of deploying apps without having to go out with new releases that are going out every six months or so you need to do that on a daily basis, even hourly basis. And that's why. >> Scales the theme again, >> Yep. >> back to scale. >> Yep. >> All right, final question. We'll wrap up this preview for KubeCon in Detroit. Whereas we start getting the lay of the land and the focus. If you had to kind of predict the psychology of the developer that's going to be attending in person and they're going to have a hybrid event. So, they will be not as good as being in person. Us, it's going to be the first time kind of post pandemic when I think everyone's going to be together in LA it was a weird time in the calendar and Valencia was the kind of the first international one but this is the first time in North America. So, we're expecting a big audience. >> Mhm. >> If you could predict or what's your view on the psychology of the attendee this year? Obviously pumped to be back. But what do you think they're going to be thinking about? what's on their mind? What are they going to be peaked on? What's the focus? Where will be the psychology? Where will be the mindset? What are people going to be looking for this year? If you had to make a prediction on what the attendees are going to be thinking about what would you say? >> Yeah. So there's always a curiosity in terms of what's new, what new cool tools that are coming out that's going to help address some of the gaps. What can I try out? That's always as I go back to my development roots, first in mind, but then very quickly it comes down to what's going to help me do my job easier, better, faster, at lower cost. And I think again, I keep going back to that theme of automation, declarative automation, automation at scale, governance at scale, these are going to be top of the mind for both developers and ops teams. >> We'll be there covering it like a blanket like we always do from day one, present at creation at KubeCon we are going to be covering again for the consecutive year in a row. We love the CNCF. We love what they do. We thank the developers this year, again continue going mainstream closer and closer to the front lines as the company is the application. As we say, here on theCUBE we'll be there bringing you all the signal. Thanks for coming in and sharing your thoughts on KubeCon 2022. >> Thank you for having me. >> Okay. I'm John Furrier here in theCUBE in Palo Alto, California. Thanks for watching. (upbeat music)
SUMMARY :
Co-Founder and VP of Product, in on the map, you got observability, think that's going to be top Yeah, and one of the What do you see emerging and it's going to be but taking them to that of product skills to the and reduce some of the chaos in CNCF, Argo cracked the top five. and it's not needed enough to get there. Us, it's going to be the first of the attendee this year? of the mind for both We thank the developers this year, in theCUBE in Palo Alto, California.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Madhura Maskasky | PERSON | 0.99+ |
Craig McLuckie | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
Lou Tucker | PERSON | 0.99+ |
California | LOCATION | 0.99+ |
LA | LOCATION | 0.99+ |
North America | LOCATION | 0.99+ |
Madhura | PERSON | 0.99+ |
Detroit | LOCATION | 0.99+ |
ORGANIZATION | 0.99+ | |
last year | DATE | 0.99+ |
first time | QUANTITY | 0.99+ |
Argo | ORGANIZATION | 0.99+ |
Palo Alto, California | LOCATION | 0.99+ |
four years ago | DATE | 0.99+ |
KubeCon | EVENT | 0.99+ |
Seattle | LOCATION | 0.99+ |
one | QUANTITY | 0.98+ |
this year | DATE | 0.98+ |
both | QUANTITY | 0.98+ |
third factor | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
Platform9 | ORGANIZATION | 0.97+ |
12,000 | QUANTITY | 0.97+ |
a day | QUANTITY | 0.97+ |
Kubernetes | TITLE | 0.97+ |
Prometheus | TITLE | 0.97+ |
Linkerd | ORGANIZATION | 0.95+ |
CNCF | ORGANIZATION | 0.95+ |
first | QUANTITY | 0.94+ |
KubeCon 2022 | EVENT | 0.93+ |
Dilbert | PERSON | 0.93+ |
theCUBE | ORGANIZATION | 0.92+ |
wave | EVENT | 0.91+ |
day one | QUANTITY | 0.91+ |
few years ago | DATE | 0.9+ |
Valencia | LOCATION | 0.9+ |
ArgoCon | EVENT | 0.9+ |
five incubated projects | QUANTITY | 0.87+ |
KubeCon | ORGANIZATION | 0.86+ |
Envoy | ORGANIZATION | 0.85+ |
first international | QUANTITY | 0.84+ |
six months | QUANTITY | 0.81+ |
Kubecon | ORGANIZATION | 0.74+ |
top five | QUANTITY | 0.71+ |
Docker | ORGANIZATION | 0.69+ |
north star | LOCATION | 0.63+ |
Arlo | TITLE | 0.54+ |
pandemic | EVENT | 0.52+ |
Github | ORGANIZATION | 0.5+ |
OpenStack | TITLE | 0.49+ |
Kubernetes | PERSON | 0.47+ |
GitHub | TITLE | 0.47+ |
Cube | EVENT | 0.3+ |
Anish Dhar & Ganesh Datta, Cortex | Kubecon + Cloudnativecon Europe 2022
>> Narrator: TheCUBE presents Kubecon and Cloudnativecon Europe, 2022. Brought to you by Red Hat, the cloud native computing foundation and its ecosystem partners. >> Welcome to Valencia, Spain in Kubecon, Cloudnativecon Europe, 2022. I'm Keith Townsend and we are in a beautiful locale. The city itself is not that big, 100,000, I mean, sorry, about 800,000 people. And we got out, got to see a little bit of the sites. It is an amazing city. I'm from the US, it's hard to put in context how a city of 800,000 people can be so beautiful. I'm here with Anish Dhar and Ganesh Datta, Co-founder and CTO of Cortex. Anish you're CEO of Cortex. We were having a conversation. One of the things that I asked my client is what is good. And you're claiming to answer the question about what is quality when it comes to measuring microservices? What is quality? >> Yeah, I think it really depends on the company and I think that's really the philosophy we have. When we built Cortex, is that we understood that different companies have different definitions of quality, but they need to be able to be represented in really objective ways. I think what ends up happening in most engineering organizations is that quality lives in people's heads. The engineers who write the services they're often the ones who understand all the intricacies with the service. What are the downstream dependencies, who's on call for this service? Where does the documentation live? All of these things I think impact the quality of the service. And as these engineers leave the company or they switch teams, they often take that tribal knowledge with them. And so I think quality really comes down to being able to objectively codify your best practices in some way and have that distributed to all engineers in the company. >> And to add to that, I think very concrete examples for an organization that's already modern like their idea of quality might be uptime incidents. For somebody that's like going through a modernization strategy, they're trying to get to the 21st century, they're trying to get to Kubernetes. For them, quality means where are we in that journey? Are you on our latest platforms? Are you running CI, are you doing continuous delivery? Like quality can mean a lot of things and so our perspective is how do we give you the tools to say as an organization, here's what quality means to us. >> So at first, my mind was going through when you said quality, Anish, you started out the conversation about having this kind of non-codified set of measurements, historical knowledge, et cetera. I was thinking observability, measuring how much time does it take to have a transaction. But Ganesh you're introducing this new thing. I'm working with this project where we're migrating a monolith application to a set of microservices. And you're telling me Cortex helps me measure the quality of what I'm doing in my project? >> Ganesh: Absolutely. >> How is that? >> Yeah, it's a great question. So I think when you think about observability, you think about uptime and latency and transactions and throughput and all this stuff. And I think that's very high level and I think that's one perspective of what quality is, but as you're going through this journey, you might say like the fact that we're tracking that stuff, the fact that you're using APM, you're using distributed tracing, that is one element of service quality. Maybe service quality means you're doing CICD, you're running vulnerability scans. You're using Docker. Like what that means to us can be very different. So observability is just one aspect of are you doing things the right way? Good to us means you're using SLOs. You are tracking those metrics. You're reporting that somewhere. And so that's like one component for our organization of what quality can mean. >> I'm kind of taken back by this because I've not seen someone kind of give the idea. And I think later on, this is the perfect segment to introduce theCUBE clock in which I'm going to give you a minute to kind of like give me the elevator pitch, but we're going to have the deep conversation right now. When you go in and you... What's the first process you do when you engage in a customer? Does a customer go and get this off of repository, install it, the open source version, and then what? I mean, what's the experience? >> Yeah, absolutely. So we have both a smart and on-prem version of Cortex. It's really straightforward. Basically we have a service discovery onboarding flow where customers can connect to different sets of source for their services. It could be Kubernetes, ECS, Git Repos, APM tools, and then we'll actually automatically map all of that service data with all of the integration data in the company. So we'll take that service and map it to its on call rotation to the JIRA tickets that have the service tag associated with it, to the data algo SLOs. And what that ends ends up producing is this service catalog that has all the information you need to understand your service. Almost like a single pane of glass to work with the service. And then once you have all of that data inside Cortex, then you can start writing scorecards, which grade the quality of those services across those different verticals Ganesh was talking about. Like whether it's a monolith, a microservice transition, whether it's production readiness or security standards, you can really start tracking that. And then engineers start understanding where the areas of risk with my service across reliability or security or operation maturity. I think it gives us in insane visibility into what's actually being built and the quality of that compared to your standards. >> So, okay, I have a standards for SLO that is usually something that is, it might not even be measured. So how do you help me understand that I'm lacking a measurable system for tracking SLO and what's the next step for helping me get that system? >> Yeah, I think our perspective is very much how do we help you create a culture where developers understand what's expected of them? So if SLOs are part of what we consider observability or reliability, then Cortex's perspective is, hey, we want to help your organization adopt SLOs. And so that service cataloging concept, the service catalog says, hey, here's my API integration. Then a scorecard, the organization goes in and says, we want every service owner to define their SLOs, we want you to define your thresholds. We want you to be tracking them, are you passing your SLOs? And so we're not being prescriptive about here's what we think your SLOs should be, ours is more around, hey, we're going to help you like if you care about SLOs, we're going to tell the service owners saying, hey, you need to have at least two SLOs for your service and you got to be tracking them. And the service catalog that data flows from a service catalog into those scorecards. And so we're helping them adopt that mindset of, hey, SLOs are important. It is a component of like a holistic service reliability excellence metric that we care about. >> So what happens when I already have systems for like SLO, how do I integrate that system with Cortex? >> That's one of the coolest things. So the service catalog can be pretty smart about it. So let's say you've sucked in your services from your GitHub. And so now your services are in Cortex. What we can do is we can actually discover from your APM tools, you can say like, hey, for this service, we have guessed that this is the corresponding APM in Datadog. And so from Datadog, here are your SLOs, here are your monitors. And so we can start mapping all the different parts of your world into the Cortex. And that's the power of the service catalog. The service catalog says, given a service, here's everything about that service. Here's the vulnerability scans. Here's the APM, the monitors, the SLOs, the JIRA ticket is like all that stuff comes into a single place. And then our scorecards product can go back out and say, hey, Datadog, tell me about this SLOs for the service. And so we're going to get that information live and then score your services against that. And so we're like integrating with all of your third party tools and integrations to create that single pan of glass. >> Yeah, and to add to that, I think one of the most interesting use cases with scorecards is, okay, which teams have actually adopted SLOs in the first place? I think a lot of companies struggle with how do we make sure engineers defined SLOs are passing them actually care about them. And scorecards can be used to one, which teams are actually meeting these guidelines? And then two, let's get those teams adopted on SLOs. Let's track that, you can do all of that in Cortex, which is I think a really interesting use case that we've seen. >> So let's talk about kind of my use case in the end to end process for integrating Cortex into migrations. So I have this monolithic application, I want to break it into microservices and then I want to ensure that I'm delivering if not, you know what, let's leave it a little bit more open ended. How do I know that I'm better at the end of I was in a monolith before, how do I measure that now that I'm in microservices and on cloud native, that I'm better? >> That's a good question. I think it comes down to, and we talk about this all the time for our customers that are going through that process. You can't define better if you don't define a baseline, like what does good mean to us? And so you need to start by saying, why are we moving to microservices? Is it because we want teams to move faster? Is it because we care about reliability up time? Like what is the core metric that we're tracking? And so you start by defining that as an organization. And that is kind of like a hand wavy thing. Why are we doing microservices? Once you have that, then you define this scorecard. And that's like our golden path. Once we're done doing this microservice migration, can we say like, yes, we have been successful and those metrics that we care about are being tracked. And so where Cortex fits in is from the very first step of creating a service, you can use Cortex to define templates. Like one click, you go in, it spins up a microservice for you that follows all your best practices. And so from there, ideally you're meeting 80% of your standards already. And then you can use scorecards to track historical progress. So you can say, are we meeting our golden path standards? Like if it's uptime, you can track uptime metrics and scorecards. If it's around velocity, you can track velocity metrics. Is it just around modernization? Are you doing CICD and vulnerability scans, like moving faster as a team? You can track that. And so you can start seeing like trends at a per team level, at a per department level, at a per product level saying, hey, we are seeing consistent progress in the metrics that we care about. And this microservice journey is helping us with that. So I think that's the kind of phased progress that we see with Cortex. >> So I'm going to give you kind of a hand wavy thing. We're told that cloud native helps me to do things faster with less defects so that I can do new opportunities. Let's stretch into kind of this non-tech, this new opportunities perspective. I want to be able to move my microservices. I want to be able to move my architecture to microservices, so I reduce call wait time on my customer service calls. So I can easily see how I can measure are we iterating faster? Are we putting out more updates quicker? That's pretty easy to measure. The number of defects, easy to measure. I can imagine a scorecard, but what about this wait time? I don't necessarily manage the call center system, but I get the data. How do I measure that the microservice migration was successful from a business process perspective? >> Yeah, that's a good question. I think it comes down to two things. One, the flexibility of scorecard means you can pipe in that data to Cortex. And what we recommend customers is track the outcome metrics and track the input metrics as well. And so what is the input metric to call wait time? Like maybe it's the fact that if something goes wrong, we have the run books to quickly roll back to an older version that we know is running. That way MTTR is faster. Or when something happens, we know the owner for that service and we can go back to them and say like, hey, we're going to ping you as an incident commander. Those are kind of the input metrics to, if we do these things, then we know our call wait time is going to drop because we're able to respond faster to incidents. And so you want to track those input metrics. And then you want to track the output metrics as well. And so if you have those metrics coming in from your Prometheus or your Datadogs or whatever, you can pipe that into Cortex and say, hey, we're going to look at both of these things holistically. So we want to see is there a correlation between those input metrics like are we doing things the right way, versus are we seeing the value that we want to come out of that? And so I think that's the value of Cortex is not so much around, hey, we're going to be prescriptive about it. It's here's this framework that will let you track all of that and say, are we doing things the right way and is it giving us the value that we want? And being able to report that update to engineer leadership and say, hey, maybe these services are not doing like we're not improving call wait time. Okay, why is that? Are these services behind on the actual input metrics that we care about? And so being able to see that I think is super valuable. >> Yeah, absolutely, I think just to touch on the reporting, I think that's one of the most value add things Cortex can provide. If you think about it, the service is atomic unit of your software. It represents everything that's being built and that bubbles up into teams, products, business units, and Cortex lets you represent that. So now I can, as a CTO, come in and say, hey, these product lines are they actually meeting our standards? Where are the areas of risk? Where should I be investing more resources? I think Cortex is almost like the best way to get the actual health of your engineering organization. >> All right Anish and Ganesh. We're going to go into the speed round here. >> Ganesh: It's time for the Q clock? >> Time for the Q clock. Start the Q clock. (upbeat music) Let's go on. >> Ganesh: Let's do it. >> Anish: Let's do it. >> Let's go on. You're you're 10 seconds in. >> Oh, we can start talking. Okay, well I would say, Anish was just touching on this. For a CTO, their question is how do I know if engineering quality is good? And they don't care about the microservice level. They care about as a business, is my engineering team actually producing. >> Keith: Follow the green, not the dream. (Ganesh laughs) >> And so the question is, well, how do we codify service quality? We don't want this to be a hand wavy thing that says like, oh, my team is good, my team is bad. We want to come in and define here's what service quality means. And we want that to be a number. You want that to be something that can- >> A goal without a timeline is just a dream. >> And CTO comes in and they say, here's what we care about. Here's how we're tracking it. Here are the teams that are doing well. We're going to reward the winners. We're going to move towards a world where every single team is doing service quality. And that's where Cortex can provide. We can give you that visibility that you never have before. >> For that five seconds. >> And hey, your SRE can't be the one handling all this. So let Cortex- >> Shoot the bad guy. >> Shot that, we're done. From Valencia Spain, I'm Keith Townsend. And you're watching theCube. The leader in high tech coverage. (soft music) (soft music) >> Narrator: TheCube presents Kubecon and Cloudnativecon Europe, 2022 brought to you by Red Hat, the cloud native computing foundation and its ecosystem partners. >> Welcome to Valencia, Spain in Kubecon, Cloudnativecon Europe, 2022. I'm Keith Townsend. And we are in a beautiful locale. The city itself is not that big 100,000, I mean, sorry, about 800,000 people. And we got out, got to see a little bit of the sites. It is an amazing city. I'm from the US, it's hard to put in context how a city of 800,000 people can be so beautiful. I'm here with Anish Dhar and Ganesh Datta, Co-founder and CTO of Cortex. Anish you're CEO of Cortex. We were having a conversation. One of the things that I asked my client is what is good. And you're claiming to answer the question about what is quality when it comes to measuring microservices? What is quality? >> Yeah, I think it really depends on the company. And I think that's really the philosophy we have when we build Cortex is that we understood that different companies have different definitions of quality, but they need to be able to be represented in really objective ways. I think what ends up happening in most engineering organizations is that quality lives in people's heads. Engineers who write the services, they're often the ones who understand all the intricacies with the service. What are the downstream I dependencies, who's on call for this service, where does the documentation live? All of these things, I think impact the quality of the service. And as these engineers leave the company or they switch teams, they often take that tribal knowledge with them. And so I think quality really comes down to being able to objectively like codify your best practices in some way, and have that distributed to all engineers in the company. >> And to add to that, I think like very concrete examples for an organization that's already modern their idea of quality might be uptime incidents. For somebody that's like going through a modernization strategy, they're trying to get to the 21st century. They're trying to get to Kubernetes. For them quality means like, where are we in that journey? Are you on our latest platforms? Are you running CI? Are you doing continuous delivery? Like quality can mean a lot of things. And so our perspective is how do we give you the tools to say as an organization here's what quality means to us. >> So at first my mind was going through when you said quality and as you started out the conversation about having this kind of non codified set of measurements, historical knowledge, et cetera. I was thinking observability measuring how much time does it take to have a transaction? But Ganesh you're introducing this new thing. I'm working with this project where we're migrating a monolith application to a set of microservices. And you're telling me Cortex helps me measure the quality of what I'm doing in my project? >> Ganesh: Absolutely. >> How is that? >> Yeah, it's a great question. So I think when you think about observability, you think about uptime and latency and transactions and throughput and all this stuff and I think that's very high level. And I think that's one perspective of what quality is. But as you're going through this journey, you might say like the fact that we're tracking that stuff, the fact that you're using APM, you're using distributed tracing, that is one element of service quality. Maybe service quality means you're doing CICD, you're running vulnerability scans. You're using Docker. Like what that means to us can be very different. So observability is just one aspect of, are you doing things the right way? Good to us means you're using SLOs. You are tracking those metrics. You're reporting that somewhere. And so that's like one component for our organization of what quality can mean. >> Wow, I'm kind of taken me back by this because I've not seen someone kind of give the idea. And I think later on, this is the perfect segment to introduce theCube clock in which I'm going to give you a minute to kind of like give me the elevator pitch, but we're going to have the deep conversation right now. When you go in and you... what's the first process you do when you engage in a customer? Does a customer go and get this off of repository, install it, the open source version and then what, I mean, what's the experience? >> Yeah, absolutely. So we have both a smart and on-prem version of Cortex. It's really straightforward. Basically we have a service discovery onboarding flow where customers can connect to different set of source for their services. It could be Kubernetes, ECS, Git Repos, APM tools, and then we'll actually automatically map all of that service data with all of the integration data in the company. So we'll take that service and map it to its on call rotation to the JIRA tickets that have the service tag associated with it, to the data algo SLOs. And what that ends up producing is this service catalog that has all the information you need to understand your service. Almost like a single pane of glass to work with the service. And then once you have all of that data inside Cortex, then you can start writing scorecards, which grade the quality of those services across those different verticals Ganesh was talking about. like whether it's a monolith, a microservice transition, whether it's production readiness or security standards, you can really start tracking that. And then engineers start understanding where are the areas of risk with my service across reliability or security or operation maturity. I think it gives us insane visibility into what's actually being built and the quality of that compared to your standards. >> So, okay, I have a standard for SLO. That is usually something that is, it might not even be measured. So how do you help me understand that I'm lacking a measurable system for tracking SLO and what's the next step for helping me get that system? >> Yeah, I think our perspective is very much how do we help you create a culture where developers understand what's expected of them? So if SLOs are part of what we consider observability and reliability, then Cortex's perspective is, hey, we want to help your organization adopt SLOs. And so that service cataloging concept, the service catalog says, hey, here's my APM integration. Then a scorecard, the organization goes in and says, we want every service owner to define their SLOs. We want to define your thresholds. We want you to be tracking them. Are you passing your SLOs? And so we're not being prescriptive about here's what we think your SLOs should be. Ours is more around, hey, we're going to help you like if you care about SLOs, we're going to tell the service owners saying, hey, you need to have at least two SLOs for your service and you've got to be tracking them. And the service catalog that data flows from the service catalog into those scorecards. And so we're helping them adopt that mindset of, hey, SLOs are important. It is a component of like a holistic service reliability excellence metric that we care about. >> So what happens when I already have systems for like SLO, how do I integrate that system with Cortex? >> That's one of the coolest things. So the service catalog can be pretty smart about it. So let's say you've sucked in your services from your GitHub. And so now your services are in Cortex. What we can do is we can actually discover from your APM tools, we can say like, hey, for this service we have guessed that this is the corresponding APM in Datadog. And so from Datadog, here are your SLOs, here are your monitors. And so we can start mapping all the different parts of your world into the Cortex. And that's the power of the service catalog. The service catalog says, given a service, here's everything about that service. Here's the vulnerability scans, here's the APM, the monitor, the SLOs, the JIRA ticket, like all that stuff comes into a single place. And then our scorecard product can go back out and say, hey, Datadog, tell me about this SLOs for the service. And so we're going to get that information live and then score your services against that. And so we're like integrating with all of your third party tools and integrations to create that single pan of glass. >> Yeah and to add to that, I think one of the most interesting use cases with scorecards is, okay, which teams have actually adopted SLOs in the first place? I think a lot of companies struggle with how do we make sure engineers defined SLOs are passing them actually care about them? And scorecards can be used to one, which teams are actually meeting these guidelines? And then two let's get those teams adopted on SLOs. Let's track that. You can do all of that in Cortex, which is, I think a really interesting use case that we've seen. >> So let's talk about kind of my use case in the end to end process for integrating Cortex into migrations. So I have this monolithic application, I want to break it into microservices and then I want to ensure that I'm delivering you know what, let's leave it a little bit more open ended. How do I know that I'm better at the end of I was in a monolith before, how do I measure that now that I'm in microservices and on cloud native, that I'm better? >> That's a good question. I think it comes down to, and we talk about this all the time for our customers that are going through that process. You can't define better if you don't define a baseline, like what does good mean to us? And so you need to start by saying, why are we moving to microservices? Is it because we want teams to move faster? Is it because we care about reliability up time? Like what is the core metric that we're tracking? And so you start by defining that as an organization. And that is kind of like a hand wavy thing. Why are we doing microservices? Once you have that, then you define the scorecard and that's like our golden path. Once we're done doing this microservice migration, can we say like, yes, we have been successful. And like those metrics that we care about are being tracked. And so where Cortex fits in is from the very first step of creating a service. You can use Cortex to define templates. Like one click, you go in, it spins up a microservice for you that follows all your best practices. And so from there, ideally you're meeting 80% of your standards already. And then you can use scorecards to track historical progress. So you can say, are we meeting our golden path standards? Like if it's uptime, you can track uptime metrics and scorecards. If it's around velocity, you can track velocity metrics. Is it just around modernization? Are you doing CICD and vulnerability scans, like moving faster as a team? You can track that. And so you can start seeing like trends at a per team level, at a per department level, at a per product level. Saying, hey, we are seeing consistent progress in the metrics that we care about. And this microservice journey is helping us with that. So I think that's the kind of phased progress that we see with Cortex. >> So I'm going to give you kind of a hand wavy thing. We're told that cloud native helps me to do things faster with less defects so that I can do new opportunities. Let's stretch into kind of this non-tech, this new opportunities perspective. I want to be able to move my microservices. I want to be able to move my architecture to microservices so I reduce call wait time on my customer service calls. So, I could easily see how I can measure are we iterating faster? Are we putting out more updates quicker? That's pretty easy to measure. The number of defects, easy to measure. I can imagine a scorecard. But what about this wait time? I don't necessarily manage the call center system, but I get the data. How do I measure that the microservice migration was successful from a business process perspective? >> Yeah, that's a good question. I think it comes down to two things. One, the flexibility of scorecard means you can pipe in that data to Cortex. And what we recommend customers is track the outcome metrics and track the input metrics as well. And so what is the input metric to call wait time? Like maybe it's the fact that if something goes wrong, we have the run book to quickly roll back to an older version that we know is running that way MTTR is faster. Or when something happens, we know the owner for that service and we can go back to them and say like, hey, we're going to ping you as an incident commander. Those are kind the input metrics to, if we do these things, then we know our call wait time is going to drop because we're able to respond faster to incidents. And so you want to track those input metrics and then you want to track the output metrics as well. And so if you have those metrics coming in from your Prometheus or your Datadogs or whatever, you can pipe that into Cortex and say, hey, we're going to look at both of these things holistically. So we want to see is there a correlation between those input metrics? Are we doing things the right way versus are we seeing the value that we want to come out of that? And so I think that's the value of Cortex is not so much around, hey, we're going to be prescriptive about it. It's here's this framework that will let you track all of that and say, are we doing things the right way and is it giving us the value that we want? And being able to report that update to engineer leadership and say, hey, maybe these services are not doing like we're not improving call wait time. Okay, why is that? Are these services behind on like the actual input metrics that we care about? And so being able to see that I think is super valuable. >> Yeah, absolutely. I think just to touch on the reporting, I think that's one of the most value add things Cortex can provide. If you think about it, the service is atomic unit of your software. It represents everything that's being built and that bubbles up into teams, products, business units, and Cortex lets you represent that. So now I can, as a CTO, come in and say, hey, these product lines are they actually meeting our standards? Where are the areas of risk? Where should I be investing more resources? I think Cortex is almost like the best way to get the actual health of your engineering organization. >> All right, Anish and Ganesh. We're going to go into the speed round here. >> Ganesh: It's time for the Q clock >> Time for the Q clock. Start the Q clock. (upbeat music) >> Let's go on. >> Ganesh: Let's do it. >> Anish: Let's do it. >> Let's go on, you're 10 seconds in. >> Oh, we can start talking. Okay, well I would say, Anish was just touching on this, for a CTO, their question is how do I know if engineering quality is good? And they don't care about the microservice level. They care about as a business, is my enduring team actually producing- >> Keith: Follow the green, not the dream. (Ganesh laughs) >> And so the question is, well, how do we codify service quality? We don't want this to be a hand wavy thing that says like, oh, my team is good, my team is bad. We want to come in and define here's what service quality means. And we want that to be a number. You want that to be something that you can- >> A goal without a timeline is just a dream. >> And a CTO comes in and they say, here's what we care about, here's how we're tracking it. Here are the teams that are doing well. We're going to reward the winners. We're going to move towards a world where every single team is doing service quality. And that's what Cortex can provide. We can give you that visibility that you never had before. >> For that five seconds. >> And hey, your SRE can't be the one handling all this. So let Cortex- >> Shoot the bad guy. >> Shot that, we're done. From Valencia Spain, I'm Keith Townsend. And you're watching theCube, the leader in high tech coverage. (soft music)
SUMMARY :
Brought to you by Red Hat, And we got out, got to see and have that distributed to how do we give you the tools the quality of what I'm So I think when you think What's the first process you do that has all the information you need So how do you help me we want you to define your thresholds. And so we can start mapping adopted SLOs in the first place? in the end to end process And so you can start seeing like trends So I'm going to give you And so if you have those metrics coming in and Cortex lets you represent that. the speed round here. Time for the Q clock. You're you're 10 seconds in. the microservice level. Keith: Follow the green, not the dream. And so the question is, well, timeline is just a dream. that you never have before. And hey, your SRE can't And you're watching theCube. 2022 brought to you by Red Hat, And we got out, got to see and have that distributed to how do we give you the tools the quality of what I'm So I think when you think And I think later on, this that has all the information you need So how do you help me And the service catalog that data flows And so we can start mapping You can do all of that in the end to end process And so you can start seeing So I'm going to give you And so if you have those metrics coming in I think just to touch on the reporting, the speed round here. Time for the Q clock. the microservice level. Keith: Follow the green, not the dream. And so the question is, well, timeline is just a dream. that you never had before. And hey, your SRE can't And you're watching theCube,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Anish | PERSON | 0.99+ |
Keith Townsend | PERSON | 0.99+ |
Cortex | ORGANIZATION | 0.99+ |
80% | QUANTITY | 0.99+ |
Keith | PERSON | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
US | LOCATION | 0.99+ |
Ganesh | PERSON | 0.99+ |
21st century | DATE | 0.99+ |
100,000 | QUANTITY | 0.99+ |
10 seconds | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
five seconds | QUANTITY | 0.99+ |
two things | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
Valencia, Spain | LOCATION | 0.99+ |
800,000 people | QUANTITY | 0.99+ |
Cortex | TITLE | 0.99+ |
Valencia Spain | LOCATION | 0.99+ |
one element | QUANTITY | 0.99+ |
one aspect | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
Cloudnativecon | ORGANIZATION | 0.99+ |
one perspective | QUANTITY | 0.99+ |
Datadog | ORGANIZATION | 0.99+ |
one component | QUANTITY | 0.99+ |
Ganesh Datta | PERSON | 0.98+ |
One | QUANTITY | 0.98+ |
SLO | TITLE | 0.98+ |
2022 | DATE | 0.98+ |
first step | QUANTITY | 0.98+ |
Kubecon | ORGANIZATION | 0.97+ |
about 800,000 people | QUANTITY | 0.97+ |
one click | QUANTITY | 0.97+ |
Kickoff with Taylor Dolezal | Kubecon + Cloudnativecon Europe 2022
>> Announcer: "theCUBE" presents "Kubecon and Cloudnativecon Europe, 2022" brought to you by Red Hat, the Cloud Native Computing Foundation and its ecosystem partners. >> Welcome to Valencia, Spain and "Kubecon + Cloudnativecon Europe, 2022." I'm Keith Townsend, and we're continuing the conversations with amazing people doing amazing things. I think we've moved beyond a certain phase of the hype cycle when it comes to Kubernetes. And we're going to go a little bit in detail with that today, and on all the sessions, I have today with me, Taylor Dolezal. New head of CNCF Ecosystem. So, first off, what does that mean new head of? You're the head of CNCF Ecosystem? What is the CNCF Ecosystem? >> Yeah. Yeah. It's really the end user ecosystem. So, the CNCF is comprised of really three pillars. And there's the governing board, they oversee the budget and fun things, make sure everything's signed and proper. Then there's the Technical Oversight Committee, TOC. And they really help decide the technical direction of the organization through deliberation and talking about which projects get invited and accepted. Projects get donated, and the TOC votes on who's going to make it in, based on all this criteria. And then, lastly, is the end user ecosystem, that encompasses a whole bunch of different working groups, special interest groups. And that's been really interesting to kind of get a deeper sense into, as of late. So, there are groups like the developer experience group, and the user research group. And those have very specific focuses that kind of go across all industries. But what we've seen lately, is that there are really deep wants to create, whether it be financial services user group, and things like that, because end users are having trouble with going to all of the different meetings. If you're a company, a vendor member company that's selling authentication software, or something in networking, makes sense to have a SIG network, SIG off, and those kinds of things. But when it comes down to like Boeing that just joined, does that make sense for them to jump into all those meetings? Or does it make sense to have some other kind of thing that is representative of them, so that they can attend that one thing, it's specific to their industry? They can get that download and kind of come up to speed, or find the best practices as quickly as possible in a nice synthesized way. >> So, you're 10 weeks into this role. You're coming from a customer environment. So, talk to me a little bit about the customer side of it? When you're looking at something, it's odd to call CNCF massive. But it is, 7.1 million members, and the number of contributing projects, et cetera. Talk to me about the view from the outside versus the view now that you're inside? >> Yeah, so honestly, it's been fun to kind of... For me, it's really mirrored the open-source journey. I've gone to Kubecon before, gotten to enjoy all of the booths, and trying to understand what's going on, and then worked for HashiCorp before coming to the CNCF. And so, get that vendor member kind of experience working the booth itself. So, kind of getting deeper and deeper into the stack of the conference itself. And I keep saying, vendor member and end user members, the difference between those, is end users are not organizations that sell cloud native services. Those are the groups that are kind of more consuming, the Airbnbs, the Boeings, the Mercedes, these people that use these technologies and want to kind of give that feedback back to these projects. But yeah, very incredibly massive and just sprawling when it comes to working in all those contexts. >> So, I have so many questions around, like the differences between having you as an end user and in inter-operating with vendors and the CNCF itself. So, let's start from the end user lens. When you're an end user and you're out discovering open-source and cloud native products, what's that journey like? How do you go from saying, okay, I'm primarily focused on vendor solutions, to let me look at this cloud native stack? >> Yeah, so really with that, there's been, I think that a lot of people have started to work with me and ask for, "Can we have recommended architectures? Can we have blueprints for how to do these things?" When the CNCF doesn't want to take that position, we don't want to kind of be the king maker and be like, this is the only way forward. We want to be inclusive, we want to pull in these projects, and kind of give everyone the same boot strap and jump... I missing the word of it, just ability to kind of like springboard off of that. Create a nice base for everybody to get started with, and then, see what works out, learn from one another. I think that when it comes to Kubernetes, and Prometheus, and some other projects, being able to share best practices between those groups of what works best as well. So, within all of the separations of the CNCF, I think that's something I've found really fun, is kind of like seeing how the projects relate to those verticals and those groups as well. Is how you run a project, might actually have a really good play inside of an organization like, "I like that idea. Let's try that out with our team." >> So, like this idea of springboarding. You know, is when an entrepreneur says, "You know what? I'm going to quit my job and springboard off into doing something new." There's a lot of uncertainty, but for enterprise, that can be really scary. Like we're used to our big vendors, HashiCorp, VMware, Cisco kind of guiding us and telling us like, what's next? What is that experience like, springboarding off into something as massive as cloud native? >> So, I think it's really, it's a great question. So, I think that's why the CNCF works so well, is the fact that it's a safe place for all these companies to come together, even companies of competing products. you know, having that common vision of, we want to make production boring again, we don't want to have so much sprawl and have to take in so much knowledge at once. Can we kind of work together to create all these things to get rid of our adminis trivia or maintenance tasks? I think that when it comes to open-source in general, there's a fantastic book it's called "Working in Public," it's by Stripe Press. I recommend it all over the place. It's orange, so you'll recognize it. Yeah, it's easy to see. But it's really good 'cause it talks about the maintainer journey, and what things make it difficult. And so, I think that that's what the CNCF is really working hard to try to get rid of, is all this monotonous, all these monotonous things, filing issues, best practices. How do you adopt open-source within your organization? We have tips and tricks, and kind of playbooks in ways that you could accomplish that. So, that's what I find really useful for those kinds of situations. Then it becomes easier to adopt that within your organization. >> So, I asked Priyanka, CNCF executive director last night, a pretty tough question. And this is kind of in the meat of what you do. What happens when you? Let's pick on service mesh 'cause everyone likes to pick on service mesh. >> XXXX: Yeah. >> What happens when there's differences at that vendor level on the direction of a CIG or a project, or the ecosystem around service mesh? >> Yeah, so that's the fun part. Honestly, is 'cause people get to hash it out. And so, I think that's been the biggest thing for me finding out, was that there's more than one way to do thing. And so, I think it always comes down to use case. What are you trying to do? And then you get to solve after that. So, it really is, I know it depends, which is the worst answer. But I really do think that's the case, because if you have people that are using something within the automotive space, or in the financial services space, they're going to have completely different needs, wants, you know, some might need to run Coball or Fortran, others might not have to. So, even at that level, just down to what your tech stack looks like, audits, and those kinds of things, that can just really differ. So, I think it does come down to something more like that. >> So, the CNCF loosely has become kind of a standards body. And it's centered around the core project Kubernetes? >> Mm-hmm. >> So, what does it mean, when we're looking at larger segments such as service mesh or observability, et cetera, to be Kubernetes compliant? Where's the point, if any, that the CNCF steps in versus just letting everyone hash it out? Is it Kubernetes just need to be Kubernetes compliant and everything else is free for all? >> Honestly, in many cases, it's up to the communities themselves to decide that. So, the groups that are running OCI, the Open Container Interface, Open Storage Interface, all of those things that we've agreed on as ways to implement those technologies, I think that's where the CNCF, that's the line. That's where the CNCF gets up to. And then, it's like we help foster those communities and those conversations and asking, does this work for you? If not, let's talk about it, let's figure out why it might not. And then, really working closely with community to kind of help bring those things forward and create action items. >> So, it's all about putting the right people in the rooms and not necessarily playing referee, but to get people in the right room to have and facilitate the conversation? >> Absolutely. Absolutely. Like all of the booths behind us could have their own conferences, but we want to bring everybody together to have those conversations. And again, sprawling can be really wild at certain times, but it's good to have those cross understandings, or to hear from somebody that you're like, "Oh, my goodness, I didn't even think about that kind of context or use case." So, really inclusive conversation. >> So, organizations like Boeing, Adobe, Microsoft, from an end user perspective, it's sometimes difficult to get those organizations into these types of communities. How do you encourage them to participate in the conversation 'cause their voice is extremely important? >> Yeah, that I'd also say it really is the community. I really liked the Kubernetes documentary that was put out, working with some of the CNCF folks and core, and beginning Kubernetes contributors and maintainers. And it just kind of blew me away when they had said, you know, what we thought was success, was seeing Kubernetes in an Amazon Data Center. That's when we knew that this was going to take root. And you'd rarely hear that, is like, "When somebody that we typically compete with, its success is seeing it, seeing them use that." And so, I thought was really cool. >> You know, I like to use this technology for my community of skipping rope. You see the girls and boys jumping double Dutch rope. And you think, "I can do that. Like it's just jumping." But there's this hesitation to actually, how do you start? How do you get inside of it? The question is how do you become a member of the community? We've talked a lot about what happens when you're in the community. But how do you join the community? >> So, really, there's a whole bunch of ways that you can. Actually, the shirt that I'm wearing, I got from the 114 Release. So, this is just a fun example of that community. And just kind of how welcoming and inviting that they are. Really, I do think it's kind of like a job breaker. Almost you start at the outside, you start using these technologies, even more generally like, what is DevOps? What is production? How do I get to infrastructure, architecture, or software engineering? Once you start there, you start working your way in, you develop a stack, and then you start to see these tools, technologies, workflows. And then, after you've kind of gotten a good amount of time spent with it, you might really enjoy it like that, and then want to help contribute like, "I like this, but it would be great to have a function that did this. Or I want a feature that does that." At that point in time, you can either take a look at the source code on GitHub, or wherever it's hosted, and then start to kind of come up with that, some ideas to contribute back to that. And then, beyond that, you can actually say, "No, I kind of want to have these conversations with people." Join in those special interest groups, and those meetings to kind of talk about things. And then, after a while, you can kind of find yourself in a contributor role, and then a maintainer role. After that, if you really like the project, and want to kind of work with community on that front. So, I think you had asked before, like Microsoft, Adobe and these others. Really it's about steering the projects. It's these communities want these things, and then, these companies say, "Okay, this is great. Let's join in the conversation with the community." And together again, inclusivity, and bringing everybody to the table to have that discussion and push things forward. >> So, Taylor, closing message. What would you want people watching this show to get when they think about ecosystem and CNCF? >> So, ecosystem it's a big place, come on in. Yeah, (laughs) the water's just fine. I really want people to take away the fact that... I think really when it comes down to, it really is the community, it's you. We are the end user ecosystem. We're the people that build the tools, and we need help. No matter how big or small, when you come in and join the community, you don't have to rewrite the Kubernetes scheduler. You can help make documentation that much more easy to understand, and in doing so, helping thousands of people, If I'm going through the instructions or reading a paragraph, doesn't make sense, that has such a profound impact. And I think a lot of people miss that. It's like, even just changing punctuation can have such a giant difference. >> Yeah, I think people sometimes forget that community, especially community-run projects, they need product managers. They need people that will help with communications, people that will help with messaging, websites updating. Just reachability, anywhere from developing code to developing documentation, there's ways to jump in and help the community. From Valencia, Spain, I'm Keith Townsend, and you're watching "theCUBE," the leader in high tech coverage. (bright upbeat music)
SUMMARY :
brought to you by Red Hat, and on all the sessions, and the user research group. and the number of contributing Those are the groups that So, let's start from the end user lens. and kind of give everyone the I'm going to quit my job and have to take in so the meat of what you do. Yeah, so that's the fun part. So, the CNCF loosely has So, the groups that are running OCI, Like all of the booths behind us participate in the conversation I really liked the Kubernetes become a member of the community? and those meetings to What would you want people it really is the community, it's you. and help the community.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Priyanka | PERSON | 0.99+ |
Boeing | ORGANIZATION | 0.99+ |
Adobe | ORGANIZATION | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Keith Townsend | PERSON | 0.99+ |
Cisco | ORGANIZATION | 0.99+ |
Cloud Native Computing Foundation | ORGANIZATION | 0.99+ |
10 weeks | QUANTITY | 0.99+ |
Taylor Dolezal | PERSON | 0.99+ |
Taylor | PERSON | 0.99+ |
TOC | ORGANIZATION | 0.99+ |
Stripe Press | ORGANIZATION | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
CNCF | ORGANIZATION | 0.99+ |
Mercedes | ORGANIZATION | 0.99+ |
Technical Oversight Committee | ORGANIZATION | 0.99+ |
Boeings | ORGANIZATION | 0.99+ |
Prometheus | TITLE | 0.99+ |
Coball | ORGANIZATION | 0.99+ |
Valencia, Spain | LOCATION | 0.99+ |
today | DATE | 0.99+ |
7.1 million members | QUANTITY | 0.99+ |
HashiCorp | ORGANIZATION | 0.98+ |
Kubecon | ORGANIZATION | 0.98+ |
Airbnbs | ORGANIZATION | 0.98+ |
VMware | ORGANIZATION | 0.98+ |
last night | DATE | 0.97+ |
GitHub | ORGANIZATION | 0.97+ |
Fortran | ORGANIZATION | 0.97+ |
first | QUANTITY | 0.96+ |
Kubernetes | TITLE | 0.95+ |
Working in Public | TITLE | 0.93+ |
Amazon Data Center | ORGANIZATION | 0.92+ |
Dutch | OTHER | 0.92+ |
thousands of people | QUANTITY | 0.91+ |
theCUBE | TITLE | 0.91+ |
more than one way | QUANTITY | 0.9+ |
Cloudnativecon | ORGANIZATION | 0.89+ |
theCUBE | ORGANIZATION | 0.86+ |
Kubernetes | ORGANIZATION | 0.84+ |
DevOps | TITLE | 0.84+ |
CNCF Ecosystem | ORGANIZATION | 0.83+ |
one thing | QUANTITY | 0.83+ |
three pillars | QUANTITY | 0.82+ |
Europe | LOCATION | 0.79+ |
Open Container Interface | OTHER | 0.77+ |
double | QUANTITY | 0.76+ |
OCI | OTHER | 0.73+ |
Cloudnativecon Europe | ORGANIZATION | 0.69+ |
Open Storage Interface | OTHER | 0.62+ |
2022 | DATE | 0.58+ |
CIG | ORGANIZATION | 0.53+ |
2022 | TITLE | 0.46+ |
114 Release | ORGANIZATION | 0.38+ |
Matt Provo & Patrick Bergstrom, StormForge | Kubecon + Cloudnativecon Europe 2022
>> Instructor: "theCUBE" presents KubeCon and CloudNativeCon Europe 2022, brought to you by Red Hat, the Cloud Native Computing Foundation and its ecosystem partners. >> Welcome to Valencia, Spain and we're at KubeCon, CloudNativeCon Europe 2022. I'm Keith Townsend, and my co-host, Enrico Signoretti. Enrico's really proud of me. I've called him Enrico instead of Enrique every session. >> Every day. >> Senior IT analyst at GigaOm. We're talking to fantastic builders at KubeCon, CloudNativeCon Europe 2022 about the projects and their efforts. Enrico, up to this point, it's been all about provisioning, insecurity, what conversation have we been missing? >> Well, I mean, I think that we passed the point of having the conversation of deployment, of provisioning. Everybody's very skilled, actually everything is done at day two. They are discovering that, well, there is a security problem. There is an observability problem a and in fact, we are meeting with a lot of people and there are a lot of conversation with people really needing to understand what is happening. I mean, in their cluster work, why it is happening and all the questions that come with it. And the more I talk with people in the show floor here or even in the various sessions is about, we are growing so that our clusters are becoming bigger and bigger, applications are becoming bigger as well. So we need to now understand better what is happening. As it's not only about cost, it's about everything at the end. >> So I think that's a great set up for our guests, Matt Provo, founder and CEO of StormForge and Patrick Brixton? >> Bergstrom. >> Bergstrom. >> Yeah. >> I spelled it right, I didn't say it right, Bergstrom, CTO. We're at KubeCon, CloudNativeCon where projects are discussed, built and StormForge, I've heard the pitch before, so forgive me. And I'm kind of torn. I have service mesh. What do I need more, like what problem is StormForge solving? >> You want to take it? >> Sure, absolutely. So it's interesting because, my background is in the enterprise, right? I was an executive at UnitedHealth Group before that I worked at Best Buy and one of the issues that we always had was, especially as you migrate to the cloud, it seems like the CPU dial or the memory dial is your reliability dial. So it's like, oh, I just turned that all the way to the right and everything's hunky-dory, right? But then we run into the issue like you and I were just talking about, where it gets very very expensive very quickly. And so my first conversations with Matt and the StormForge group, and they were telling me about the product and what we're dealing with. I said, that is the problem statement that I have always struggled with and I wish this existed 10 years ago when I was dealing with EC2 costs, right? And now with Kubernetes, it's the same thing. It's so easy to provision. So realistically what it is, is we take your raw telemetry data and we essentially monitor the performance of your application, and then we can tell you using our machine learning algorithms, the exact configuration that you should be using for your application to achieve the results that you're looking for without over-provisioning. So we reduce your consumption of CPU, of memory and production which ultimately nine times out of 10, actually I would say 10 out of 10, reduces your cost significantly without sacrificing reliability. >> So can your solution also help to optimize the application in the long run? Because, yes, of course-- >> Yep. >> The lowering fluid as you know optimize the deployment. >> Yeah. >> But actually the long-term is optimizing the application. >> Yes. >> Which is the real problem. >> Yep. >> So, we're fine with the former of what you just said, but we exist to do the latter. And so, we're squarely and completely focused at the application layer. As long as you can track or understand the metrics you care about for your application, we can optimize against it. We love that we don't know your application, we don't know what the SLA and SLO requirements are for your app, you do, and so, in our world it's about empowering the developer into the process, not automating them out of it and I think sometimes AI and machine learning sort of gets a bad rap from that standpoint. And so, at this point the company's been around since 2016, kind of from the very early days of Kubernetes, we've always been, squarely focused on Kubernetes, using our core machine learning engine to optimize metrics at the application layer that people care about and need to go after. And the truth of the matter is today and over time, setting a cluster up on Kubernetes has largely been solved. And yet the promise of Kubernetes around portability and flexibility, downstream when you operationalize, the complexity smacks you in the face and that's where StormForge comes in. And so we're a vertical, kind of vertically oriented solution, that's absolutely focused on solving that problem. >> Well, I don't want to play, actually. I want to play the devils advocate here and-- >> You wouldn't be a good analyst if you didn't. >> So the problem is when you talk with clients, users, there are many of them still working with Java, something that is really tough. I mean, all of us loved Java. >> Yeah, absolutely. >> Maybe 20 years ago. Yeah, but not anymore, but still they have developers, they have porting applications, microservices. Yes, but not very optimized, et cetera, cetera, et cetera. So it's becoming tough. So how you can interact with this kind of old hybrid or anyway, not well engineered applications. >> Yeah. >> We do that today. We actually, part of our platform is we offer performance testing in a lower environment and stage and we, like Matt was saying, we can use any metric that you care about and we can work with any configuration for that application. So perfect example is Java, you have to worry about your heap size, your garbage collection tuning and one of the things that really struck me very early on about the StormForge product is because it is true machine learning. You remove the human bias from that. So like a lot of what I did in the past, especially around SRE and performance tuning, we were only as good as our humans were because of what they knew. And so, we kind of got stuck in these paths of making the same configuration adjustments, making the same changes to the application, hoping for different results. But then when you apply machine learning capability to that the machine will recommend things you never would've dreamed of. And you get amazing results out of that. >> So both me and Enrico have been doing this for a long time. Like, I have battled to my last breath the argument when it's a bare metal or a VM, look, I cannot give you any more memory. >> Yeah. >> And the argument going all the way up to the CIO and the CIO basically saying, you know what, Keith you're cheap, my developer resources are expensive, buy bigger box. >> Yeah. >> Yap. >> Buying a bigger box in the cloud to your point is no longer a option because it's just expensive. >> Yeah. >> Talk to me about the carrot or the stick as developers are realizing that they have to be more responsible. Where's the culture change coming from? Is it the shift in responsibility? >> I think the center of the bullseye for us is within those sets of decisions, not in a static way, but in an ongoing way, especially as the development of applications becomes more and more rapid and the management of them. Our charge and our belief wholeheartedly is that you shouldn't have to choose. You should not have to choose between costs or performance. You should not have to choose where your applications live, in a public private or hybrid cloud environment. And so, we want to empower people to be able to sit in the middle of all of that chaos and for those trade offs and those difficult interactions to no longer be a thing. We're at a place now where we've done hundreds of deployments and never once have we met a developer who said, "I'm really excited to get out of bed and come to work every day and manually tune my application." One side, secondly, we've never met, a manager or someone with budget that said, please don't increase the value of my investment that I've made to lift and shift us over to the cloud or to Kubernetes or some combination of both. And so what we're seeing is the converging of these groups, their happy place is the lack of needing to be able to make those trade offs, and that's been exciting for us. >> So, I'm listening and looks like that your solution is right in the middle in application performance, management, observability. >> Yeah. >> And, monitoring. >> Yeah. >> So it's a little bit of all of this. >> Yeah, so we want to be, the intel inside of all of that, we often get lumped into one of those categories, it used to be APM a lot, we sometimes get, are you observability or and we're really not any of those things, in and of themselves, but we instead we've invested in deep integrations and partnerships with a lot of that tooling 'cause in a lot of ways, the tool chain is hardening in a cloud native and in Kubernetes world. And so, integrating in intelligently, staying focused and great at what we solve for, but then seamlessly partnering and not requiring switching for our users who have already invested likely, in a APM or observability. >> So to go a little bit deeper. What does it mean integration? I mean, do you provide data to this, other applications in the environment or are they supporting you in the work that you do. >> Yeah, we're a data consumer for the most part. In fact, one of our big taglines is take your observability and turn it into action ability, right? Like how do you take that, it's one thing to collect all of the data, but then how do you know what to do with it, right? So to Matt's point, we integrate with folks like Datadog, we integrate with Prometheus today. So we want to collect that telemetry data and then do something useful with it for you. >> But also we want Datadog customers, for example, we have a very close partnership with Datadog so that in your existing Datadog dashboard, now you have-- >> Yeah. >> The StormForge capability showing up in the same location. >> Yep. >> And so you don't have to switch out. >> So I was just going to ask, is it a push pull? What is the developer experience when you say you provide developer this resolve ML learnings about performance, how do they receive it? Like, what's the developer experience. >> They can receive it, for a while we were CLI only, like any good developer tool. >> Right. >> And, we have our own UI. And so it is a push in a lot of cases where I can come to one spot, I've got my applications and every time I'm going to release or plan for a release or I have released and I want to pull in observability data from a production standpoint, I can visualize all of that within the StormForge UI and platform, make decisions, we allow you to set your, kind of comfort level of automation that you're okay with. You can be completely set and forget or you can be somewhere along that spectrum and you can say, as long as it's within, these thresholds, go ahead and release the application or go ahead and apply the configuration. But we also allow you to experience the same, a lot of the same functionality right now, in Grafana, in Datadog and a bunch of others that are coming. >> So I've talked to Tim Crawford who talks to a lot of CIOs and he's saying one of the biggest challenges or if not, one of the biggest challenges CIOs are facing are resource constraints. >> Yeah. >> They cannot find the developers to begin with to get this feedback. How are you hoping to address this biggest pain point for CIOs-- >> Yeah.6 >> And developers? >> You should take that one. >> Yeah, absolutely. So like my background, like I said at UnitedHealth Group, right. It's not always just about cost savings. In fact, the way that I look about at some of these tech challenges, especially when we talk about scalability there's kind of three pillars that I consider, right? There's the tech scalability, how am I solving those challenges? There's the financial piece 'cause you can only throw money at a problem for so long and it's the same thing with the human piece. I can only find so many bodies and right now that pool is very small, and so, we are absolutely squarely in that footprint of we enable your team to focus on the things that they matter, not manual tuning like Matt said. And then there are other resource constraints that I think that a lot of folks don't talk about too. Like, you were talking about private cloud for instance and so having a physical data center, I've worked with physical data centers that companies I've worked for have owned where it is literally full, wall to wall. You can't rack any more servers in it, and so their biggest option is, well, I could spend $1.2 billion to build a new one if I wanted to, or if you had a capability to truly optimize your compute to what you needed and free up 30% of your capacity of that data center. So you can deploy additional name spaces into your cluster, like that's a huge opportunity. >> So I have another question. I mean, maybe it doesn't sound very intelligent at this point, but, so is it an ongoing process or is it something that you do at the very beginning, I mean you start deploying this. >> Yeah. >> And maybe as a service. >> Yep. >> Once in a year I say, okay, let's do it again and see if something change it. >> Sure. >> So one spot, one single.. >> Yeah, would you recommend somebody performance test just once a year? Like, so that's my thing is, at previous roles, my role was to do performance test every single release, and that was at a minimum once a week and if your thing did not get faster, you had to have an executive exception to get it into production and that's the space that we want to live in as well as part of your CICD process, like this should be continuous verification, every time you deploy, we want to make sure that we're recommending the perfect configuration for your application in the name space that you're deploying into. >> And I would be as bold as to say that we believe that we can be a part of adding, actually adding a step in the CICD process that's connected to optimization and that no application should be released, monitored, and sort of analyzed on an ongoing basis without optimization being a part of that. And again, not just from a cost perspective, but for cost and performance. >> Almost a couple of hundred vendors on this floor. You mentioned some of the big ones Datadog, et cetera, but what happens when one of the up and comings out of nowhere, completely new data structure, some imaginative way to click to telemetry data. >> Yeah. >> How do, how do you react to that? >> Yeah, to us it's zeros and ones. >> Yeah. >> And, we really are data agnostic from the standpoint of, we're fortunate enough from the design of our algorithm standpoint, it doesn't get caught up on data structure issues, as long as you can capture it and make it available through one of a series of inputs, one would be load or performance tests, could be telemetry, could be observability, if we have access to it. Honestly, the messier the better from time to time from a machine learning standpoint, it's pretty powerful to see. We've never had a deployment where we saved less than 30%, while also improving performance by at least 10%. But the typical results for us are 40 to 60% savings and 30 to 40% improvement in performance. >> And what happens if the application is, I mean, yes Kubernetes is the best thing of the world but sometimes we have to, external data sources or, we have to connect with external services anyway. >> Yeah. >> So, can you provide an indication also on this particular application, like, where the problem could be? >> Yeah. >> Yeah, and that's absolutely one of the things that we look at too, 'cause it's, especially when you talk about resource consumption it's never a flat line, right? Like depending on your application, depending on the workloads that you're running it varies from sometimes minute to minute, day to day, or it could be week to week even. And so, especially with some of the products that we have coming out with what we want to do, integrating heavily with the HPA and being able to handle some of those bumps and not necessarily bumps, but bursts and being able to do it in a way that's intelligent so that we can make sure that, like I said, it's the perfect configuration for the application regardless of the time of day that you're operating in or what your traffic patterns look like, or, what your disc looks like, right. Like 'cause with our low environment testing, any metric you throw at us, we can optimize for. >> So Matt and Patrick, thank you for stopping by. >> Yeah. >> Yes. >> We can go all day because day two is I think the biggest challenge right now, not just in Kubernetes but application re-platforming and transformation, very, very difficult. Most CTOs and EASs that I talked to, this is the challenge space. From Valencia, Spain, I'm Keith Townsend, along with my host Enrico Signoretti and you're watching "theCube" the leader in high-tech coverage. (whimsical music)
SUMMARY :
brought to you by Red Hat, and we're at KubeCon, about the projects and their efforts. And the more I talk with I've heard the pitch and then we can tell you know optimize the deployment. is optimizing the application. the complexity smacks you in the face I want to play the devils analyst if you didn't. So the problem is when So how you can interact and one of the things that last breath the argument and the CIO basically saying, Buying a bigger box in the cloud Is it the shift in responsibility? and the management of them. that your solution is right in the middle we sometimes get, are you observability or in the work that you do. consumer for the most part. showing up in the same location. What is the developer experience for a while we were CLI only, and release the application and he's saying one of the They cannot find the developers and it's the same thing or is it something that you do Once in a year I say, okay, and that's the space and that no application You mentioned some of the and 30 to 40% improvement in performance. Kubernetes is the best thing of the world so that we can make So Matt and Patrick, Most CTOs and EASs that I talked to,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Keith Townsend | PERSON | 0.99+ |
Enrico | PERSON | 0.99+ |
Enrico Signoretti | PERSON | 0.99+ |
Matt | PERSON | 0.99+ |
Jeff | PERSON | 0.99+ |
Tim Crawford | PERSON | 0.99+ |
Patrick | PERSON | 0.99+ |
2003 | DATE | 0.99+ |
Keith Townsend | PERSON | 0.99+ |
UnitedHealth Group | ORGANIZATION | 0.99+ |
40 | QUANTITY | 0.99+ |
Alex | PERSON | 0.99+ |
Jeff Frick | PERSON | 0.99+ |
Santa Clara | LOCATION | 0.99+ |
30 | QUANTITY | 0.99+ |
$1.2 billion | QUANTITY | 0.99+ |
Alex Wolf | PERSON | 0.99+ |
Enrique | PERSON | 0.99+ |
StormForge | ORGANIZATION | 0.99+ |
Alexander Wolf | PERSON | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
ACG | ORGANIZATION | 0.99+ |
January | DATE | 0.99+ |
Matt Provo | PERSON | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
Santa Cruz | LOCATION | 0.99+ |
Cloud Native Computing Foundation | ORGANIZATION | 0.99+ |
Patrick Bergstrom | PERSON | 0.99+ |
Best Buy | ORGANIZATION | 0.99+ |
30% | QUANTITY | 0.99+ |
first time | QUANTITY | 0.99+ |
Bergstrom | ORGANIZATION | 0.99+ |
nine times | QUANTITY | 0.99+ |
10 | QUANTITY | 0.99+ |
Valencia, Spain | LOCATION | 0.99+ |
300 people | QUANTITY | 0.99+ |
millions | QUANTITY | 0.99+ |
Datadog | ORGANIZATION | 0.99+ |
Java | TITLE | 0.99+ |
GigaOm | ORGANIZATION | 0.99+ |
Baskin School of Engineering | ORGANIZATION | 0.99+ |
two things | QUANTITY | 0.99+ |
third year | QUANTITY | 0.99+ |
Mountain View, California | LOCATION | 0.99+ |
KubeCon | EVENT | 0.99+ |
ACGSV | ORGANIZATION | 0.99+ |
both | QUANTITY | 0.99+ |
once a week | QUANTITY | 0.99+ |
less than 30% | QUANTITY | 0.99+ |
ACGSV GROW! Awards | EVENT | 0.98+ |
2016 | DATE | 0.98+ |
one | QUANTITY | 0.98+ |
Kubernetes | TITLE | 0.98+ |
40% | QUANTITY | 0.98+ |
Santa Cruz UC Santa Cruz School of Engineering | ORGANIZATION | 0.98+ |
today | DATE | 0.98+ |
ACG Silicon Valley | ORGANIZATION | 0.98+ |
60% | QUANTITY | 0.98+ |
once a year | QUANTITY | 0.98+ |
one spot | QUANTITY | 0.98+ |
10 years ago | DATE | 0.97+ |
Patrick Brixton | PERSON | 0.97+ |
Prometheus | TITLE | 0.97+ |
20 years ago | DATE | 0.97+ |
CloudNativeCon Europe 2022 | EVENT | 0.97+ |
secondly | QUANTITY | 0.97+ |
one single | QUANTITY | 0.96+ |
first conversations | QUANTITY | 0.96+ |
millions of dollars | QUANTITY | 0.96+ |
ACGSV GROW! Awards 2018 | EVENT | 0.96+ |
Michael Cade, Veeam | VeeamON 2022
(calm music) >> Hi everybody. We're here at VeeamON 2022. This is day two of the CUBE's continuous coverage. I'm Dave Vellante. My co-host is Dave Nicholson. A ton of energy. The keynotes, day two keynotes are all about products at Veeam. Veeam, the color of green, same color as money. And so, and it flows in this ecosystem. I'll tell you right now, Michael Cade is here. He's the senior technologist for product strategy at Veeam. Michael, fresh off the keynotes. >> Yeah, yeah. >> Welcome. Danny Allen's keynote was fantastic. I mean, that story he told blew me away. I can't wait to have him back. Stay tuned for that one. But we're going to talk about protecting containers, Kasten. You guys got announcements of Kasten by Veeam, you call it K10 version five, I think? >> Yeah. So just rolled into 5.0 release this week. Now, it's a bit different to what we see from a VBR release cycle kind of thing, cause we're constantly working on a two week sprint cycle. So as much as 5.0's been launched and announced, we're going to see that trickling out over the next couple of months until we get round to Cube (indistinct) and we do all of this again, right? >> So let's back up. I first bumped into Kasten, gosh, it was several years ago at VeeamON. Like, wow this is a really interesting company. I had deep conversations with them. They had a sheer, sheer cat grin, like something was going on and okay finally you acquire them, but go back a little bit of history. Like why the need for this? Containers used to be ephemeral. You know, you didn't have to persist them. That changed, but you guys are way ahead of that trend. Talk a little bit more about the history there and then we'll get into current day. >> Yeah, I think the need for stateful workloads within Kubernetes is absolutely grown. I think we just saw 1.24 of Kubernetes get released last week or a couple of weeks ago now. And really the focus there, you can see, at least three of the big ticket items in that release are focused around storage and data. So it just encourages that the community is wanting to put these data services within that. But it's also common, right? It's great to think about a stateless... If you've got stateless application but even a web server's got some state, right? There's always going to be some data associated to an application. And if there isn't then like, great but that doesn't really work- >> You're right. Where'd they click, where'd they go? I mean little things like that, right? >> Yeah. Yeah, exactly. So one of the things that we are seeing from that is like obviously the requirement to back up and put in a lot of data services in there, and taking full like exposure of the Kubernetes ecosystem, HA, and very tiny containers versus these large like virtual machines that we've always had the story at Veeam around the portability and being able to move them left, right, here, there, and everywhere. But from a K10 point of view, the ability to not only protect them, but also move those applications or move that data wherever they need to be. >> Okay. So, and Kubernetes of course has evolved. I mean the early days of Kubernetes, they kept it simple, kind of like Veeam actually. Right? >> Yeah. >> And then, you know, even though Mesosphere and even Docker Swarm, they were trying to do more sophisticated cluster management. Kubernetes has now got projects getting much more complicated. So more complicated workloads mean more data, more critical data means more protection. Okay, so you acquire Kasten, we know that's a small part of your business today but it's going to be growing. We know this cause everybody's developing applications. So what's different about protecting containers? Danny talks about modern data protection. Okay, when I first heard that, I'm like, eh, nice tagline, but then he peel the onion. He explains how in virtualization, you went from agents to backing up of VMware instance, a virtual instance. What's different about containers? What constitutes modern data protection for containers? >> Yeah, so I think the story that Danny tells as well, is so when we had our physical agents and virtualization came along and a lot of... And this is really where Veeam was born, right, we went into the virtualization API, the VMware API, and we started leveraging that to be more storage efficient. The admin overhead around those agents weren't there then, we could just back up using the API. Whereas obviously a lot of our competition would use agents still and put that resource overhead on top of that. So that's where Veeam initially got the kickstart in that world. I think it's very similar to when it comes to Kubernetes because K10 is deployed within the Kubernetes cluster and it leverages the Kubernetes API to pull out that data in a more efficient way. You could use image based backups or traditional NAS based backups to protect some of the data, and backup's kind of the... It's only one of the ticks in the boxes, right? You have to be able to restore and know what that data is. >> But wait, your competitors aren't as fat, dumb and happy today as they were back then, right? So it can't... They use the same APIs and- >> Yeah. >> So what makes you guys different? >> So I think that's testament to the Kubernetes and the community behind that and things like the CSI driver, which enables the storage vendors to take that CSI abstraction layer and then integrate their storage components, their snapshot technologies, and other efficiency models in there, and be able to leverage that as part of a universal data protection API. So really that's one tick in the box and you're absolutely right, there's open source tools that can do exactly what we're doing to a degree on that backup and recovery. Where it gets really interesting is the mobility of data and how we're protecting that. Because as much as stateful workloads are seen within the Kubernetes environments now, they're also seen outside. So things like Amazon RDS, but the front end lives in Kubernetes going to that stateless point. But being able to protect the whole application and being very application aware means that we can capture everything and restore wherever we want that to go as well. Like, so the demo that I just did was actually a Postgres database in AWS, and us being able to clone or migrate that out into an EKS cluster as a staple set. So again, we're not leveraging RDS at that point, but it gives us the freedom of movement of that data. >> Yeah, I want to talk about that, what you actually demoed. One of the interesting things, we were talking earlier, I didn't see any CLI when you were going through the integration of K10 V5 and V12. >> Yeah. >> That was very interesting, but I'm more skeptical of this concept, of the single pane of glass and how useful that is. Who is this integration targeting? Are you targeting the sort of traditional Veeam user who is now adding as a responsibility, the management of protecting these Kubernetes environments? Or are you at the same time targeting the current owners of those environments? Cause I know you talk about shift left and- >> Yeah. >> You know, nobody needs Kubernetes if you only have one container and one thing you're doing. So at some point it's all about automation, it's about blueprints, it's about getting those things in early. So you get up, you talk about this integration, who cares about that kind of integration? >> Yeah, so I think it's a bit of both, right? So we're definitely focused around the DevOps focused engineer. Let's just call it that. And under an umbrella, the cloud engineer that's looking after Kubernetes, from an application delivery perspective. But I think more and more as we get further up the mountain, CIS admin, obviously who we speak to the tech decision makers, the solutions architects systems engineers, they're going to inherit and be that platform operator around the Kubernetes clusters. And they're probably going to land with the requirement around data management as well. So the specific VBR centralized management is very much for the backup admin, the infrastructure admin or the cloud based engineer that's looking after the Kubernetes cluster and the data within that. Still we speak to app developers who are conscious of what their database looks like, because that's an external data service. And the biggest question that we have or the biggest conversation we have with them is that the source code, the GitHub or the source repository, that's fine, that will get your... That'll get some of the way back up and running, but when it comes to a Postgres database or some sort of data service, oh, that's out of the CI/CD pipeline. So it's whether they're interested in that or whether that gets farmed out into another pre-operations, the traditional operations team. >> So I want to unpack your press release a little bit. It's full of all the acronyms, so maybe you can help us- >> Sure. >> Cipher. You got security everywhere enhance platform hardening, including KMS. That's key- >> Yeah, key management service, yeah. >> System, okay. With AWS, KMS and HashiCorp vault. Awesome, love to see HashiCorp company. >> Yeah. >> RBAC objects in UI dashboards, ransomware attacks, AWS S3. So anyway, security everywhere. What do you mean by that? >> So I think traditionally at Veeam, and continue that, right? From a security perspective, if you think about the failure scenario and ransomware's, the hot topic, right, when it comes to security, but we can think about security as, if we think about that as the bang, right, the bang is something bad's happen, fire, flood, blood, type stuff. And we tend to be that right hand side of that, we tend to be the remediation. We're definitely the one, the last line of defense to get stuff back when something really bad happens. And I think what we've done from a K10 point of view, is not only enhance that, so with the likes of being able to... We're not going to reinvent the wheel, let's use the services that HashiCorp have done from a HashiCorp vault point of view and integrate from a key management system. But then also things like S3 or ransomware prevention. So I want to know if something bad's happened and Kasten actually did something more generic from a Veeam ONE perspective, but one of the pieces that we've seen since we've then started to send our backups to an immutable object storage, is let's be more of that left as well and start looking at the preventative tasks that we can help with. Now, we're not going to be a security company, but you heard all the way through Danny's like keynote, and probably when he is been on here, is that it's always, we're always mindful of that security focus. >> On that point, what was being looked for? A spike in CPU utilization that would be associated with encryption? >> Yeah, exactly that. >> Is that what was being looked- >> That could be... Yeah, exactly that. So that could be from a virtual machine point of view but from a K10, and it specifically is that we're going to look at the S3 bucket or the object storage, we're going to see if there's a rate of change that's out of the normal. It's an abnormal rate. And then with that, we can say, okay, that doesn't look right, alert us through observability tools, again, around the cloud native ecosystem, Prometheus Grafana. And then we're going to get insight into that before the bang happens, hopefully before the bang. >> So that's an interesting when we talk about adjacencies and moving into this area of security- >> We're talking to Zeus about that too. >> Exactly. That's that sort of creep where you can actually add value. It's interesting. >> So, okay. So we talked about shift left, get that, and then expanded ecosystem, industry leading technologies. By the way, one of them is the Red Hat Marketplace. And I think, I heard Anton's... Anton was amazing. He is the head of product management at Veeam. Is been to every VeeamON. He's got family in Ukraine. He's based in Switzerland. >> Yeah. >> But he chose not to come here because he's obviously supporting, you know, the carnage that's going on in Ukraine. But anyway, I think he said the Red Hat team is actually in Ukraine developing, you know, while the bombs are dropping. That's amazing. But anyway, back to our interview here, expanded ecosystem, Red Hat, SUSE with Rancher, they've got some momentum. vSphere with Tanzu, they're in the game. Talk about that ecosystem and its importance. >> Yeah, and I think, and it goes back to your point around the CLI, right? Is that it feels like the next stage of Kubernetes is going to be very much focused towards the operator or the operations team. The CIS admin of today is going to have to look after that. And at the moment it's all very command line, it's all CLI driven. And I think the marketplace is OpenShift, being our biggest foothold around our customer base, is definitely around OpenShift. But things like, obviously we are a longstanding alliance partner with VMware as well. So their Tanzu operations actually there's support for TKGS, so vSphere Tanzu grid services is another part of the big release of 5.0. But all three of those and the common marketplace gives us a UI, gives us a way of being able to see and visualize that rather than having to go and hunt down the commands and get our information through some- >> Oh, some people are going to be unhappy about that. >> Yeah. >> But I contend the human eye has evolved to see in color for a very good reason. So I want to see things in red, yellow, and green at times. >> There you go, yeah. >> So when we hear a company like Veeam talk about, look we have no platform agenda, we don't care which cloud it's in. We don't care if it's on-prem or Google Azure, AWS. We had Wasabi on, we have... Great, they got an S3 compatible, you know, target, and others as well. When we hear them, companies like you, talk about that consistent experience, single pane of glass that you're skeptical of, maybe cause it's technically challenging, one of the things, we call it super cloud, right, that's come up. Danny and I were riffing on that the other day and we'll do that more this afternoon. But it brings up something that we were talking about with Zeus, Dave, which is the edge, right? And it seems like Kubernetes, and we think about OpenShift. >> Yeah. >> We were there last week at Red Hat Summit. It's like 50% of the conversation, if not more, was the edge. Right, and really true edge, worst cases, use cases. Two weeks ago we were at Dell Tech, there was a lot of edge talk, but it was retail stores, like Lowe's. Okay, that's kind of near edge, but the far edge, we're talking space, right? So seems like Kubernetes fits there and OpenShift, you know, particularly, as well as some of the others that we mentioned. What about edge? How much of what you're doing with container data protection do you see as informing you about the edge opportunity? Are you seeing any patterns there? Nobody's really talking about it in data protection yet. >> So yeah, large scale numbers of these very small clusters that are out there on farms or in wind turbines, and that is definitely something that is being spoken about. There's not much mention actually in this 5.0 release because we actually support things like K3s,(indistinct), that all came in 4.5, but I think, to your first point as well, David, is that, look, we don't really care what that Kubernetes distribution is. So you've got K3s lightweight Kubernetes distribution, we support it, because it uses the same native Kubernetes APIs, and we get deployed inside of that. I think where we've got these large scale and large numbers of edge deployments of Kubernetes and that you require potentially some data management down there, and they might want to send everything into a centralized location or a more centralized location than a farm shed out in the country. I think we're going to see a big number of that. But then we also have our multi cluster dashboard that gives us the ability to centralize all of the control plane. So we don't have to go into each individual K10 deployment to manage those policies. We can have one big centralized management multi cluster dashboard, and we can set global policies there. So if you're running a database and maybe it's the same one across all of your different edge locations, where you could just set one policy to say I want to protect that data on an hourly basis, a daily basis, whatever that needs to be, rather than having to go into each individual one. >> And then send it back to that central repository. So that's the model that you see, you don't see the opportunity, at least at this point in time, of actually persisting it at the edge? >> So I think it depends. I think we see both, but again, that's the footprint. And maybe like you mentioned about up in space having a Kubernetes cluster up there. You don't really want to be sending up a NAS device or a storage device, right, to have to sit alongside it. So it's probably, but then equally, what's the art of the possible to get that back down to our planet, like as part of a consistent copy of data? >> Or even a farm or other remote locations. The question is, I mean, EVs, you know, we believe there's going to be tons of data, we just don't.. You think about Tesla as a use case, they don't persist a ton of their data. Maybe if a deer runs across, you know, the front of the car, oh, persist that, send that back to the cloud. >> I don't want anyone knowing my Tesla data. I'll tell you that right now. (all laughing) >> Well, there you go, that one too. All right, well, that's future discussion, we're still trying to squint through those patterns. I got so many questions for you, Michael, but we got to go. Thanks so much for coming to theCUBE. >> Always. >> Great job on the keynote today and good luck. >> Thank you. Thanks for having me. >> All right, keep it right there. We got a ton of product talk today. As I said, Danny Allan's coming back, we got the ecosystem coming, a bunch of the cloud providers. We have, well, iland was up on stage. They were just recently acquired by 11:11 Systems. They were an example today of a cloud service provider. We're going to unpack it all here on theCUBE at VeeamON 2022 from Las Vegas at the Aria. Keep it right there. (calm music)
SUMMARY :
Veeam, the color of green, I mean, that story he told blew me away. and we do all of this again, right? about the history there So it just encourages that the community I mean little things like that, right? So one of the things that I mean the early days of Kubernetes, but it's going to be growing. and it leverages the Kubernetes API So it can't... and be able to leverage that One of the interesting things, of the single pane of glass So you get up, you talk And the biggest question that we have It's full of all the acronyms, You got security everywhere With AWS, KMS and HashiCorp vault. So anyway, security everywhere. and ransomware's, the hot topic, right, or the object storage, That's that sort of creep where He is the head of product said the Red Hat team and the common marketplace gives us a UI, to be unhappy about that. But I contend the human eye on that the other day It's like 50% of the and maybe it's the same one So that's the model that you see, but again, that's the footprint. that back to the cloud. I'll tell you that right now. Thanks so much for coming to theCUBE. on the keynote today and good luck. Thanks for having me. a bunch of the cloud providers.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Nicholson | PERSON | 0.99+ |
David | PERSON | 0.99+ |
Michael | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Danny Allen | PERSON | 0.99+ |
Switzerland | LOCATION | 0.99+ |
Ukraine | LOCATION | 0.99+ |
Danny | PERSON | 0.99+ |
Michael Cade | PERSON | 0.99+ |
Tesla | ORGANIZATION | 0.99+ |
50% | QUANTITY | 0.99+ |
Las Vegas | LOCATION | 0.99+ |
Lowe | ORGANIZATION | 0.99+ |
Anton | PERSON | 0.99+ |
VeeamON | ORGANIZATION | 0.99+ |
Veeam | ORGANIZATION | 0.99+ |
last week | DATE | 0.99+ |
Dave | PERSON | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
Two weeks ago | DATE | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
last week | DATE | 0.99+ |
two week | QUANTITY | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
Veeam | PERSON | 0.99+ |
11:11 Systems | ORGANIZATION | 0.99+ |
Danny Allan | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
both | QUANTITY | 0.98+ |
SUSE | ORGANIZATION | 0.98+ |
one | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
one policy | QUANTITY | 0.98+ |
first point | QUANTITY | 0.98+ |
Rancher | ORGANIZATION | 0.98+ |
K10 | COMMERCIAL_ITEM | 0.98+ |
this week | DATE | 0.98+ |
S3 | TITLE | 0.98+ |
one container | QUANTITY | 0.98+ |
several years ago | DATE | 0.97+ |
Kubernetes | TITLE | 0.97+ |
CIS | ORGANIZATION | 0.97+ |
KMS | TITLE | 0.96+ |
Dell Tech | ORGANIZATION | 0.96+ |
Zeus | ORGANIZATION | 0.96+ |
K10 V5 | COMMERCIAL_ITEM | 0.95+ |
OpenShift | TITLE | 0.95+ |
VMware | TITLE | 0.95+ |
first | QUANTITY | 0.95+ |
this afternoon | DATE | 0.95+ |
V12 | COMMERCIAL_ITEM | 0.94+ |
iland | ORGANIZATION | 0.94+ |
GitHub | ORGANIZATION | 0.94+ |
One | QUANTITY | 0.94+ |
TKGS | ORGANIZATION | 0.93+ |
S3 | COMMERCIAL_ITEM | 0.92+ |
Red Hat Summit | EVENT | 0.92+ |
day two | QUANTITY | 0.92+ |
Tanzu | ORGANIZATION | 0.92+ |
Matt Provo & Patrick Bergstrom, StormForge | Kubecon + Cloudnativecon Europe 2022
>>The cube presents, Coon and cloud native con Europe 22, brought to you by the cloud native computing foundation. >>Welcome to Melissa Spain. And we're at cuon cloud native con Europe, 2022. I'm Keith Townsend. And my co-host en Rico senior Etti en Rico's really proud of me. I've called him en Rico and said IK, every session, senior it analyst giga, O we're talking to fantastic builders at Cuban cloud native con about the projects and the efforts en Rico up to this point, it's been all about provisioning insecurity. What, what conversation have we been missing? >>Well, I mean, I, I think, I think that, uh, uh, we passed the point of having the conversation of deployment of provisioning. You know, everybody's very skilled, actually everything is done at day two. They are discovering that, well, there is a security problem. There is an observability problem. And in fact, we are meeting with a lot of people and there are a lot of conversation with people really needing to understand what is happening. I mean, in their classroom, what, why it is happening and all the, the questions that come with it. I mean, and, uh, the more I talk with, uh, people in the, in the show floor here, or even in the, you know, in the various sessions is about, you know, we are growing, the, our clusters are becoming bigger and bigger. Uh, applications are becoming, you know, bigger as well. So we need to know, understand better what is happening. It's not only, you know, about cost it's about everything at the >>End. So I think that's a great set up for our guests, max, Provo, founder, and CEO of storm for forge and Patrick Britton, Bergstrom, Brookstone. Yeah, I spelled it right. I didn't say it right. Berg storm CTO. We're at Q con cloud native con we're projects are discussed, built and storm forge. I I've heard the pitch before, so forgive me. And I'm, I'm, I'm, I'm, I'm, I'm kind of torn. I have service mesh. What do I need more like, what problem is storm for solving? >>You wanna take it? >>Sure, absolutely. So it it's interesting because, uh, my background is in the enterprise, right? I was an executive at United health group. Um, before that I worked at best buy. Um, and one of the issues that we always had was, especially as you migrate to the cloud, it seems like the CPU dial or the memory dial is your reliability dial. So it's like, oh, I just turned that all the way to the right and everything's hunky Dory. Right. Uh, but then we run into the issue like you and I were just talking about where it gets very, very expensive, very quickly. Uh, and so my first conversations with Matt and the storm forge group, and they were telling me about the product and, and what we're dealing with. I said, that is the problem statement that I have always struggled with. And I wish this existed 10 years ago when I was dealing with EC two costs, right? And now with Kubernetes, it's the same thing. It's so easy to provision. So realistically, what it is is we take your raw telemetry data and we essentially monitor the performance of your application. And then we can tell you using our machine learning algorithms, the exact configuration that you should be using for your application to achieve the results that you're looking for without over provisioning. So we reduce your consumption of CPU of memory and production, which ultimately nine times outta 10, actually I would say 10 out of 10 reduces your cost significantly without sacrificing reliability. >>So can your solution also help to optimize the application in the long run? Because yes, of course, yep. You know, the lowing fluid is, you know, optimize the deployment. Yeah. But actually the long term is optimizing the application. Yes. Which is the real problem. >>Yep. So we actually, um, we're fine with the, the former of what you just said, but we exist to do the latter. And so we're squarely and completely focused at the application layer. Um, we are, uh, as long as you can track or understand the metrics you care about for your application, uh, we can optimize against it. Um, we love that we don't know your application. We don't know what the SLA and SLO requirements are for your app. You do. And so in, in our world, it's about empowering the developer into the process, not automating them out of it. And I think sometimes AI and machine learning sort of gets a bad wrap from that standpoint. And so, uh, we've at this point, the company's been around, you know, since 2016, uh, kind of from the very early days of Kubernetes, we've always been, you know, squarely focused on Kubernetes using our core machine learning, uh, engine to optimize metrics at the application layer, uh, that people care about and, and need to need to go after. And the truth of the matter is today. And over time, you know, setting a cluster up on Kubernetes has largely been solved. Um, and yet the promise of, of Kubernetes around portability and flexibility, uh, downstream when you operationalize the complexity, smacks you in the face. And, uh, and that's where, where storm forge comes in. And so we're a vertical, you know, kind of vertically oriented solution. Um, that's, that's absolutely focused on solving that problem. >>Well, I don't want to play, actually. I want to play the, uh, devils advocate here and, you know, >>You wouldn't be a good analyst if you didn't. >>So the, the problem is when you talk with clients, users, they, there are many of them still working with Java with, you know, something that is really tough. Mm-hmm <affirmative>, I mean, we loved all of us loved Java. Yeah, absolutely. Maybe 20 years ago. Yeah. But not anymore, but still they have developers. They are porting applications, microservices. Yes. But not very optimized, etcetera. C cetera. So it's becoming tough. So how you can interact with these kind of yeah. Old hybrid or anyway, not well in generic applications. >>Yeah. We, we do that today. We actually, part of our platform is we offer performance testing in a lower environment and stage. And we like Matt was saying, we can use any metric that you care about and we can work with any configuration for that application. So the perfect example is Java, you know, you have to worry about your heap size, your garbage collection tuning. Um, and one of the things that really struck, struck me very early on about the storm forage product is because it is true machine learning. You remove the human bias from that. So like a lot of what I did in the past, especially around SRE and, and performance tuning, we were only as good as our humans were because of what they knew. And so we were, we kind of got stuck in these paths of making the same configuration adjustments, making the same changes to the application, hoping for different results. But then when you apply machine learning capability to that, the machine will recommend things you never would've dreamed of. And you get amazing results out of >>That. So both me and an Rico have been doing this for a long time. Like I have battled to my last breath, the, the argument when it's a bare metal or a VM. Yeah. Look, I cannot give you any more memory. Yeah. And the, the argument going all the way up to the CIO and the CIO basically saying, you know what, Keith you're cheap, my developer resources expensive, my bigger box. Yep. Uh, buying a bigger box in the cloud to your point is no longer a option because it's just expensive. Talk to me about the carrot or the stick as developers are realizing that they have to be more responsible. Where's the culture change coming from? So is it, that is that if it, is it the shift in responsibility? >>I think the center of the bullseye for us is within those sets of decisions, not in a static way, but in an ongoing way, especially, um, especially as the development of applications becomes more and more rapid. And the management of them, our, our charge and our belief wholeheartedly is that you shouldn't have to choose, you should not have to choose between costs or performance. You should not have to choose where your, you know, your applications live, uh, in a public private or, or hybrid cloud environment. And so we want to empower people to be able to sit in the middle of all of that chaos and for those trade-offs and those difficult interactions to no, no longer be a thing. You know, we're at, we're at a place now where we've done, you know, hundreds of deployments and never once have we met a developer who said, I'm really excited to get outta bed and come to work every day and manually tune my application. <laugh> One side, secondly, we've never met, uh, you know, uh, a manager or someone with budget that said, uh, please don't, you know, increase the value of my investment that I've made to lift and shift us over mm-hmm <affirmative>, you know, to the cloud or to Kubernetes or, or some combination of both. And so what we're seeing is the converging of these groups, um, at, you know, their happy place is the lack of needing to be able to, uh, make those trade offs. And that's been exciting for us. So, >>You know, I'm listening and looks like that your solution is right in the middle in application per performance management, observability. Yeah. And, uh, and monitoring. So it's a little bit of all of this. >>So we, we, we, we want to be, you know, the Intel inside of all of that, mm-hmm, <affirmative>, we don't, you know, we often get lumped into one of those categories. It used to be APM a lot. We sometimes get a, are you observability or, and we're really not any of those things in and of themselves, but we, instead of invested in deep integrations and partnerships with a lot of those, uh, with a lot of that tooling, cuz in a lot of ways, the, the tool chain is hardening, uh, in a cloud native and, and Kubernetes world. And so, you know, integrating in intelligently staying focused and great at what we solve for, but then seamlessly partnering and not requiring switching for, for our users who have already invested likely in a APM or observability. >>So to go a little bit deeper. Sure. What does it mean integration? I mean, do you provide data to this, you know, other applications in, in the environment or are they supporting you in the work that you >>Yeah, we're, we're a data consumer for the most part. Um, in fact, one of our big taglines is take your observability and turn it into actionability, right? Like how do you take the it's one thing to collect all of the data, but then how do you know what to do with it? Right. So to Matt's point, um, we integrate with folks like Datadog. Um, we integrate with Prometheus today. So we want to collect that telemetry data and then do something useful with it for you. >>But, but also we want Datadog customers. For example, we have a very close partnership with, with Datadog, so that in your existing data dog dashboard, now you have yeah. This, the storm for capability showing up in the same location. Yep. And so you don't have to switch out. >>So I was just gonna ask, is it a push pull? What is the developer experience? When you say you provide developer, this resolve ML, uh, learnings about performance mm-hmm <affirmative> how do they receive it? Like what, yeah, what's the, what's the, what's the developer experience >>They can receive it. So we have our own, we used to for a while we were CLI only like any good developer tool. Right. Uh, and you know, we have our own UI. And so it is a push in that, in, in a lot of cases where I can come to one spot, um, I've got my applications and every time I'm going to release or plan for a release or I have released, and I want to take, pull in, uh, observability data from a production standpoint, I can visualize all of that within the storm for UI and platform, make decisions. We allow you to, to set your, you know, kind of comfort level of automation that you're, you're okay with. You can be completely set and forget, or you can be somewhere along that spectrum. And you can say, as long as it's within, you know, these thresholds, go ahead and release the application or go ahead and apply the configuration. Um, but we also allow you to experience, uh, the same, a lot of the same functionality right now, you know, in Grafana in Datadog, uh, and a bunch of others that are coming. >>So I've talked to Tim Crawford who talks to a lot of CIOs and he's saying one of the biggest challenges, or if not, one of the biggest challenges CIOs are facing are resource constraints. Yeah. They cannot find the developers to begin with to get this feedback. How are you hoping to address this biggest pain point for CIOs? Yeah. >>Development? >>Just take that one. Yeah, absolutely. That's um, so like my background, like I said, at United health group, right. It's not always just about cost savings. In fact, um, the way that I look about at some of these tech challenges, especially when we talk about scalability, there's kind of three pillars that I consider, right? There's the tech scalability, how am I solving those challenges? There's the financial piece, cuz you can only throw money at a problem for so long. And it's the same thing with the human piece. I can only find so many bodies and right now that pool is very small. And so we are absolutely squarely in that footprint of, we enable your team to focus on the things that they matter, not manual tuning like Matt said. And then there are other resource constraints that I think that a lot of folks don't talk about too. >>Like we were, you were talking about private cloud for instance. And so having a physical data center, um, I've worked with physical data centers that companies I've worked for have owned where it is literally full wall to wall. You can't rack any more servers in it. And so their biggest option is, well, I could spend 1.2 billion to build a new one if I wanted to. Or if you had a capability to truly optimize your compute to what you needed and free up 30% of your capacity of that data center. So you can deploy additional name spaces into your cluster. Like that's a huge opportunity. >>So either out of question, I mean, may, maybe it, it doesn't sound very intelligent at this point, but so is it an ongoing process or is it something that you do at the very beginning mean you start deploying this. Yeah. And maybe as a service. Yep. Once in a year I say, okay, let's do it again and see if something changes. Sure. So one spot 1, 1, 1 single, you know? >>Yeah. Um, would you recommend somebody performance tests just once a year? >>Like, so that's my thing is, uh, previous at previous roles I had, uh, my role was you performance test, every single release. And that was at a minimum once a week. And if your thing did not get faster, you had to have an executive exception to get it into production. And that's the space that we wanna live in as well as part of your C I C D process. Like this should be continuous verification every time you deploy, we wanna make sure that we're recommending the perfect configuration for your application in the name space that you're deploying >>Into. And I would be as bold as to say that we believe that we can be a part of adding, actually adding a step in the C I C D process that's connected to optimization and that no application should be released monitored and sort of, uh, analyzed on an ongoing basis without optimization being a part of that. And again, not just from a cost perspective, yeah. Cost end performance, >>Almost a couple of hundred vendors on this floor. You know, you mentioned some of the big ones, data, dog, et cetera. But what happens when one of the up and comings out of nowhere, completely new data structure, some imaginable way to click to elementry data. Yeah. How do, how do you react to that? >>Yeah. To us it's zeros and ones. Yeah. Uh, and you know, we're, we're, we're really, we really are data agnostic from the standpoint of, um, we're not, we we're fortunate enough to, from the design of our algorithm standpoint, it doesn't get caught up on data structure issues. Um, you know, as long as you can capture it and make it available, uh, through, you know, one of a series of inputs, what one, one would be load or performance tests, uh, could be telemetry, could be observability if we have access to it. Um, honestly the messier, the, the better from time to time, uh, from a machine learning standpoint, um, it, it, it's pretty powerful to see we've, we've never had a deployment where we, uh, where we saved less than 30% while also improving performance by at least 10%. But the typical results for us are 40 to 60% savings and, you know, 30 to 40% improvement in performance. >>And what happens if the application is, I, I mean, yes, Kubernetes is the best thing of the world, but sometimes we have to, you know, external data sources or, or, you know, we have to connect with external services anyway. Mm-hmm <affirmative> yeah. So can you, you know, uh, can you provide an indication also on, on, on this particular application, like, you know, where the problem could >>Be? Yeah, yeah. And that, that's absolutely one of the things that we look at too, cuz it's um, especially when you talk about resource consumption, it's never a flat line, right? Like depending on your application, depending on the workloads that you're running, um, it varies from sometimes minute to minute, day to day, or it could be week to week even. Um, and so especially with some of the products that we have coming out with what we want to do, you know, partnering with, uh, you know, integrating heavily with the HPA and being able to handle some of those bumps and not necessarily bumps, but bursts and being able to do it in a way that's intelligent so that we can make sure that, like I said, it's the perfect configuration for the application regardless of the time of day that you're operating in or what your traffic patterns look like. Um, or you know, what your disc looks like, right? Like cuz with our, our low environment testing, any metric you throw at us, we can, we can optimize for. >>So Madden Patrick, thank you for stopping by. Yeah. Yes. We can go all day. Because day two is I think the biggest challenge right now. Yeah. Not just in Kubernetes, but application replatforming and re and transformation. Very, very difficult. Most CTOs and S that I talked to, this is the challenge space from Valencia Spain. I'm Keith Townsend, along with my host en Rico senior. And you're watching the queue, the leader in high tech coverage.
SUMMARY :
brought to you by the cloud native computing foundation. And we're at cuon cloud native you know, in the various sessions is about, you know, we are growing, I I've heard the pitch before, and one of the issues that we always had was, especially as you migrate to the cloud, You know, the lowing fluid is, you know, optimize the deployment. And so we're a vertical, you know, devils advocate here and, you know, So the, the problem is when you talk with clients, users, So the perfect example is Java, you know, you have to worry about your heap size, And the, the argument going all the way up to the CIO and the CIO basically saying, you know what, that I've made to lift and shift us over mm-hmm <affirmative>, you know, to the cloud or to Kubernetes or, You know, I'm listening and looks like that your solution is right in the middle in all of that, mm-hmm, <affirmative>, we don't, you know, we often get lumped into one of those categories. this, you know, other applications in, in the environment or are they supporting Like how do you take the it's one thing to collect all of the data, And so you don't have to switch out. Um, but we also allow you to experience, How are you hoping to address this And it's the same thing with the human piece. Like we were, you were talking about private cloud for instance. is it something that you do at the very beginning mean you start deploying this. And that's the space that we wanna live in as well as part of your C I C D process. actually adding a step in the C I C D process that's connected to optimization and that no application You know, you mentioned some of the big ones, data, dog, Um, you know, as long as you can capture it and make it available, or, you know, we have to connect with external services anyway. we want to do, you know, partnering with, uh, you know, integrating heavily with the HPA and being able to handle some So Madden Patrick, thank you for stopping by.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Tim Crawford | PERSON | 0.99+ |
Keith Townsend | PERSON | 0.99+ |
30 | QUANTITY | 0.99+ |
40 | QUANTITY | 0.99+ |
1.2 billion | QUANTITY | 0.99+ |
Matt | PERSON | 0.99+ |
Matt Provo | PERSON | 0.99+ |
Datadog | ORGANIZATION | 0.99+ |
storm for forge | ORGANIZATION | 0.99+ |
Patrick Bergstrom | PERSON | 0.99+ |
2016 | DATE | 0.99+ |
Java | TITLE | 0.99+ |
10 | QUANTITY | 0.99+ |
Melissa Spain | PERSON | 0.99+ |
nine times | QUANTITY | 0.99+ |
Valencia Spain | LOCATION | 0.99+ |
40% | QUANTITY | 0.99+ |
less than 30% | QUANTITY | 0.99+ |
10 years ago | DATE | 0.98+ |
United health group | ORGANIZATION | 0.98+ |
both | QUANTITY | 0.98+ |
20 years ago | DATE | 0.98+ |
one | QUANTITY | 0.98+ |
Keith | PERSON | 0.98+ |
once a year | QUANTITY | 0.98+ |
once a week | QUANTITY | 0.98+ |
HPA | ORGANIZATION | 0.98+ |
2022 | DATE | 0.98+ |
Coon | ORGANIZATION | 0.98+ |
30% | QUANTITY | 0.98+ |
first conversations | QUANTITY | 0.97+ |
Cloudnativecon | ORGANIZATION | 0.97+ |
60% | QUANTITY | 0.97+ |
Kubernetes | TITLE | 0.97+ |
Etti | PERSON | 0.97+ |
today | DATE | 0.96+ |
Patrick Britton | PERSON | 0.96+ |
Kubecon | ORGANIZATION | 0.96+ |
StormForge | ORGANIZATION | 0.95+ |
data dog | ORGANIZATION | 0.94+ |
Prometheus | TITLE | 0.94+ |
three pillars | QUANTITY | 0.94+ |
secondly | QUANTITY | 0.94+ |
Rico | ORGANIZATION | 0.93+ |
Q con cloud | ORGANIZATION | 0.93+ |
hundreds of deployments | QUANTITY | 0.92+ |
day two | QUANTITY | 0.92+ |
Europe | LOCATION | 0.92+ |
Kubernetes | ORGANIZATION | 0.92+ |
Intel | ORGANIZATION | 0.92+ |
one spot | QUANTITY | 0.89+ |
at least 10% | QUANTITY | 0.87+ |
one thing | QUANTITY | 0.85+ |
hundred vendors | QUANTITY | 0.83+ |
Once in a year | QUANTITY | 0.83+ |
cuon cloud native con | ORGANIZATION | 0.81+ |
Rico | LOCATION | 0.81+ |
Brookstone | ORGANIZATION | 0.8+ |
Grafana | ORGANIZATION | 0.8+ |
Berg storm CTO | ORGANIZATION | 0.8+ |
SRE | TITLE | 0.79+ |
SLA | TITLE | 0.79+ |
Bergstrom | ORGANIZATION | 0.79+ |
cloud native con | ORGANIZATION | 0.78+ |
single release | QUANTITY | 0.77+ |
storm forge group | ORGANIZATION | 0.75+ |
1 | QUANTITY | 0.75+ |
One side | QUANTITY | 0.74+ |
EC two | TITLE | 0.74+ |
1 single | QUANTITY | 0.74+ |
Patrick | PERSON | 0.74+ |
Ian Massingham, MongoDB | AWS Summit SF 2022
>>Okay, welcome back everyone. Cube's coverage here. Live on the floor at AWS summit, 2022, an in person event in San Francisco. Of course, AWS summit, 2022 in New York city is coming up this summer. The cube will be there as well. Make sure you check us out then too, but we day two of coverage had a great guest here. I Han VP of developer relations, Mongo DB, formally of AWS. We've been known each other for a long time doing, uh, developer relations at Mongo DB. Welcome to the queue. Good to see >>You. Thank to be here. Thanks for inviting me, John. It's great >>To, so Mongo DB is, um, first of all, stocks' doing really well right now. Businesswise is good, but I still think it's undervalue. A lot of people think is, is a lot more going huge success with Atlas. So congratulations to the team over there. Um, what's the update? What's the relationship withs, you know, guys have been great partners for years. What's the new thing. Yeah. >>So MongoDB Atlas obviously runs on several different major cloud providers, but AWS is the largest partner that we work with in the public cloud. So the majority of our Atlas workloads for our customers are running on the AWS platform. And just earlier this year, we announced a new strategic collaboration agreement with AWS. That's gonna further strengthen and deepen that partnership that we have with them. >>What's the main product value right now on the scale on, on Atlas, what's the drive in the revenue momentum. >>So, I mean, you know, there's a huge trend in the industry towards cloud managed databases, right? You look back 10, 15 years ago when we first met, most customers were only and operating their own data infrastructure, either running it in their own data centers, or maybe if they were really early using the primitives that cloud providers like AWS offered to run their databases in the cloud when Amazon launched RDS back in 2009, I think it was, we started to see this trend towards cloud managed databases. We followed that with our own Atlas offering back in 2016. And as Andy jazzy from AWS would say very often it's offloading that UND differentiated, heavy lifting, allowing developers to focus on building applications. They don't have to win and operate the data infrastructure. We do it for them, and that has proven incredibly popular amongst our customers. You know, Atlas route right now is growing at 50, sorry, 85% car year on year growth. >>You know, um, I've been following MongoDB for a long, long time. I mean, going back to the lamp stack days, you know, and you think about what Mongo has done as a product because of the developer traction, you know, Mongo can't do this, just keeps getting better every year. And, and the, I think the stickiness with developers is a real big part of that. Can you your view there cuz you're in VE relations. I mean, developers all love Mongo. They're teaching in school. People are picking up a side hustles, they're coding on it, using it all everywhere. I mean it's well known. >>There's a few different reasons for that. I think the main one is the, the document orientated model that we use, the document data models that are used by Mongo DB, just a net way for developers to work with data. And then, uh, we've invested in creating 16 first party drivers that allow developers using various different programming languages, whether that's JavaScript or Python or rust to integrate MongoDB, natively and idiomatic with their software. So it's very, very easy for a developer to pick up MongoDB, grab one of these drivers from their package manager of their choice and then build applications that natively manipulate data inside MongoDB, whether that's MongoDB Atlas or our enterprise edition on their own premises. They get a very consistent and very easy to, I easy to use developer experience with our, with our platform. >>Talk about the go to market with AWS. You guys also have a tightly coupled relationships. There's been announcements there recently. Uh, what's changing most right now that people should pay attention to. Well, >>The first thing is there's a huge amount of technical integration between MongoDB and AWS services. And that's the basis for many of our customers choosing to run Mon Mongo DB on AWS. We're active in 23 AWS regions around the world. And there's many other integration points as well, like cryptographic protection of Mongo MongoDB, stored data using Amazon cryptographic services, for example, or building serverless applications with AWS Lambda and MongoDB servers. So there's a ton of technical integration. Yeah, but what we started to work on now is go to market integration with AWS as well. So you can buy Mongo DB Atlas through AWS's marketplace. You can use the payer, you go offering to pay for it with your AWS bill. And then we're collaborating with AWS on migrations and other joint go to market activities as well. That >>Means get incentives, the sales people at AWS. >>Of course our moreover I mean, it's just really easy for customers, really easy for developers to consume. Yeah, they don't need to contract with MongoDB. They can use their existing AWS contracting, their existing discounting relationships and pre purchasing arrangements with AWS to consume Atlas. >>It's the classic meet the customers where they >>Are exactly right. Meet the developer where they are and meet the customers where they are now with this new model as well. >>Yeah. I love marketplace. I think it's been great. You know, even with its kind of catalog and vibe, I think it's gonna get better and better, uh, over there teams doing good work. Um, and it's easy to consume. That's key. >>Yeah. Super easy. Reduce that friction and make it real easy for developers to adopt this. Right. >>Talk about some of the top customers that you guys share with AWS. What are some of the customers you guys have together and what the benefits of the >>Relationship joint references that we talk about? A lot, one of them is Shutterfly. So in the photographic products area, they built a eCommerce offering with MongoDB and AWS. The second is seven 11 with seven 11. We're doing a lot in the mobile space. So edge applications, we've got a feature in MongoDB Atlas that allows you to synchronize data with databases on mobile devices. Those can be phones point of sale devices or handheld devices that might be used in the parcel industry, for example. So seven 11 using us in that way. And then lastly with Pitney Bowes, we've got a big digital transformation project with Pitney Bowes where they've reimagined their, uh, postage and packaging services, delivering those to their customers, using MongoDB as a data store as well. >>I wanna get in some of the trends, you've got a great per you know, you know, Mongo from Amazon side and now you're there. Um, Mongo's, as you pointed out has, has been around for a long time. What are some of the stats? I mean, how many customers, how many countries? Well, it's pretty massive >>Mind. We've got almost quarter of a billion downloads today, 240 million MongoDB downloads since we launched the first product <laugh>, we've got 33,000 active customers that are using MongoDB Atlas today and we're running well over a million free tier clusters on MongoDB Atlas across all of the different providers where we operate the service as well. So these numbers are, you know, mind blowing in terms of scale. Uh, but of course at the core of that is operational excellence. Customers love Mongo DBS because they don't have to operate it themselves. They don't have to deal with fairly conditions. They don't have to deal with scaling. They don't have to deal with deployment. We all, we do all of those things as part of the service offering and customers get an endpoint that they can use with their applications to store and retrieve data reliably. And with consistently high perform, >>You know, it's, you know, in the media, something has to be dead. Someone's the death of the iPhone, the death of this, nothing that really dies. Mongo DB has always been kind of like talked about, well, it doesn't scale on the high end. Of course, Oracle was saying that, I mean, all the, all the big database vendors were kind of throwing darts at, at Mongo, uh, DB, uh, but it kept scaling. Atlas is a whole nother. Could you just unpack that a little bit more? Why is it so important? Because scale is just, I mean, it's, it's horizontal, but it's also performant. >>Exactly. Right. So with, uh, Mongo DB's document access model that I've described already, you break some of the limitations that exist inside traditional relational databases. So, you know, they don't scale well, if you've got high concurrent and see of data access, and they're typically difficult and expensive to scale because you need to share data. Once you grow beyond individual cluster nodes, and you'll know that all relational databases suffer from these same kinds of issues with non relational systems, no SQL systems like MongoDB, you have to think a little bit more about design at the beginning. So designing database to cater for the different access patterns that you have, but in return for that upfront preparation, that design work, you get near limitless, scalability and performance will scale nearly linearly with that scalability as well. So very much more high performance, very much more simplicity for the developer as their database gets larger and their cluster gets larger to support it. >>Yeah. You know, Amazon web service has always had an a and D jazz. We talk to us all the time, every interview I've done with Swami and Matt wood or whoever on the team and executive levels always said the same thing. There's not one database to rule the world, right? Obvious you're talking about Oracle, but even within AWS customers, they're mixing and matching databases based on use cases. So in distributed environment, they're all working together. So, um, you guys fit nicely into that. So how does that, >>I think strategy slightly counterbalances that so, you know, they would say use the specific tool for the specific task that you have in hand. Yeah. What we try to focus on is creating the simple and most effective developer experience that we can, and then supporting different facets to the product in order to allow developers to different use cases. A really good example with something like MongoDB Atlas search. So we integrated Apache Luine into MongoDB Atlas. Customers can very simply apply Apache Luine search indexes to the data that they've got in MongoDB. And then they can interact with that search data using the same drivers as an API. Yeah, yeah. That they use for regular queries. So if you want to run search on your application data, you don't need a separate open search or elastic search cluster, just turn on MongoDB Atlas search and use that, that search facet. So it's interest and we have other capabilities that it's >>Vertically integrating inside within Mongo, >>Correct? Yes. That's better. Yeah. With the guy, all of creating a really simple and effective developer experience, boosting developer productivity and helping developers get more done in less time. >>You mentioned serverless earlier, what's the serverless angle with AWS when Mongo, >>Is there one? Yeah. So we have MongoDB serverless currently in preview, uh, has the same kind of characteristics that you would, or the characteristics that you would expect from a serverless data base. So consumption based model, you provision an endpoint and that will scale elastically in accordance with your usage and you get billed by consumption units so much like the serverless paradigm that we've seen delivered by AWS, the same kind of model for Mongo, DB, Atlas serverless. >>What, what attracted you to Mongo DBS? So you knew them before, or you moved over there. Um, what's going on there? What's the culture like right now? Oh, >>The culture's great. I mean, it's a much smaller company than AWS where I was before, you know, it's a very large organization. And one of the things that I really like about MongoDB is, as I've said earlier, we can serve the different use cases that a developer might have with a single product, with different aspects, to it, different facets to it. Uh, and it's a really great conversation to have with a, with a developer, with a developer customer, to be able to offer one thing that helps them solve five or six different problems that have traditionally been quite hard for them to wrestle with quite difficult for them to, to deal with. And then we've got this focus on developer experience through these driver packages that we have as well. So it's really great to have as a developer relations pro have that kind of tooling in my kit bag that can help developers become more effective. >>Talk about tooling, cuz you know, I always have, uh, kind of moments where I waffle between more. I love platforms, tools are being over overused, too many tools tool with the tool, you know, the expressions, but we're seeing from developers, the ones that don't want to go into the hood, we serverless plays beautifully. Yep. They want tools. They do. And, and the, the new engineering developers that are coming outta college and universities, they love tools. >>Yeah. And we actually have quite a few of those built into Mongo, DB Atlas. So inside Mongo, DB Atlas, we've got things like an index optimizer, which will suggest the best way that you might index your data for better perform months inside MongoDB, running on Atlas, we've got a data Explorer, which is much like another product that we've got called MongoDB compass that allows you to see and manipulate the data that you have stored within your database natively within the Atlas interface. Uh, and then we also have, uh, whole slew of different metrics, monitoring capabilities built into the platform as well. So these are aspects of Atlas that developers can take advantage of. And then over on the client side, visual studio code plugins. Yeah. So you can manipulate and operate with data directly inside visual studio code, which is obviously the most common and popular IDE out there today, as well as integration with things like infrastructure is code tools. So we support cloud formation for provisioning. We have CDK constructs inside. Yeah. The CDK construct library. We also have a lot of customers using Terraform to provision MongoDB across both AWS and other providers. So having that developer tooling of course is super important. Yeah. Aspect of the developer experience, trying to >>Build out deploying observability is a big one. How does that fit in? Cuz you knew need to talk and not only measure everything here, but talk to other systems. >>Yeah. So we recently announced a provider for Prometheus and Grafana. So we can emit metrics into those providers. Obviously CNCF projects, very common and popular inside customers that are running on Kubernetes. We've got a Kubernetes operator for MongoDB Atlas as well. Good. So you can provision MongoDB Atlas from within Kubernetes as well as having our own native metrics directly within Atlas as well. >>Ian you're crushing it. You got all the, the data, the fingertips. Are you gonna be at Cuban this year? Uh, >>I will be, but some of our team members will definitely be there. >>Yeah, we'll be at, uh, EU. The cube will be there. Great. Thanks for coming on. Appreciate the insight final world. I'll give you the last word. Tell the audience what's going on. What's at Mongo DB. What should they pay attention to? If they've used Mongo and are aware of it? What's the update. What's >>The so you should come to MongoDB world actually in New York at the beginning of June, June 7th, the ninth in the Javit center in New York. Gonna have our own show there. And of course we'd love to see you there. >>Okay. Cube comes here day two of eight, us summit, 2020, this Cub I'm John for your host. Stay with us more. Our coverage as day two winds down. Great coverage.
SUMMARY :
Make sure you check Thanks for inviting me, John. So congratulations to the team over there. That's gonna further strengthen and deepen that partnership that we have with them. So, I mean, you know, there's a huge trend in the industry towards cloud managed databases, right? I think the stickiness with developers is a real big part of that. or Python or rust to integrate MongoDB, natively and idiomatic with their software. Talk about the go to market with AWS. And that's the basis for many of our customers choosing to run Mon Mongo DB on AWS. Yeah, they don't need to contract with MongoDB. Meet the developer where they are and meet the customers where they are now with this new model as well. You know, even with its kind of catalog and vibe, Reduce that friction and make it real easy for developers to adopt this. Talk about some of the top customers that you guys share with AWS. Atlas that allows you to synchronize data with databases on mobile devices. Um, Mongo's, as you pointed out has, has been around for a long time. part of the service offering and customers get an endpoint that they can use with their applications to store and You know, it's, you know, in the media, something has to be dead. cater for the different access patterns that you have, but in return for that upfront preparation, So, um, you guys fit nicely into that. the specific task that you have in hand. boosting developer productivity and helping developers get more done in less time. that you would, or the characteristics that you would expect from a serverless data base. So you knew them before, or you moved over Uh, and it's a really great conversation to have with a, Talk about tooling, cuz you know, I always have, uh, kind of moments where I waffle between more. So you can manipulate and operate with data directly inside visual studio code, Cuz you knew need to talk and not only measure everything So you can provision MongoDB Are you gonna be at Cuban this year? I'll give you the last word. And of course we'd love to see you there. Stay with us more.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
AWS | ORGANIZATION | 0.99+ |
five | QUANTITY | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
2016 | DATE | 0.99+ |
John | PERSON | 0.99+ |
Ian Massingham | PERSON | 0.99+ |
Mongo | ORGANIZATION | 0.99+ |
2009 | DATE | 0.99+ |
San Francisco | LOCATION | 0.99+ |
Andy jazzy | PERSON | 0.99+ |
New York | LOCATION | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
Ian | PERSON | 0.99+ |
85% | QUANTITY | 0.99+ |
Atlas | ORGANIZATION | 0.99+ |
240 million | QUANTITY | 0.99+ |
Atlas | TITLE | 0.99+ |
33,000 active customers | QUANTITY | 0.99+ |
iPhone | COMMERCIAL_ITEM | 0.99+ |
Python | TITLE | 0.99+ |
MongoDB | TITLE | 0.99+ |
50 | QUANTITY | 0.99+ |
MongoDB Atlas | TITLE | 0.99+ |
MongoDB Atlas | TITLE | 0.99+ |
first | QUANTITY | 0.99+ |
2020 | DATE | 0.99+ |
one | QUANTITY | 0.99+ |
JavaScript | TITLE | 0.99+ |
today | DATE | 0.99+ |
23 | QUANTITY | 0.98+ |
Mongo DB | ORGANIZATION | 0.98+ |
first product | QUANTITY | 0.98+ |
second | QUANTITY | 0.98+ |
Swami | PERSON | 0.98+ |
Grafana | ORGANIZATION | 0.98+ |
eight | QUANTITY | 0.98+ |
10, 15 years ago | DATE | 0.97+ |
Kubernetes | TITLE | 0.97+ |
Matt wood | PERSON | 0.97+ |
this year | DATE | 0.97+ |
single product | QUANTITY | 0.97+ |
six different problems | QUANTITY | 0.96+ |
Morgan McLean & Danielle Greshock | AWS Partner Showcase S1E2
(gentle music) >> Hello, welcome to theCUBE's presentation of the AWS Showcase season one, episode two with the ISV Startups partners. I'm John Furrier, your host of theCUBE. We're joined by Morgan McLean, director of product management at Splunk, and Danielle Greshock, who is the director of ISVs solution architects at AWS. Welcome to the show. Thanks for coming on. >> Thanks for having us. >> And great. Thanks for having us. >> Great to see both of you, both theCUBE alumni, but the Splunk-AWS relationship has been going very, very well. You guys are doing great business enabling this app revolution. And cloud scale has been going extremely well. So let's get into it. You guys are involved in a lot of action around application revolution, around OpenTelemetry and open source. So let's get into it. What's the latest? >> Danielle, you go ahead. >> Well, I'll just jump in first. Obviously last year, not last year, but in 2020, we launched the AWS Distro for OpenTelemetry. The idea being essentially, we're able to bring in data from partners, from infrastructure running on AWS, from apps running on AWS, to really be able to increase observability across all cloud assets at your entire cloud platform. So, Morgan, if you want to chime in on how Splunk >> Morgan: Certainly. >> has worked out OpenTelemetry. >> Yeah. I mean, OpenTelemetry is super exciting. Obviously, there's a lot of partnership points between Amazon and Splunk, but OpenTelemetry is probably one of them that's the most visible to people who aren't already maybe using these two products together. And so, as Danielle mentioned, Amazon has their own distribution of OpenTelemetry, Splunk has their own, as well, and of course there's the main open source distribution that everybody knows and loves. Just for our viewers, just for clarity's sake, the separate distributions are fundamentally very similar to, almost identical to what's offered in the open source space, but they come preconfigured and they come with support guarantees from each company, meaning that you can actually get paid full support for an open source project, which is really fantastic for customers. And as Danielle mentioned, it's a great demonstration of the alliance between Splunk and Amazon Web Services. For example, the AWS Distro, when you use it, can export data to Amazon CloudWatch, various Amazon backed open source initiatives like Prometheus and others, and to Splunk Observability Cloud and to Splunk Enterprise. So it's a place that we've worked very closely together, and it's something that we're very excited about. >> So, Morgan, I want to get your take on the on the product management side and also how product are built these days. >> One of the big things we're seeing in cloud is that open source has been the big enabler for a lot of refactoring. And you got multiple distributions, but the innovations on top of that, can you talk about how you see the productization of new innovations with open source as you guys go into this market, because this is the new dynamic with cloud. We're seeing examples all over the place. Obviously, Amazon's going next level with what they're doing, and that open source, it's not a one game for all of it. You can have mix and match. Take us through the product angle. >> And in many ways, this is just another wave of the same thing, right? Like, if you think back in time, we all used and still use in many cases, virtual machines, most of those are based on Linux, right? Another large open source project. And so, open source software has been accelerating innovation in the cloud space and in the computing space generally for a very long time, which is fantastic. Our excitement with something like OpenTelemetry comes from both the project's capabilities but also what we can do with it. So for those who aren't already familiar with OpenTelemetry, OpenTelemetry allows you to extract really critical system telemetry, application signals and everything else you need from your own applications, from new services, from your infrastructure, from everything that you're running in a cloud environment. You can then send that data to another location for processing. And so John, you ask like, how does this accelerate innovation? What does it unlock? Well, the insight you can gain from this data means you can become so much more efficient as a development organization. You can make your applications so much more effective because when you send that data to something like Splunk Observability Cloud, to something like Amazon CloudWatch, to various other solutions on the market, they can give you deep, deep insight into your application's performance, to its structure, they can help you reduce outages. And so, it's very, very powerful because it allows organizations to use tools like Splunk, like Amazon, like other things to innovate so much more effectively. >> Danielle, can you comment >> If I could... >> on the AWS side because this is again on the big point. You guys are going next level, and you're starting to see patterns in the ISV world, certainly on the architecture side of partners doing things differently now on top of what they've already done. Could you share how AWS is helping customers accelerate? >> Well, just as Morgan was talking about what OpenTelemetry provides, you can see how from a partnership perspective, this is so valuable, right? What the partner team here at AWS is in the business of doing, is really enabling customer choice, right? And having that ability to plug in and pull data from different sources, post it to different sources, make it available for visibility across all of your resources is very powerful and it's something from the partner community that we really value because we want customers to be able to select best of breed solutions, what works for their business, which businesses are different and they may have different needs, and that also fosters that true innovation. A small company is going to develop and release software a lot differently than a large enterprise. And so, being able to support something like OpenTelemetry just enables that for all different kinds of customers. >> Morgan, add to that because the velocity of releases, certainly operational, stability, is key every predominant security, uptime, these are top concerns. And, you mention data too, >> And you mention challenges. >> You got the data in here. So you got a lot of data moving around, a lot of value. What's your take? >> Yeah. So, I'll speak with some specifics. So a challenge that developers have had for years when you're developing large services, which you can now do with platforms like AWS. So, it's very easy to go develop huge deployments. But a challenge they have is you go and build a mess, right? And like, I've worked earlier in my career in Web Services. And I remember in one of the first orgs I was in, I was one of the five people who really understood our ecommerce stack. Right? And so like, I would get dragged into all these meetings and I'd have to go draw like the 50 services we had, and how they interacted, and the changes that were made in the last week. And without observability tools like Splunk Observability Cloud, like the ones offered by Amazon, like the ones that are backed by the data that comes with OpenTelemetry, organizations basically rely on people like this, to go draw out their deployments so they understand what it is they've built. Well, as you can imagine, this crimps your development velocity, because most of your engineers, most of your tech leads, most of everyone else don't actually understand what it is they've built what it is they're running, because they need that global context. You get something like OpenTelemetry and the solutions that consume the data from it, and suddenly now, all your developers have that context, all of them when they're adding functionality to a service or they're updating their infrastructure, can actually understand how it interacts with the rest of the broader application. This lets you speed up your time to development, this lets you ship more safely, more securely. And finally, when things do go wrong, which will be less frequent, but when they do go wrong, you can fix them super rapidly. >> If I'm a customer, let me ask a question. I'm a customer and I say, "Okay, I love AWS, I love Splunk, I love OpenTelemetry. I got to have open sources, technology innovation is happening." What's the integration? What are some of the standards? Can you take us through how that's working together with you guys as a shared platform? >> Yeah. So let's take the Amazon distribution for OpenTelemetry or even the Splunk one. One of the first things they do is they include all of the receivers, all of the sort of data capture components that you need, out of box for platforms like AWS, right? And so, right away, you get that power and flexibility where you're getting access to all of these data sources, right? And so, that's part of that partnership. And additionally, once the data comes into OpenTelemetry, you can now send that to various different data sources, including, as Danielle mentioned, to multiple at the same time. So you can use whatever tools you want. And so when you talk about like what the partnership is actually providing to you as a customer and still, this is just within the context of OpenTelemetry, obviously there's a much broader partnership between these two companies than just that. But within the context of OpenTelemetry means you can download one of these distributions. It's fully supported. It works with both solutions and everything is just great, right? You don't need to go fiddle with that out of the box. To be clear, OpenTelemetry is a batteries included project, right? This means that even the standard distributions of OpenTelemetry include the components you need. You have to go directly, reference them and ensure that they're packaged in there, but they exist, right. But the nice thing about these distributions is that it's done, it's out of the box, you don't even have to worry about is something missing or do I need to include new exporters or new receivers? It's all there. It's preconfigured. It just works. And if something goes wrong and you have a support contact, you pick up the phone, you talk to someone to get it fixed. >> Danielle, what's the Amazon side 'cause agility and scale is one of the highlights you guys are seeing. How does this tie into that and how are you guys working backwards from the customers to support the partners? >> Well, I think just to add on essentially to what Morgan said, I think that AWS is a cloud platform, has always really had a focus on developers. And, we talk a lot about how AWS and Amazon as a whole really embraces this continuous integration and continuous deployment methods inside of our organization. And we talk about services, and observability is a huge part of that. The only way that you're actually able to release hundreds, thousands of times a day like Amazon does, is by having an observability platform, to be able to measure metrics, see changes in the environment, to be able to roll back if you need to, and to be able to quickly mitigate any challenges or anything that goes wrong at any part of the process. And so, when we preach that to our customers, I think it's something that we do that because we live it and breathe it. And so, things such as OpenTelemetry and such as the products that Splunk builds, those are also ways in which we believe our customers can achieve that. >> Yeah. And we can... I mean, as I mentioned before, this partnership goes well beyond OpenTelemetry, right? And so, if you go use like Splunk Enterprise, Enterprise Cloud, Splunk Observability Cloud, and you're running on AWS, you have excellent support and excellent visibility into your Amazon infrastructure, into the services and applications you've deployed on top of that infrastructure. We try and give you, and I think we do succeed in this. We give you the best possible experience, the deepest possible visibility, into what it is you've deployed on AWS, so that you can be even more successful as a business, and so that you can be even more successful on AWS as a platform. >> Yeah. This is a great conversation, Morgan. You mentioned the early days of Web Services. AWS stands for Amazon Web Services built on web services. So interesting throwback there, but made me think about the days of the early days of web services. And if you look at data, what's going on now, the top partners in AWS, you're seeing a lot of people thinking about data differently, they're refactoring, a lot of machine learning, a lot of AI going on at scale. So then, you got cloud native, things like Kubernetes and these new services being stood up and teared down with automation. A whole new operating model's coming. And so when you think about observability, the importance of it, I mean, can you share your perspective on this whole 'nother level? I mean, I always say that whole another level sounds cliche, but it is next level. I mean, this is completely different. What's your reaction? >> Yeah. There there's a ton of factors here, right? So as you point out, companies are totally shifting how they use their cloud infrastructure. And part of this you see during their cloud migrations, a part of it you see after, and they're shifting from their sort of stateful VMs that they may have had in the past to infrastructure that they tear down and put up regularly. And there's a lot more automation. With this, comes as I mentioned before, complexity, right? And also, with this comes more and more businesses becoming even more reliant on their digital infrastructure. And so, not having observability into your applications, into your services, into your infrastructure, to me, is akin to running a business, say running a large warehousing or distribution company, but not having any idea where you're shipping products or where things are, or not having any accounting or CFO, right? Like, business has become so digital. Business is so reliant on technology, and that's unlocked a ton of new things. It's great. But not having visibility into how that technology works or what it is that's deployed or how to fix it is akin to having no visibility to anything else in your business. It's nuts. And so, observability is super, super critical, particularly for customers who are adopting this new wave of cloud technologies on platforms like AWS. >> Danielle, on your side too, you're enabling this new capability so that businesses can do it, the partners do it, we're calling it super cloud. We've been calling it super cloud kind of dynamic where new things are happening with the data. And you guys are evolving with that. Can you share what you're seeing on your side as your partners start to go to the next level? What are you guys doing? How does it all come together? >> Well, we always talk about what has happened with data in the last couple of years, which the cloud has really enabled around, you know, variety and velocity and there's one other "V" that's escaping me right now, but essentially, all of this data is coming in and providing the ability for us to make better decisions, to build better products, to provide better experiences for customers. And so, I just think, the OpenTelemetry project, as well as what Splunk is doing is just another example of how we're taking this massive amount of data and being able to provide better experiences and outcomes for customers. >> And you guys have been working along together for long time, Splunk, and, it's been a great partners, if we're going back with that been covering it on theCUBE and SiliconANGLE. So, we know that, the change is key observability. Can you imagine a company without a CFO, Morgan? That's just boggles your mind, but that's what it's like right now. So... >> It is, yeah. >> And the people who take advantage of that are winning, right? So it's like, that's the key. >> Yeah, I know. I mean, even in my own career, right, I've moved between different companies. And I remember, when I joined Google in particular, which is where I worked at previously, I was very impressed with their internal observability tools. And I'm certain, I haven't worked at Amazon. I'm certainly, I just assume inside of Amazon they're excellent as well, so a lot of the large cloud firms these days. But it was so refreshing going from an organization where if we had some outage or something went wrong, there were like a very small set of people who could actually understand what was going on. And then you would just have to manually dive through logs and correlate requests manually between services. It's very challenging. And so, when things went wrong, they went wrong for a long, long time. And so, the companies that understood this even in the past are already very successful as a result. I think now, the rest of the industry is really in the midst of adopting these observability practices and the tools that are required to implement them, because you're right. Otherwise your development velocity slows down. Now you're getting out competed by your competition. And then, when you have a problem, it blows up for ages. And once again, your competition can take advantage of it. >> And, can you just summarize the observability piece relative to the OpenTelemetry? Where is that going to go? Where do you see that evolving? >> Sure. >> I see open source is growing like crazy, we all know that. >> Of course. >> But OpenTelemetry in particular and open source, 'cause this is a big hot area. >> Yes. So to set the stage for people, OpenTelemetry, unlocks observability in many ways. As I mentioned earlier, OpenTelemetry is how you capture data out of your application. It doesn't process it. It's not a replacement for something like Amazon CloudWatch or any Splunk's products, but it's how we get the data out of your system, which is a remarkably difficult problem. I won't dive into it today, but, those who work in this space are very aware. That's why this project exists and it's so big, that actually extracting information, metrics, logs, distributed traces, profiles, everything else, from your applications and from your infrastructure is very, very difficult. So for OpenTelemetry, where it's going is just continually getting better at extracting more types of data from more sources, and doing that more effectively for people in a more standardized way. That will unlock firms like Splunk, firms, like Amazon and others to better process this data. In terms of where that's going, the sky's the limit, right? Like, everyone's familiar with APM, people are familiar with infrastructure monitoring, but there's a lot more capabilities coming there for security analytics, for network performance monitoring, for getting down all the way to single lines of coding your application, how they impact everything. There's just so much power that's coming to the industry right now. I'm really excited to see where things go in the next few years. >> And Danielle, you're in the middle of all the action as a solution architect, really set the stage for their companies and the ISVs, and this is a big, hot area. What are the patterns you're seeing and what are some of the best practices that you're doing will help companies? >> Right. So I think, summarizing our entire conversation, the big things that we're seeing in the market is essentially more and more companies are looking to move to a continuous deployment and a continuous integration environment. And they're looking to innovate faster and spend less time hot patching or hot fixing their environments and they want to spend more time innovating. And so, that you know, the patterns that we're seeing is... What I see and what I actually experience firsthand at re:Invent when I talk to probably over 40 or 50 ISVs, is customers want to know in their environment, where are their changes? Where are their security vulnerabilities? Where are their data changes, and what are customers really experiencing, whether it's latency, poor experience throughout their products, those types of things? So security, data, and observability are just key to all of that experience and that's what we're definitely seeing as patterns, what we're seeing with our customers and also what value our ISVs are providing in that space. >> That's awesome. And the other thing I would observe is that there's more of an integration story going on around joint projects, whether it's open source. >> Absolutely. >> Because this is where we want to get that services connected. And it's mutual beneficial. I mean, this is really >> Exactly. >> whole 'nother, new kind of interoperable cloud scale. >> Yeah, if I could say one thing else there, I think that, a lot of the customers who are trying to move into the cloud now are, maybe not technology forward companies and they really need that solution. And that's very important. I think COVID has pushed a lot of companies into the cloud maybe very quickly. And, that has been something else we've observed in the market. So, solutions and full solutions between ISVs and ISVs, or ISVs and AWS is just becoming more and more common thing that we see. >> And, you mentioned John, in the open source space as well. Like, we're certainly from Amazon to Splunk. So we're talking a lot about those, but there's a lot of other firms involved in projects like OpenTelemetry. And I think it's very endearing, very heartening to see how well they cooperate in this community and how, when their interests are aligned, how effective they can be. And it's been very exciting to work in the space and very pleasant, honestly, to see everything come together with this huge set of customers and partners. >> Yeah. The pleasant surprise of the pandemic has been that people come into the cloud and they like it and they, "Hey, this works," and they double down on it. Then they realize, there's more there and they refactor. So, you're seeing real examples of that. So, this is a great discussion, great success story. Congratulations Morgan, Danielle. >> Thank you. >> Great partnership between Splunk and AWS. We've been following for a long time. And again, this highlights this whole another level of integrating super cloud kind of experience where people are getting more capabilities and doing more together, so great stuff. >> And this is just one facet of that, right? Like, there's all the other connections of Splunk Enterprise, Splunk security analytics products, and others. It's a deep, deep partnership between these firms. >> Yeah. And the companies that innovate and get that new capability are going to have an advantage. And you're seeing... >> Yes. >> Right? >> Agreed. >> And this is awesome, and great stuff, thank you for coming on and sharing that insight. >> Thank you. >> Congratulations Morgan over there at Splunk, great stuff. And Danielle, thanks for coming on and sharing the AWS perspective. >> Thanks for having me. >> And you guys are going to the next level. You moving up to stack as they say, all good stuff for customers. Thanks. >> Thank you. >> Okay. >> Thank you. >> This is season one, episode two of the AWS Partner Showcase. I'm John Furrier with theCUBE. Thanks for watching. (gentle music)
SUMMARY :
of the AWS Showcase And great. but the Splunk-AWS relationship So, Morgan, if you want it's a great demonstration of the alliance on the on the product management side One of the big things Well, the insight you on the AWS side And having that ability to plug in the velocity of releases, You got the data in here. and the changes that were What are some of the standards? is actually providing to you as a customer from the customers to to be able to roll back if you need to, and so that you can be And so when you think about observability, And part of this you see And you guys are evolving with that. and providing the ability for And you guys have been And the people who And so, the companies that is growing like crazy, 'cause this is a big hot area. OpenTelemetry is how you capture data What are the patterns you're seeing And so, that you know, And the other thing I I mean, this is really new kind of interoperable cloud scale. into the cloud maybe very quickly. And I think it's very has been that people come into the cloud And again, this highlights And this is just one And the companies that innovate And this is awesome, and great stuff, and sharing the AWS perspective. And you guys are of the AWS Partner Showcase.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
John | PERSON | 0.99+ |
Danielle | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Morgan | PERSON | 0.99+ |
Danielle Greshock | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
ORGANIZATION | 0.99+ | |
2020 | DATE | 0.99+ |
Morgan McLean | PERSON | 0.99+ |
last year | DATE | 0.99+ |
Splunk | ORGANIZATION | 0.99+ |
two companies | QUANTITY | 0.99+ |
Amazon Web Services | ORGANIZATION | 0.99+ |
50 services | QUANTITY | 0.99+ |
Amazon Web Services | ORGANIZATION | 0.99+ |
five people | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
last week | DATE | 0.99+ |
both | QUANTITY | 0.99+ |
ISV | ORGANIZATION | 0.99+ |
OpenTelemetry | TITLE | 0.99+ |
both solutions | QUANTITY | 0.99+ |
Prometheus | TITLE | 0.98+ |
ISVs | ORGANIZATION | 0.98+ |
Linux | TITLE | 0.98+ |
SiliconANGLE | ORGANIZATION | 0.97+ |
first | QUANTITY | 0.97+ |
One | QUANTITY | 0.97+ |
Morgan McLean, Splunk & Danielle Greshock, AWS | AWS Partner Showcase
(gentle music) >> Hello, welcome to theCUBE's presentation of the AWS Showcase season one, episode two with the ISV Startups partners. I'm John Furrier, your host of theCUBE. We're joined by Morgan McLean, director of product management at Splunk, and Danielle Greshock, who is the director of ISVs solution architects at AWS. Welcome to the show. Thanks for coming on. >> Thanks for having us. >> And great. Thanks for having us. >> Great to see both of you, both theCUBE alumni, but the Splunk-AWS relationship has been going very, very well. You guys are doing great business enabling this app revolution. And cloud scale has been going extremely well. So let's get into it. You guys are involved in a lot of action around application revolution, around OpenTelemetry and open source. So let's get into it. What's the latest? >> Danielle, you go ahead. >> Well, I'll just jump in first. Obviously last year, not last year, but in 2020, we launched the AWS Distro for OpenTelemetry. The idea being essentially, we're able to bring in data from partners, from infrastructure running on AWS, from apps running on AWS, to really be able to increase observability across all cloud assets at your entire cloud platform. So, Morgan, if you want to chime in on how Splunk >> Morgan: Certainly. >> has worked out OpenTelemetry. >> Yeah. I mean, OpenTelemetry is super exciting. Obviously, there's a lot of partnership points between Amazon and Splunk, but OpenTelemetry is probably one of them that's the most visible to people who aren't already maybe using these two products together. And so, as Danielle mentioned, Amazon has their own distribution of OpenTelemetry, Splunk has their own, as well, and of course there's the main open source distribution that everybody knows and loves. Just for our viewers, just for clarity's sake, the separate distributions are fundamentally very similar to, almost identical to what's offered in the open source space, but they come preconfigured and they come with support guarantees from each company, meaning that you can actually get paid full support for an open source project, which is really fantastic for customers. And as Danielle mentioned, it's a great demonstration of the alliance between Splunk and Amazon Web Services. For example, the AWS Distro, when you use it, can export data to Amazon CloudWatch, various Amazon backed open source initiatives like Prometheus and others, and to Splunk Observability Cloud and to Splunk Enterprise. So it's a place that we've worked very closely together, and it's something that we're very excited about. >> So, Morgan, I want to get your take on the on the product management side and also how product are built these days. >> One of the big things we're seeing in cloud is that open source has been the big enabler for a lot of refactoring. And you got multiple distributions, but the innovations on top of that, can you talk about how you see the productization of new innovations with open source as you guys go into this market, because this is the new dynamic with cloud. We're seeing examples all over the place. Obviously, Amazon's going next level with what they're doing, and that open source, it's not a one game for all of it. You can have mix and match. Take us through the product angle. >> And in many ways, this is just another wave of the same thing, right? Like, if you think back in time, we all used and still use in many cases, virtual machines, most of those are based on Linux, right? Another large open source project. And so, open source software has been accelerating innovation in the cloud space and in the computing space generally for a very long time, which is fantastic. Our excitement with something like OpenTelemetry comes from both the project's capabilities but also what we can do with it. So for those who aren't already familiar with OpenTelemetry, OpenTelemetry allows you to extract really critical system telemetry, application signals and everything else you need from your own applications, from new services, from your infrastructure, from everything that you're running in a cloud environment. You can then send that data to another location for processing. And so John, you ask like, how does this accelerate innovation? What does it unlock? Well, the insight you can gain from this data means you can become so much more efficient as a development organization. You can make your applications so much more effective because when you send that data to something like Splunk Observability Cloud, to something like Amazon CloudWatch, to various other solutions on the market, they can give you deep, deep insight into your application's performance, to its structure, they can help you reduce outages. And so, it's very, very powerful because it allows organizations to use tools like Splunk, like Amazon, like other things to innovate so much more effectively. >> Danielle, can you comment >> If I could... >> on the AWS side because this is again on the big point. You guys are going next level, and you're starting to see patterns in the ISV world, certainly on the architecture side of partners doing things differently now on top of what they've already done. Could you share how AWS is helping customers accelerate? >> Well, just as Morgan was talking about what OpenTelemetry provides, you can see how from a partnership perspective, this is so valuable, right? What the partner team here at AWS is in the business of doing, is really enabling customer choice, right? And having that ability to plug in and pull data from different sources, post it to different sources, make it available for visibility across all of your resources is very powerful and it's something from the partner community that we really value because we want customers to be able to select best of breed solutions, what works for their business, which businesses are different and they may have different needs, and that also fosters that true innovation. A small company is going to develop and release software a lot differently than a large enterprise. And so, being able to support something like OpenTelemetry just enables that for all different kinds of customers. >> Morgan, add to that because the velocity of releases, certainly operational, stability, is key every predominant security, uptime, these are top concerns. And, you mention data too, >> And you mention challenges. >> You got the data in here. So you got a lot of data moving around, a lot of value. What's your take? >> Yeah. So, I'll speak with some specifics. So a challenge that developers have had for years when you're developing large services, which you can now do with platforms like AWS. So, it's very easy to go develop huge deployments. But a challenge they have is you go and build a mess, right? And like, I've worked earlier in my career in Web Services. And I remember in one of the first orgs I was in, I was one of the five people who really understood our ecommerce stack. Right? And so like, I would get dragged into all these meetings and I'd have to go draw like the 50 services we had, and how they interacted, and the changes that were made in the last week. And without observability tools like Splunk Observability Cloud, like the ones offered by Amazon, like the ones that are backed by the data that comes with OpenTelemetry, organizations basically rely on people like this, to go draw out their deployments so they understand what it is they've built. Well, as you can imagine, this crimps your development velocity, because most of your engineers, most of your tech leads, most of everyone else don't actually understand what it is they've built what it is they're running, because they need that global context. You get something like OpenTelemetry and the solutions that consume the data from it, and suddenly now, all your developers have that context, all of them when they're adding functionality to a service or they're updating their infrastructure, can actually understand how it interacts with the rest of the broader application. This lets you speed up your time to development, this lets you ship more safely, more securely. And finally, when things do go wrong, which will be less frequent, but when they do go wrong, you can fix them super rapidly. >> If I'm a customer, let me ask a question. I'm a customer and I say, "Okay, I love AWS, I love Splunk, I love OpenTelemetry. I got to have open sources, technology innovation is happening." What's the integration? What are some of the standards? Can you take us through how that's working together with you guys as a shared platform? >> Yeah. So let's take the Amazon distribution for OpenTelemetry or even the Splunk one. One of the first things they do is they include all of the receivers, all of the sort of data capture components that you need, out of box for platforms like AWS, right? And so, right away, you get that power and flexibility where you're getting access to all of these data sources, right? And so, that's part of that partnership. And additionally, once the data comes into OpenTelemetry, you can now send that to various different data sources, including, as Danielle mentioned, to multiple at the same time. So you can use whatever tools you want. And so when you talk about like what the partnership is actually providing to you as a customer and still, this is just within the context of OpenTelemetry, obviously there's a much broader partnership between these two companies than just that. But within the context of OpenTelemetry means you can download one of these distributions. It's fully supported. It works with both solutions and everything is just great, right? You don't need to go fiddle with that out of the box. To be clear, OpenTelemetry is a batteries included project, right? This means that even the standard distributions of OpenTelemetry include the components you need. You have to go directly, reference them and ensure that they're packaged in there, but they exist, right. But the nice thing about these distributions is that it's done, it's out of the box, you don't even have to worry about is something missing or do I need to include new exporters or new receivers? It's all there. It's preconfigured. It just works. And if something goes wrong and you have a support contact, you pick up the phone, you talk to someone to get it fixed. >> Danielle, what's the Amazon side 'cause agility and scale is one of the highlights you guys are seeing. How does this tie into that and how are you guys working backwards from the customers to support the partners? >> Well, I think just to add on essentially to what Morgan said, I think that AWS is a cloud platform, has always really had a focus on developers. And, we talk a lot about how AWS and Amazon as a whole really embraces this continuous integration and continuous deployment methods inside of our organization. And we talk about services, and observability is a huge part of that. The only way that you're actually able to release hundreds, thousands of times a day like Amazon does, is by having an observability platform, to be able to measure metrics, see changes in the environment, to be able to roll back if you need to, and to be able to quickly mitigate any challenges or anything that goes wrong at any part of the process. And so, when we preach that to our customers, I think it's something that we do that because we live it and breathe it. And so, things such as OpenTelemetry and such as the products that Splunk builds, those are also ways in which we believe our customers can achieve that. >> Yeah. And we can... I mean, as I mentioned before, this partnership goes well beyond OpenTelemetry, right? And so, if you go use like Splunk Enterprise, Enterprise Cloud, Splunk Observability Cloud, and you're running on AWS, you have excellent support and excellent visibility into your Amazon infrastructure, into the services and applications you've deployed on top of that infrastructure. We try and give you, and I think we do succeed in this. We give you the best possible experience, the deepest possible visibility, into what it is you've deployed on AWS, so that you can be even more successful as a business, and so that you can be even more successful on AWS as a platform. >> Yeah. This is a great conversation, Morgan. You mentioned the early days of Web Services. AWS stands for Amazon Web Services built on web services. So interesting throwback there, but made me think about the days of the early days of web services. And if you look at data, what's going on now, the top partners in AWS, you're seeing a lot of people thinking about data differently, they're refactoring, a lot of machine learning, a lot of AI going on at scale. So then, you got cloud native, things like Kubernetes and these new services being stood up and teared down with automation. A whole new operating model's coming. And so when you think about observability, the importance of it, I mean, can you share your perspective on this whole 'nother level? I mean, I always say that whole another level sounds cliche, but it is next level. I mean, this is completely different. What's your reaction? >> Yeah. There there's a ton of factors here, right? So as you point out, companies are totally shifting how they use their cloud infrastructure. And part of this you see during their cloud migrations, a part of it you see after, and they're shifting from their sort of stateful VMs that they may have had in the past to infrastructure that they tear down and put up regularly. And there's a lot more automation. With this, comes as I mentioned before, complexity, right? And also, with this comes more and more businesses becoming even more reliant on their digital infrastructure. And so, not having observability into your applications, into your services, into your infrastructure, to me, is akin to running a business, say running a large warehousing or distribution company, but not having any idea where you're shipping products or where things are, or not having any accounting or CFO, right? Like, business has become so digital. Business is so reliant on technology, and that's unlocked a ton of new things. It's great. But not having visibility into how that technology works or what it is that's deployed or how to fix it is akin to having no visibility to anything else in your business. It's nuts. And so, observability is super, super critical, particularly for customers who are adopting this new wave of cloud technologies on platforms like AWS. >> Danielle, on your side too, you're enabling this new capability so that businesses can do it, the partners do it, we're calling it super cloud. We've been calling it super cloud kind of dynamic where new things are happening with the data. And you guys are evolving with that. Can you share what you're seeing on your side as your partners start to go to the next level? What are you guys doing? How does it all come together? >> Well, we always talk about what has happened with data in the last couple of years, which the cloud has really enabled around, you know, variety and velocity and there's one other "V" that's escaping me right now, but essentially, all of this data is coming in and providing the ability for us to make better decisions, to build better products, to provide better experiences for customers. And so, I just think, the OpenTelemetry project, as well as what Splunk is doing is just another example of how we're taking this massive amount of data and being able to provide better experiences and outcomes for customers. >> And you guys have been working along together for long time, Splunk, and, it's been a great partners, if we're going back with that been covering it on theCUBE and SiliconANGLE. So, we know that, the change is key observability. Can you imagine a company without a CFO, Morgan? That's just boggles your mind, but that's what it's like right now. So... >> It is, yeah. >> And the people who take advantage of that are winning, right? So it's like, that's the key. >> Yeah, I know. I mean, even in my own career, right, I've moved between different companies. And I remember, when I joined Google in particular, which is where I worked at previously, I was very impressed with their internal observability tools. And I'm certain, I haven't worked at Amazon. I'm certainly, I just assume inside of Amazon they're excellent as well, so a lot of the large cloud firms these days. But it was so refreshing going from an organization where if we had some outage or something went wrong, there were like a very small set of people who could actually understand what was going on. And then you would just have to manually dive through logs and correlate requests manually between services. It's very challenging. And so, when things went wrong, they went wrong for a long, long time. And so, the companies that understood this even in the past are already very successful as a result. I think now, the rest of the industry is really in the midst of adopting these observability practices and the tools that are required to implement them, because you're right. Otherwise your development velocity slows down. Now you're getting out competed by your competition. And then, when you have a problem, it blows up for ages. And once again, your competition can take advantage of it. >> And, can you just summarize the observability piece relative to the OpenTelemetry? Where is that going to go? Where do you see that evolving? >> Sure. >> I see open source is growing like crazy, we all know that. >> Of course. >> But OpenTelemetry in particular and open source, 'cause this is a big hot area. >> Yes. So to set the stage for people, OpenTelemetry, unlocks observability in many ways. As I mentioned earlier, OpenTelemetry is how you capture data out of your application. It doesn't process it. It's not a replacement for something like Amazon CloudWatch or any Splunk's products, but it's how we get the data out of your system, which is a remarkably difficult problem. I won't dive into it today, but, those who work in this space are very aware. That's why this project exists and it's so big, that actually extracting information, metrics, logs, distributed traces, profiles, everything else, from your applications and from your infrastructure is very, very difficult. So for OpenTelemetry, where it's going is just continually getting better at extracting more types of data from more sources, and doing that more effectively for people in a more standardized way. That will unlock firms like Splunk, firms, like Amazon and others to better process this data. In terms of where that's going, the sky's the limit, right? Like, everyone's familiar with APM, people are familiar with infrastructure monitoring, but there's a lot more capabilities coming there for security analytics, for network performance monitoring, for getting down all the way to single lines of coding your application, how they impact everything. There's just so much power that's coming to the industry right now. I'm really excited to see where things go in the next few years. >> And Danielle, you're in the middle of all the action as a solution architect, really set the stage for their companies and the ISVs, and this is a big, hot area. What are the patterns you're seeing and what are some of the best practices that you're doing will help companies? >> Right. So I think, summarizing our entire conversation, the big things that we're seeing in the market is essentially more and more companies are looking to move to a continuous deployment and a continuous integration environment. And they're looking to innovate faster and spend less time hot patching or hot fixing their environments and they want to spend more time innovating. And so, that you know, the patterns that we're seeing is... What I see and what I actually experience firsthand at re:Invent when I talk to probably over 40 or 50 ISVs, is customers want to know in their environment, where are their changes? Where are their security vulnerabilities? Where are their data changes, and what are customers really experiencing, whether it's latency, poor experience throughout their products, those types of things? So security, data, and observability are just key to all of that experience and that's what we're definitely seeing as patterns, what we're seeing with our customers and also what value our ISVs are providing in that space. >> That's awesome. And the other thing I would observe is that there's more of an integration story going on around joint projects, whether it's open source. >> Absolutely. >> Because this is where we want to get that services connected. And it's mutual beneficial. I mean, this is really >> Exactly. >> whole 'nother, new kind of interoperable cloud scale. >> Yeah, if I could say one thing else there, I think that, a lot of the customers who are trying to move into the cloud now are, maybe not technology forward companies and they really need that solution. And that's very important. I think COVID has pushed a lot of companies into the cloud maybe very quickly. And, that has been something else we've observed in the market. So, solutions and full solutions between ISVs and ISVs, or ISVs and AWS is just becoming more and more common thing that we see. >> And, you mentioned John, in the open source space as well. Like, we're certainly from Amazon to Splunk. So we're talking a lot about those, but there's a lot of other firms involved in projects like OpenTelemetry. And I think it's very endearing, very heartening to see how well they cooperate in this community and how, when their interests are aligned, how effective they can be. And it's been very exciting to work in the space and very pleasant, honestly, to see everything come together with this huge set of customers and partners. >> Yeah. The pleasant surprise of the pandemic has been that people come into the cloud and they like it and they, "Hey, this works," and they double down on it. Then they realize, there's more there and they refactor. So, you're seeing real examples of that. So, this is a great discussion, great success story. Congratulations Morgan, Danielle. >> Thank you. >> Great partnership between Splunk and AWS. We've been following for a long time. And again, this highlights this whole another level of integrating super cloud kind of experience where people are getting more capabilities and doing more together, so great stuff. >> And this is just one facet of that, right? Like, there's all the other connections of Splunk Enterprise, Splunk security analytics products, and others. It's a deep, deep partnership between these firms. >> Yeah. And the companies that innovate and get that new capability are going to have an advantage. And you're seeing... >> Yes. >> Right? >> Agreed. >> And this is awesome, and great stuff, thank you for coming on and sharing that insight. >> Thank you. >> Congratulations Morgan over there at Splunk, great stuff. And Danielle, thanks for coming on and sharing the AWS perspective. >> Thanks for having me. >> And you guys are going to the next level. You moving up to stack as they say, all good stuff for customers. Thanks. >> Thank you. >> Okay. >> Thank you. >> This is season one, episode two of the AWS Partner Showcase. I'm John Furrier with theCUBE. Thanks for watching. (gentle music)
SUMMARY :
of the AWS Showcase And great. but the Splunk-AWS relationship So, Morgan, if you want it's a great demonstration of the alliance on the on the product management side One of the big things Well, the insight you on the AWS side And having that ability to plug in the velocity of releases, You got the data in here. and the changes that were What are some of the standards? is actually providing to you as a customer from the customers to to be able to roll back if you need to, and so that you can be And so when you think about observability, And part of this you see And you guys are evolving with that. and providing the ability for And you guys have been And the people who And so, the companies that is growing like crazy, 'cause this is a big hot area. OpenTelemetry is how you capture data What are the patterns you're seeing And so, that you know, And the other thing I I mean, this is really new kind of interoperable cloud scale. into the cloud maybe very quickly. And I think it's very has been that people come into the cloud And again, this highlights And this is just one And the companies that innovate And this is awesome, and great stuff, and sharing the AWS perspective. And you guys are of the AWS Partner Showcase.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
John | PERSON | 0.99+ |
Danielle | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Morgan | PERSON | 0.99+ |
Danielle Greshock | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
ORGANIZATION | 0.99+ | |
2020 | DATE | 0.99+ |
Morgan McLean | PERSON | 0.99+ |
last year | DATE | 0.99+ |
Splunk | ORGANIZATION | 0.99+ |
two companies | QUANTITY | 0.99+ |
Amazon Web Services | ORGANIZATION | 0.99+ |
50 services | QUANTITY | 0.99+ |
Amazon Web Services | ORGANIZATION | 0.99+ |
five people | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
last week | DATE | 0.99+ |
both | QUANTITY | 0.99+ |
ISV | ORGANIZATION | 0.99+ |
OpenTelemetry | TITLE | 0.99+ |
both solutions | QUANTITY | 0.99+ |
Prometheus | TITLE | 0.98+ |
ISVs | ORGANIZATION | 0.98+ |
Linux | TITLE | 0.98+ |
SiliconANGLE | ORGANIZATION | 0.97+ |
Splunk | PERSON | 0.97+ |
Matt Provo, StormForge
(bright upbeat music) >> The adoption of container orchestration platforms is accelerating at a rate as fast or faster than any category in enterprise IT. Survey data from Enterprise Technology Research shows Kubernetes specifically leads the pack into both spending velocity and market share. Now like virtualization in its early days, containers bring many new performance and tuning challenges in particular insuring consistent and predictable application performance is tricky especially because containers, they're so flexible and they enable portability. Things are constantly changing. DevOps pros have to way through a sea of observability data and tuning the environment becomes a continuous exercise of trial and error. This endless cycle taxes resources and kills operational efficiency. So teams often just capitulate and simply dial up and throw unnecessary resources at the problem. StormForge is a company founded mid last decade that is attacking these issues with a combination of machine learning and data analysis. And with me to talk about a new offering that directly addresses these concerns is Matt Provo, founder and CEO of StormForge. Matt, welcome to theCUBE. Good to see you. >> Good to see you. Thanks for having me. >> Yeah, so we saw you guys at a KubeCon sort of first introduce you to our community but add a little color to my intro there if you will. >> Yeah, well, Semi stole my thunder but I'm okay with that. Absolutely agree with everything you said in the intro. You know, the problem that we have set out to solve which is tailor made for the use of real machine learning not machine learning kind of as a marketing tag is connected to how workloads on Kubernetes are really managed from a resource efficiency standpoint. And so a number of years ago, we built the core machine learning engine and have now turned that into a platform around how Kubernetes resources are managed at scale. And so organizations today as they're moving more workloads over, sort of drink the Kool-Aid of the flexibility that comes with Kubernetes and how many knobs you can turn. And developers in many ways love it. Once they start to operationalize the use of Kubernetes and move workloads from pre-production into production, they run into a pretty significant complexity wall. And this is where StormForge comes in to try to help them manage those resources more effectively in ensuring and implementing the right kind of automation that empowers developers into the process ultimately does not automate them out of it. >> So you've got news. You had launch coming to further address these problems. Tell us about that. >> Yeah, so historically, you know, like any machine learning engine, we think about data inputs and what kind of data is going to feed our system to be able to draw the appropriate insights out for the user. And so historically we've kind of been single threaded on load and performance tests in a pre-production environment. And there's been a lot of adoption of that, a lot of excitement around it and frankly amazing results. My vision has been for us to be able to close the loop, however, between data coming out of pre-production and the associated optimizations and data coming out of production environment and our ability to optimize that. A lot of our users along the way have said these results in pre-production are fantastic. How do I know they reflect reality of what my application is going to experience in a production environment? And so we're super excited to announce kind of the a second core module for our platform called Optimize Live. The data input for that is observability and telemetry data coming out of APM platforms and other data sources. >> So this is like Nirvana. So I wonder if we could talk a little bit more about the challenges that this addresses. I mean, I've been around a while and it really have observed... And I used to ask, you know, technology companies all the time. Okay, so you're telling me beforehand what the optimal configuration should be and resource allocation. What happens if something changes? >> Yeah. >> And then it's always, always a pause. >> Yeah. >> And Kubernetes is more of a rapidly changing environment than anything we've ever seen. So specifically the problem you're addressing. Maybe talk about that a little bit. >> Yeah, so we view what happens in pre-production as sort of the experimentation phase. And our machine learning is allowing the user to experiment in scenario plan. What we're doing with Optimize Live and adding the the production piece is what we kind of also call kind of our observation phase. And so you need to be able to run the appropriate checks and balances between those two environments to ensure that what you're actually deploying and monitoring from an application performance, from a cost standpoint is with your SLOs and your SLAs as well as your business objectives. And so that's the entire point of this edition is to allow our users to experience hopefully the Nirvana associated with that because it's an exciting opportunity for them and really something that no else is doing from the standpoint of closing that loop. >> So you said front machine learning not as a marketing tag. So I want you to sort of double click on that. What's different than how other companies approach this problem? >> Yeah, I mean, part of it is a bias for me and a frustration as a founder of the reason I started the company in the first place. I think machine learning or AI gets tagged to a lot of stuff. It's very buzzwordy. It looks good. I'm fortunate to have found a number of folks from the outset of the company with, you know, PhDs in Applied Mathematics and a focus on actually building real AI at the core that is connected to solving the right kind of actual business problems. And so, you know, for the first three or four years of the company's history, we really operated as a lab. And that was our focus. We then decided, we're trying to connect a fantastic team with differentiated technology to the right market timing. And when we saw all these pain points around how fast the adoption of containers and Kubernetes have taken place but the pain that the developers are running into, we actually found for ourselves that this was the perfect use case. >> So how specifically does Optimize Live work? Can you add a little detail on that? >> Yes, so when you... Many organizations today have an existing monitoring APM observability suite really in place. They've also got a metric source. So this could be something like Datadog or Prometheus. And once that data starts flowing, there's an out of the box or kind of a piece of Kubernetes that ships with it called the VPA or the Vertical Pod Autoscaler. And less than, really than 1% of Kubernetes users take advantage of the VPA mostly because it's really challenging to configure and it's not super compatible with the the tool set or, you know, the ecosystem of tools in a Kubernetes environment. And so our biggest competitor is the VPA. And what's happening in this environment or in this world for developers is they're having to make decisions on a number of different metrics or resource elements typically things like memory and CPU. And they have to decide what are the requests I'm going to allow application and what are the limits? So what are those thresholds that I'm going to be okay with so that I can, again, try to hit my business objectives and keep in line with my SLAs? And to your earlier point in the intro, it's often guesswork. You know, they either have to rely on out of the box recommendations that ship with the databases and other services that they are using or it's a super manual process to go through and try to configure and tune this. And so with Optimize Live, we're making that one click. And so we're continuously and consistently observing and watching the data that's flowing through these tools and we're serving back recommendations for the user. They can choose to let those recommendations automatically patch and deploy or they can retain some semblance of control over are the recommendations and manually deploy them into their environment themselves. And we, again, really believe that the user knows their application. They know the goals that they have and we don't. But we have a system that's smart enough to align with the business objectives and ultimately provide the relevant recommendations at that point. >> So the business objectives are an input from the application team? >> Yep. >> And then your system is smart enough to adapt and address those. >> Application over application, right? And so the thresholds in any given organization across their different ecosystem of apps or environment could be different. The business objectives could be different. And so we don't want to predefine that for people. We want to give them the opportunity to build those thresholds in and then allow the machine learning to learn and to send recommendations within those bounds. >> And we're going to hear later from a customer who's hosting a Drupal, one of the largest Drupal hosts. So it's all do it yourself across thousands of customers so it's, you know, very unpredictable. I want to make something clear though as to where you fit in the ecosystem. You're not an observability platform, you leverage observability platforms, right? So talk about that and where you fit into the ecosystem. >> Yeah, so it's a great point. We're also, you know, a series B startup and growing. We've the choice to be very intentionally focused on the problems that we've solve. And we've chosen to partner or integrate otherwise. And so we do get put into the APM category from time to time. We are really an intelligence platform. And that intelligence and insights that we're able to draw is because of the core machine learning we've built over the years. And we also don't want organizations or users to have to switch from tools and investments that they've already made. And so we were never going to catch up to to Datadog or Dynatrace or Splunk or AppDynamics or some of the other. And we're totally fine with that. They've got great market share and penetration. They do solve real problems. Instead, we felt like users would want a seamless integration into the tools they're already using. And so we view ourselves as kind of the Intel inside for that kind of a scenario. And it takes observability and APM data and insights that were somewhat reactive. They're visualized and somewhat reactive. And we add that proactive nature onto it, the insights and ultimately the appropriate level of automation. >> So when I think, Matt, about cloud native and I go back to the sort of origins of CNCF who's a, you know, handful of companies. And now you look at the participants it'll, you know, make your eyes bleed. How do you address dealing with all those companies and what is the partnership strategy? >> Yeah, it's so interesting because it's just that even that CNCF landscape has exploded. It was not too long ago where it was as small or smaller than the FinOps landscape today which by the way, the FinOps piece is also on a a neck breaking, you know, growth curve. We, I do see, although there are a lot of companies and a lot of tools, we're starting to see a significant amount of consistency or hardening of the tool chain, you know, with our customers and users. And so we've made strategic and intentional decisions on deep partnerships in some cases like OEM uses of our technology and certainly, you know, intelligent and seamless integrations into a few. So, you know, we'll be announcing a really exciting partnership with AWS and that specifically what they're doing with EKS, their Kubernetes distribution and services. We've got a deep partnership and integration with Datadog and then with Prometheus and specifically a few other cloud providers that are operating, manage Prometheus environments. >> Okay, so where do you want to take this thing? You're not taking the observability guys head on, smart move. So many of those even entering the market now. But what is the vision? >> Yeah, so we've had this debate a lot as well 'cause it's super difficult to create a category. You know, on one hand, you know, I have a lot of respect for founders and companies that do that. On the other hand from a market timing standpoint, you know we fit into AIOps, that's really where we fit. You know, we've made a bet on the future of Kubernetes and what that's going to look like. And so from a containers and Kubernetes standpoint, that's our bet. But we're an AIOps platform. You know, we'll continue getting better at the problems we solve with machine learning and we'll continue adding data inputs. So we'll go, you know, we'll go beyond the application layer which is really where we play now. We'll add, you know, kind of whole cluster optimization capabilities across the full stack. And the way we will get there is by continuing to add different data inputs that make sense across the different layers of the stack. And it's exciting. We can stay vertically oriented on the problems that we're really good at solving but we can become more applicable and compatible over time. >> So that's your next concentric circle. As the observability vendors expand their observation space, you can just play right into that. >> Yeah. >> The more data you get because your purpose built to solving these types of problems. >> Yeah, so you can imagine a world right now out of observability, we're taking things like telemetry data pretty quickly. You can imagine a world where we take traces and logs and other data inputs as that ecosystem continues to grow, it just feeds our own, you know, we are reliant on data. >> Excellent, Matt, thank you so much. >> Thanks for having me. >> Appreciate for coming on. Okay, keep it right there in a moment. We're going to hear from a customer with a highly diverse and constantly changing environment that I mentioned earlier. They went through a major replatforming with Kubernetes on AWS. You're watching theCUBE, you are leader in enterprise tech coverage. (bright upbeat music)
SUMMARY :
and CEO of StormForge. Good to see you. Yeah, so we saw you guys at a KubeCon that empowers developers into the process You had launch coming to and the associated optimizations And I used to ask, you know, And Kubernetes is more of And so that's the entire So I want you to sort And so, you know, for the And so our biggest competitor is the VPA. is smart enough to adapt And so the thresholds in as to where you fit in the ecosystem. We've the choice to be and I go back to the or hardening of the tool chain, you know, Okay, so where do you And the way we will get there As the observability vendors to solving these types of problems. as that ecosystem continues to grow, and constantly changing environment
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
AWS | ORGANIZATION | 0.99+ |
StormForge | ORGANIZATION | 0.99+ |
Matt | PERSON | 0.99+ |
Matt Provo | PERSON | 0.99+ |
KubeCon | EVENT | 0.99+ |
Dynatrace | ORGANIZATION | 0.99+ |
first | QUANTITY | 0.99+ |
Datadog | ORGANIZATION | 0.99+ |
Splunk | ORGANIZATION | 0.99+ |
four years | QUANTITY | 0.99+ |
AppDynamics | ORGANIZATION | 0.99+ |
CNCF | ORGANIZATION | 0.98+ |
mid last decade | DATE | 0.98+ |
Prometheus | TITLE | 0.98+ |
Kubernetes | TITLE | 0.98+ |
DevOps | TITLE | 0.97+ |
one click | QUANTITY | 0.97+ |
first three | QUANTITY | 0.97+ |
Semi | PERSON | 0.97+ |
two environments | QUANTITY | 0.97+ |
Drupal | TITLE | 0.96+ |
both | QUANTITY | 0.96+ |
thousands of customers | QUANTITY | 0.95+ |
Kool-Aid | ORGANIZATION | 0.94+ |
EKS | ORGANIZATION | 0.92+ |
today | DATE | 0.92+ |
of years ago | DATE | 0.88+ |
1% | QUANTITY | 0.86+ |
one | QUANTITY | 0.85+ |
Datadog | TITLE | 0.85+ |
Optimize Live | TITLE | 0.84+ |
second core module | QUANTITY | 0.83+ |
Intel | ORGANIZATION | 0.82+ |
Live | TITLE | 0.8+ |
single | QUANTITY | 0.76+ |
Enterprise Technology | ORGANIZATION | 0.74+ |
AIOps | ORGANIZATION | 0.72+ |
theCUBE | TITLE | 0.72+ |
theCUBE | ORGANIZATION | 0.72+ |
Kubernetes | ORGANIZATION | 0.68+ |
double | QUANTITY | 0.6+ |
Optimize | ORGANIZATION | 0.59+ |
Nirvana | ORGANIZATION | 0.46+ |
FinOps | ORGANIZATION | 0.42+ |
FinOps | TITLE | 0.41+ |
Nirvana | TITLE | 0.39+ |
Matt Provo, StormForge
(upbeat music) >> The adoption of container orchestration platforms is accelerating at a rate as fast or faster than any category in enterprise IT. Survey data from enterprise technology research shows Kubernetes specifically, leads the pack into both spending velocity and market share. Now like virtualization in its early days, containers bring many new performance and tuning challenges, in particular ensuring consistent and predictable application performance is tricky especially because containers they're so flexible and they enable portability, things are constantly changing. DevOps Pros have to wade through a sea of observability data and tuning the environment becomes a continuous exercise of trial and error. This endless cycle taxes resources and kills operational efficiency. So teams often just capitulate and simply dial up and throw unnecessary resources at the problem. StormForge is a company founded mid last decade that is attacking these issues with a combination of machine learning and data analysis. And with me to talk about a new offering that directly addresses these concerns is Matt Provo, founder and CEO of StormForge. Matt, welcome to The CUBE. Good to see you. >> Good to see you. Thanks for having me. >> Yeah. So we saw you guys at CUBE con, sort of first introduce you to our community, but add a little color to my intro there if you want. >> Well, you semi stole my thunder, but I'm okay with that. Absolutely agree with everything you said in the intro. The problem that we have set out to solve, which is tailor made for the use of real machine learning, not machine learning kind of as a marketing tag is connected to how workloads on Kubernetes are really managed from a resource efficiency standpoint. And so a number of years ago, we built the core machine learning engine and have now turned that into a platform around how Kubernetes resources are managed at scale. And so organizations today, as they're moving more workloads over, sort of drink the cool-Aid of the flexibility that comes with Kubernetes and how many knobs you can turn and developers in many ways love it. Once they start to operationalize use of Kubernetes and move workloads from pre-production into production, they run into a pretty significant complexity wall. And this is where StormForge comes in to try to help them manage those resources more effectively and ensuring and implementing the right kind of automation that empowers developers into the process, ultimately does not automate them out of it. >> So you've got news, you a hard launch coming to further address these problems. Tell us about that. >> Yeah. So historically, like any machine learning engine, we think about data inputs and what kind of data is going to feed our system to be able to draw the appropriate insights out for the user. And so historically we've kind of been single threaded on load and performance tests in a pre-production environment. And there's been a lot of adoption of that, a lot of excitement around it and frankly, amazing results. My vision has been for us to be able to close the loop, however, between data coming out of pre-production and the associated optimizations and data coming out of production environment and our ability to optimize that. A lot of our users along the way have said, these results in pre-production are fantastic. How do I know they reflect reality of what my application is going to experience in a production environment? And so we're super excited to announce, kind of the second core module for our platform called optimized live. The data input for that is observability and telemetry data coming out of APM platforms and other data sources. >> So this is like Nirvana. So I wonder if we could talk a little bit more about the challenges that this addresses. I mean, I've been around a while and it really have observed, and I used to ask technology companies all the time. Okay. So you're telling me beforehand what the optimal configuration should be and resource allocation, what happens if something changes? And then it's always, always a pause. And Kubernetes is more of a rapidly changing environment than anything we've ever seen. So this is specifically the problem you're addressing, maybe talk about that a little bit more. >> Yeah. So we view what happens in pre-production as sort of the experimentation phase. And our machine learning is is allowing the user to experiment and scenario plan. What we're doing with optimized live and adding the production piece is what we kind of also call, kind of our observation phase. And so you need to be able to run the appropriate checks and balances between those two environments to ensure that what you're actually deploying and monitoring from an application performance, from a cost standpoint is aligning with your SLOs and your SLAs, as well as your business objectives. And so that's the entire point of this edition, is to allow our users to experience, hopefully the the Nirvana associated with that, because it's an exciting opportunity for them and really something that nobody else is doing from the standpoint of closing that loop. >> So you said up front, machine learning not as a marketing tag. So I want you to sort of double click on that. What's different than how other companies approach this problem? >> Yeah. I mean, part of it is a bias for me and a frustration as a founder of the reason I started the company in the first place. I think machine learning or AI gets tagged to a lot of stuff. It's very buzz wordy, it looks good. I'm fortunate to have found a number of folks from the outset of the company with PhDs and applied mathematics and a focus on actually building real AI at the core that is connected to solving the right kind of actual business problems. And so for the first three or four years of the company's history, we really operated as a lab. And that was our focus. We then decided, we're trying to connect a fantastic team with differentiated technology to the right market timing. And when we saw all these pain points around, how fast the adoption of containers and Kubernetes have taken place, but the pain that developers are running into, we actually found for ourselves that this was the perfect use case. >> So how specifically does optimize live work? Can you add a little detail on that? >> Yeah. So when you... Many organizations today have an existing monitoring APM, observability suite really in place, they've also got a metric source. So this could be something like Datadog, or Prometheus. And once that data starts flowing there's an out of the box or kind of a piece of Kubernetes that ships with it called the VPA or the vertical pod auto scaler. And less than, really less than 1% of Kubernetes users take advantage of of the VPA, mostly because it's really challenging to configure and it's not super compatible with the tool set or the ecosystem of tools in a Kubernetes environment. And so our biggest competitor is the VPA. And what's happening in this world for developers is they're having to make decisions on a number of different metrics or resource elements, typically things like memory and CPU, and they have to decide, what are the requests I'm going to allow for this application and what are the limits? So what are those thresholds that I'm going to be okay with? So that I can, again, try to hit my business objectives and keep in line with my SLAs. And to your earlier point in the intro, it's often guesswork. They either have to rely on out of the box recommendations that ship with the databases and other services that they are using, or it's a super manual process to go through and try to configure and tune this. And so with optimized live, we're making that one click. And so we're continuously and consistently observing and watching the data that's flowing through these tools and we're serving back recommendations for the user. They can choose to let those recommendations automatically patch and deploy, or they can retain some semblance of control over the recommendations and manually deploy them into their environment themselves. And we, again, really believe that the user knows their application. They know the goals that they have, we don't, but we have a system that's smart enough to align with the business objectives and ultimately provide the relevant recommendations at that point. >> So the business objectives are an input from the application team. And then your system is smart enough to a adapt and address those? >> Application over application. And so the thresholds in any given organization across their different ecosystem of apps or environment could be different. The business objectives could be different. And so we don't want to predefine that for people. We want to give them the opportunity to build those thresholds in and then allow the machine learning to learn and to send recommendations within those bounds. >> And we're going to hear later from a customer who's hosting a Drupal, one of the largest Drupal hosts. So it's all do it yourself across that of customers. So it's very unpredictable. I want to make something clear though. As to where you fit in the ecosystem, you're not an observability platform, you leverage observability platforms. So talk about that and where you fit in into the ecosystem. >> Yeah. So this is a great point. We're also a series B startup and growing where we've the choice to be very intentionally focused on the problems that we've solve and we've chosen to partner or integrate otherwise. And so we do get put into the APM category from time to time. We are really an intelligence platform and that intelligence and insights that we're able to draw is because of the core machine learning we've built over the years. And we also don't want organizations or users to have to switch from tools and investments that they've already made. And so we were never going to catch up to Datadog or Dynatrace or Splunk or UpDynamics or some of the other. And we're totally fine with that. They've got great market share and penetration. They do solve real problems. Instead, we felt like users would want a seamless integration into the tools they're already using. And so we view ourselves as kind of the Intel inside for that kind of a scenario. And it takes observability and APM data and insights that were somewhat reactive, they're visualized and somewhat reactive and we make those, we add that proactive nature onto it, the insights and ultimately the appropriate level of automation. >> So when I think Matt about cloud native and I go back to the sort of origins of CNCF, it was a handful of companies, and now you look at the participants make your eyes bleed. How do you address dealing with all those companies and what's the partnership strategy? >> Yeah, it's so interesting because, just that even that CNCF landscape has exploded. It was not too long ago where it was as small or smaller than the Finops landscape today, which by the way, the Finops piece is also on a neck breaking growth curve. I do see, although there are a lot of companies and a lot of tools, we're starting to see a significant amount of consistency or hardening of the tool chain with our customers and users. And so we've made strategic and intentional decisions on deep partnerships, in some cases like OEM, uses of our technology and certainly, intelligent and seamless integrations into a few. So we'll be announcing a really exciting partnership with AWS and that specifically what they're doing with EKS, their Kubernetes distribution and services. We've got a deep partnership and integration with Datadog and then with Prometheus, and specifically a few other cloud providers that are operating manage Prometheus environments. >> Okay. So where do you want to take this thing? You're not taking the observability guys head on, smart move. So many of those even entering the market now. But what is the vision? >> Yeah. So we've had this debate a lot as well 'cause it's super difficult to create a category. On one hand, I have a lot of respect for founders and companies that do that, on the other hand, from a market timing standpoint, we fit into AI Ops, that's really where we fit. We've made a bet on the future of Kubernetes and what that's going to look like. And so from a containers and Kubernetes standpoint that's our bet, but we're an AI Ops platform, we'll continue getting better at the problems we solve with machine learning and we'll continue adding data inputs. So we'll go beyond the application layer, which is really where we play now. We'll add kind of whole cluster optimization capabilities across the full stack. And the way we will get there is by continuing to add different data inputs that make sense across the different layers of the stack. And it's exciting. We can stay vertically oriented on the problems that we're really good at solving but we can become more applicable and compatible over time. >> So that's your next concentric circle. As the observability vendors expand their observation space, you can just play right into that? More data you get because your purpose built to solving these types of problems. >> Yeah. So you can imagine a world right now out of observability, we're taking things like telemetry data. Pretty quickly you can imagine a world where we take traces and logs and other data inputs as that ecosystem continues to grow. It just feeds our own, we are reliant on data. >> Excellent. Matt, thank you so much. Appreciate you coming on. >> Thanks for having me. >> Okay. Keep it right there. In a moment, we're going to hear from a customer with a highly diverse and constantly changing environment that I mentioned earlier. They went through a major replatforming with Kubernetes on AWS. You're watching The CUBE, your leader in enterprise tech coverage. (upbeat music)
SUMMARY :
And with me to talk about a new offering Good to see you. but add a little color to that empowers developers into the process, to further address these problems. and the associated optimizations And Kubernetes is more of a And so that's the entire So I want you to sort And so for the first three or four years And so our biggest competitor is the VPA. So the business objectives are an input And so the thresholds in of the largest Drupal hosts. is because of the core machine learning and I go back to the and that specifically what So many of those even And the way we will get there As the observability vendors as that ecosystem continues to grow. Matt, thank you so much. to hear from a customer
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
AWS | ORGANIZATION | 0.99+ |
StormForge | ORGANIZATION | 0.99+ |
Matt Provo | PERSON | 0.99+ |
Matt | PERSON | 0.99+ |
Dynatrace | ORGANIZATION | 0.99+ |
Datadog | ORGANIZATION | 0.99+ |
UpDynamics | ORGANIZATION | 0.99+ |
Splunk | ORGANIZATION | 0.99+ |
mid last decade | DATE | 0.99+ |
four years | QUANTITY | 0.99+ |
first | QUANTITY | 0.98+ |
Kubernetes | TITLE | 0.98+ |
first three | QUANTITY | 0.98+ |
Prometheus | TITLE | 0.98+ |
EKS | ORGANIZATION | 0.98+ |
second core | QUANTITY | 0.98+ |
less than 1% | QUANTITY | 0.98+ |
DevOps | TITLE | 0.98+ |
Drupal | TITLE | 0.96+ |
one click | QUANTITY | 0.95+ |
both | QUANTITY | 0.95+ |
CNCF | ORGANIZATION | 0.92+ |
Datadog | TITLE | 0.91+ |
Intel | ORGANIZATION | 0.91+ |
CUBE | ORGANIZATION | 0.91+ |
two environments | QUANTITY | 0.9+ |
today | DATE | 0.9+ |
Finops | ORGANIZATION | 0.89+ |
The CUBE | TITLE | 0.84+ |
Kubernetes | ORGANIZATION | 0.81+ |
of years ago | DATE | 0.77+ |
single | QUANTITY | 0.76+ |
one | QUANTITY | 0.72+ |
CUBE con | EVENT | 0.68+ |
double | QUANTITY | 0.68+ |
Prometheus | ORGANIZATION | 0.67+ |
Nirvana | ORGANIZATION | 0.61+ |
Nirvana | TITLE | 0.48+ |
Matt Provo, StormForge
[Music] the adoption of container orchestration platforms is accelerating at a rate as fast or faster than any category in enterprise i.t survey data from enterprise technology research shows kubernetes specifically leads the pack in both spending velocity and market share now like virtualization in its early days containers bring many new performance and tuning challenges in particular ensuring consistent and predictable application performance is tricky especially because containers they're so flexible and they enable portability things are constantly changing devops pros have to wade through a sea of observability data and tuning the environment becomes a continuous exercise of trial and error this endless cycle taxes resources and kills operational efficiency so teams often just capitulate and simply dial up and throw unnecessary resources at the problem stormforge is a company founded mid last decade that is attacking these issues with a combination of machine learning and data analysis and with me to talk about a new offering that directly addresses these concerns is matt provo founder and ceo of stormforge matt welcome to the cube good to see you good to see you thanks for having me yeah so we saw you guys at a kubecon sort of first introduce you to our community but add a little color to my intro there yeah well you semi stole my thunder but uh i'm okay with that uh absolutely agree with everything you said in the intro um you know the the problem that we have set out to solve which is tailor-made for the use of real machine learning not machine learning kind of as a as a marketing tag uh is is connected to how workloads on kubernetes are are really managed from a resource efficiency standpoint and so a number of years ago we built uh the the core machine learning engine and have now turned that into a platform around how kubernetes resources are managed at scale and so organizations today as they're moving more workloads over uh sort of drink the kool-aid of the flexibility that comes with kubernetes and how many knobs you can turn and developers in many many ways love it once they start to operationalize the use of kubernetes and move uh workloads from pre-production into production they run into a pretty significant complexity wall and and this is where stormforge comes in to try to help them manage those resources more effectively in ensuring and implementing the right kind of automation that empowers developers into the process ultimately does not automate them out of it so you've got news yeah hard launch coming and to further address these problems tell us about that yeah so historically um uh you know like any machine learning engine we think about data inputs and what kind of data is going to feed our our system to be able to draw the appropriate insights out out for the user and so historically we are we've kind of been single threaded on load and performance tests in a pre-production environment and there's been a lot of adoption of that a lot of excitement around it and and frankly amazing results my vision has been uh for us to be able to close the loop however between uh data coming out of pre-production and opt in the associated optimizations and data coming out of production a production environment uh and and our ability to optimize that a lot of our users along the way have have said these results in pre-production are are fantastic how do i know they reflect reality of what my application is going to experience in a production environment and so we're super excited to to announce kind of the second core module for our platform called optimizelive the data input for that is uh observability and telemetry data coming out of apm platforms and and other data sources so this is like nirvana so i wonder if we could talk a little bit more about the the challenges that this address is i mean i've been around a while and it really have observed and i used to ask you know technology companies all the time okay so you're telling me beforehand what the optimal configuration should be and resource allocation what happens if something changes yeah and then it's always always a pause yeah and kubernetes is more of a rapidly changing environment than anything we've ever seen yeah so this is specifically the problem you're addressing maybe talk about that yeah so we view what happens in pre-production as sort of the experimentation phase and our machine learning is is allowing the user to experiment and design and scenario plan what we're doing uh with optimize live and adding the the production piece is uh what we kind of also call kind of our observation phase and so you need to be able to to to run the appropriate checks and balances between those two environments to ensure that what you're actually deploying and monitoring from an application performance from a cost standpoint is aligning with your slos and your slas as well as your business objectives and so that's the entire point of of this edition is to is to allow our users uh to experience uh hopefully the nirvana associated with that because it's an exciting er it's an exciting opportunity for them and really something that uh nobody else is doing from the standpoint of of closing that loop so you said upfront machine learning not as a marketing tag so i want you to sort of double click on that what's different than how other companies approach this problem yeah i mean part of it is a bias for me and a frustration as a founder of of the reason i started the company in the first place i think machine learning or ai gets tagged to a lot of stuff it's very buzz wordy it's it looks good i'm fortunate to have found a number of folks from the outset of the company with you know phds in applied mathematics and a focus on actually building real ai at the core uh that is connected to solving the right kind of actual business problems and so you know for the first three or four years of the company's history we really operated as a lab and that was our our focus we were we then decided we're trying to connect a fantastic team with differentiated technology to the right market timing and when we saw all these pain points around how fast the adoption of containers and kubernetes have taken place but the pain that the developers are running into we found it we actually found for ourselves uh that this was the perfect use case so how specifically does optimize live work can you add a little detail on that yeah so when you um many organizations today have an existing monitoring apm observability suite really in in place they've also got they've also got a metric source so this could be something like datadog or prometheus and once that data starts flowing there's an out of the box or or kind of a piece of kubernetes that ships with it called the vpa or the vertical pod auto scaler and uh less than really less than one percent of kubernetes users take advantage of the of the vpa mostly because it's really challenging to configure and it's not super compatible with the the tool set or the eco you know the ecosystem of tools uh in a kubernetes environment and so our biggest competitor is the vpa and what's happening in this environment or in in this world for developers is they're having to make decisions on on a number of different metrics or or resource elements typically things like memory and cpu and they have to decide what are the what are the limitations what are the requests i'm going to allow for this uh application and what are the limits so what are those thresholds that i'm going to be okay with so that i can again try to hit my business objectives and keep in line with my slas and to your earlier point in the intro it's often guesswork um you know they either have to rely on out of the box recommendations that ship with the databases and other services that they are using or it's a super manual process to go through and try to configure and tune this and so with optimize live we're making that one click and so we're continuously and consistently uh observing and watching the data that's flowing through these tools and we're serving back recommendations for the user they can choose to let those recommendations automatically patch and deploy or they can retain some semblance of control over the recommendations and manually deploy them into their environment themselves and we again really believe that the the user knows their application they know their the goals that they have we don't uh but we have a system that's smart enough to align with the business objectives and ultimately provide the relevant recommendations so the business objectives are an input from the application team yeah and then your system is smart enough to adapt and address those application over application right and and so the the thresholds in any given organization across their different ecosystem of apps or environment could be different the business objectives could be different and so we don't want to predefine that for people we want to give them the opportunity to build those thresholds in and then allow the machine learning to uh to learn and to send recommendations within those bounds and we're going to hear later from a customer who's uh hosting a drupal one of the largest drupal hosts so it's all do-it-yourself across thousands of customers so it's you know very unpredictable i want to make something clear though as to where you fit in the ecosystem you're not an observability platform you leverage observability platforms right so talk about that and where you fit in into the ecosystem yeah so that's a great point um we uh we're also you know a series b startup and and growing where we've made the choice to be very intentionally focused on the problems that we've solved and we've uh chosen to partner or integrate otherwise and so we do get put into the apm category from from time to time we're really an intelligence platform and that intelligence and insights that we're able to draw is because we because of the core machine learning we've built over the years and we also don't want organizations or users to have to switch from tools and investments that they've already made and so we were never going to we were never going to catch up to to to datadog or dynatrace or or splunk or app dynamics or some of the other and and we're totally fine with that they've got great market share and penetration they they do solve real problems instead we felt like users would want a seamless integration uh into the the tools they're already using and so we we we view ourselves as uh kind of the intel inside uh for that kind of a scenario and uh it takes observability and apm data and insights that were somewhat reactive uh they're visualized and somewhat reactive and and we make those uh we add that we add that proactive nature onto it the insights and and ultimately the the appropriate level of automation so when i think matt about cloud native and i go back to the sort of origins of cncf it was a handful of companies and now you look at the participants it'll you know make your eyes bleed how do you address dealing with all those companies and what are the what's the partnership strategy yeah it's so interesting because um it's just that even that cncf landscape has exploded um it was not too long ago where it was as small or smaller than the finnops landscape today which by the way the phenops pieces is also on a neck-breaking you know growth curve we i do see although there are a lot of companies and a lot of tools we're starting to see a significant amount of consistency or hardening of the tool chain uh you know for with our customers and end users and so we've made strategic and intentional decisions on deep partnerships in some cases like oem uh uses of our technology and and certainly you know intelligent and seamless integrations uh into a few so you know we're we'll be announcing uh a really exciting partnership with aws uh and and uh specifically what they're doing with eks their their kubernetes distribution and services we've got a deep partnership and integration with datadog and then with prometheus and specifically cloud provider a few other cloud providers that are operating managed prometheus environments okay so where do you want to take this thing it's not you're not taking the observability guys head on smart move so many of those even entering the market now but what is the vision yeah so we've had this debate a lot as well because it's super difficult to create a category uh you know on one hand um you know you know i have a lot of respect for founders and and companies that do that on the other hand um from a market timing standpoint you know we fit into ai ops that's really where we fit um you know we are we've made a bet on the future of kubernetes uh and and what that's going to look like and so um from a containers and kubernetes standpoint that's our bet uh but we're an aiops platform you know we'll continue getting better at what at the problems we solve with machine learning and we'll continue adding data inputs so we'll go you know we'll go beyond the application layer which is really where we play now we'll add kind of whole cluster optimization capabilities across across the full stack and the way we'll get there is by continuing to add different data inputs that make sense across the different layers of the stack and it's exciting we can stay vertically oriented on the problems that we're really good at solving but we can become more applicable and compatible over time so that's your next concentric circle as the observability vendors expand their observation space you can just play right into that yeah more data you get because you're a purpose built to solving these types of problems yeah so you can imagine a world right now out of observability we're taking things like telemetry data pretty quickly you can imagine a world where we take traces and logs and other data inputs as as that ecosystem continues to grow it just feeds our own uh you know we are reliant on data um so excellent matt thank you so much appreciate you for having me okay keep it right there in a moment we're gonna hear from a customer with a highly diverse and constantly changing environment that i mentioned earlier they went through a major re-platforming with kubernetes on aws you're watching thecube your leader in enterprise tech coverage [Music] you
SUMMARY :
the tool set or the eco you know the
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Jim | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
John | PERSON | 0.99+ |
Jeff | PERSON | 0.99+ |
Paul Gillin | PERSON | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
David | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
PCCW | ORGANIZATION | 0.99+ |
Dave Volante | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Michelle Dennedy | PERSON | 0.99+ |
Matthew Roszak | PERSON | 0.99+ |
Jeff Frick | PERSON | 0.99+ |
Rebecca Knight | PERSON | 0.99+ |
Mark Ramsey | PERSON | 0.99+ |
George | PERSON | 0.99+ |
Jeff Swain | PERSON | 0.99+ |
Andy Kessler | PERSON | 0.99+ |
Europe | LOCATION | 0.99+ |
Matt Roszak | PERSON | 0.99+ |
Frank Slootman | PERSON | 0.99+ |
John Donahoe | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Dan Cohen | PERSON | 0.99+ |
Michael Biltz | PERSON | 0.99+ |
Dave Nicholson | PERSON | 0.99+ |
Michael Conlin | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Melo | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
NVIDIA | ORGANIZATION | 0.99+ |
Joe Brockmeier | PERSON | 0.99+ |
Sam | PERSON | 0.99+ |
Matt | PERSON | 0.99+ |
Jeff Garzik | PERSON | 0.99+ |
Cisco | ORGANIZATION | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Joe | PERSON | 0.99+ |
George Canuck | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Apple | ORGANIZATION | 0.99+ |
Rebecca Night | PERSON | 0.99+ |
Brian | PERSON | 0.99+ |
Dave Valante | PERSON | 0.99+ |
NUTANIX | ORGANIZATION | 0.99+ |
Neil | PERSON | 0.99+ |
Michael | PERSON | 0.99+ |
Mike Nickerson | PERSON | 0.99+ |
Jeremy Burton | PERSON | 0.99+ |
Fred | PERSON | 0.99+ |
Robert McNamara | PERSON | 0.99+ |
Doug Balog | PERSON | 0.99+ |
2013 | DATE | 0.99+ |
Alistair Wildman | PERSON | 0.99+ |
Kimberly | PERSON | 0.99+ |
California | LOCATION | 0.99+ |
Sam Groccot | PERSON | 0.99+ |
Alibaba | ORGANIZATION | 0.99+ |
Rebecca | PERSON | 0.99+ |
two | QUANTITY | 0.99+ |
Matt Provo | ** Do not make public **
(bright upbeat music) >> The adoption of container orchestration platforms is accelerating at a rate as fast or faster than any category in enterprise IT. Survey data from Enterprise Technology Research shows Kubernetes specifically leads the pack in both spending velocity and market share. Now like virtualization in its early days, containers bring many new performance and tuning challenges. In particular, ensuring consistent and predictable application performance is tricky especially because containers they're so flexible and the enabled portability things are constantly changing. DevOps pros have to wade through a sea of observability data and tuning the environment becomes a continuous exercise of trial and error. This endless cycle taxes, resources, and kills operational efficiencies so teams often just capitulate and simply dial up and throw unnecessary resources at the problem. StormForge is a company founded in mid last decade that is attacking these issues with a combination of machine learning and data analysis. And with me to talk about a new offering that directly addresses these concerns, is Matt Provo, founder and CEO of StormForge. Matt, welcome to thecube. Good to see you. >> Good to see you, thanks for having me. >> Yeah. So we saw you guys at CubeCon, sort of first introduce you to our community but add a little color to my intro if you will. >> Yeah, well you semi stole my thunder but I'm okay with that. Absolutely agree with everything you said in the intro. You know, the problem that we have set out to solve which is tailor made for the use of real machine learning not machine learning kind of as a marketing tag is connected to how workloads on Kubernetes are really managed from a resource efficiency standpoint. And so a number of years ago we built the core machine learning engine and have now turned that into a platform around how Kubernetes resources are managed at scale. And so organizations today as they're moving more workloads over sort of drink the Kool-Aid of the flexibility that comes with Kubernetes and how many knobs you can turn and developers in many ways love it. Once they start to operationalize the use of Kubernetes and move workloads from pre-production into production, they run into a pretty significant complexity wall. And this is where StormForge comes in to try to help them manage those resources more effectively in ensuring and implementing the right kind of automation that empowers developers into the process ultimately does not automate them out of it. >> So you've got news, your hard launch coming in to further address these problems. Tell us about that. >> Yeah so historically, you know, like any machine learning engine, we think about data inputs and what kind of data is going to feed our system to be able to draw the appropriate insights out for the user. And so historically we are, we've kind of been single-threaded on load and performance tests in a pre-production environment. And there's been a lot of adoption of that, a lot of excitement around it and frankly, amazing results. My vision has been for us to be able to close the loop however between data coming out of pre-production and the associated optimizations and data coming out of production, a production environment, and our ability to optimize that. A lot of our users along the way have said these results in pre-production are fantastic. How do I know they reflect reality of what my application is going to experience in a production environment? And so we're super excited to announce kind of the second core module for our platform called Optimize Live. The data input for that is observability and telemetry data coming out of APM platforms and other data sources. >> So this is like Nirvana. So I wonder if we could talk a little bit more about the challenges that this addresses. I mean, I've been around a while and it really have observed and I used to ask technology companies all the time, okay, so you're telling me beforehand what the optimal configuration should be in resource allocation, what happens if something changes? And then it's always a pause. And Kubernetes is more of a rapidly changing environment than anything we've ever seen. So this is specifically the problem you're addressing. Maybe talk about that a little bit. >> Yeah so we view what happens in pre-production as sort of the experimentation phase and our machine learning is allowing the user to experiment and scenario plan. What we're doing with Optimize Live and adding the production piece is what we kind of also call kind of our observation phase. And so you need to be able to run the appropriate checks and balances between those two environments to ensure that what you're actually deploying and monitoring from an application performance, from a cost standpoint, is aligning with your SLOs and your SLAs as well as your business objectives. And so that's the entire point of this addition is to allow our users to experience hopefully the Nirvana associated with that because it's an exciting opportunity for them and really something that nobody else is doing from the standpoint of closing that loop. >> So you said upfront machine learning not as a marketing tag. So I want you to sort of double click on that. What's different than how other companies approach this problem? >> Yeah I mean, part of it is a bias for me and a frustration as a founder of the reason I started the company in the first place. I think machine learning our AI gets tagged to a lot of stuff. It's very buzzwordy, it looks good. I'm fortunate to have found a number of folks from the outset of the company with, you know, PhDs in Applied Mathematics and a focus on actually building real AI at the core that is connected to solving the right kind of actual business problems. And so, you know, for the first three or four years of the company's history, we really operated as a lab and that was our focus. We then decided we're trying to connect a fantastic team with differentiated technology to the right market timing. And when we saw all of these pain points around how fast the adoption of containers and Kubernetes have taken place but the pain that the developers are running into, we found it, we actually found for ourselves that this was the perfect use case. >> So how specifically does Optimize Live work? Can you add a little detail on that? >> Yeah so when you, many organizations today have an existing monitoring APM observability suite really in place. They've also got, they've also got a metric source, so this could be something like Datadog or Prometheus. And once that data starts flowing, there's an out of the box or kind of a piece of Kubernetes that ships with it called the VPA or the Vertical Pod Autoscaler. And less than really less than 1% of Kubernetes users take advantage of the VPA mostly because it's really challenging to configure and it's not super compatible with the tool set or the, you know, the ecosystem of tools in a Kubernetes environment. And so our biggest competitor is the VPA. And what's happening in this environment or in this world for developers is they're having to make decisions on a number of different metrics or resource elements typically things like memory and CPU. And they have to decide what are the, what are the requests I'm going to allow for this application and what are the limits? So what are those thresholds that I'm going to be okay with? So that I can again try to hit my business objectives and keep in line with my SLAs. And to your earlier point in the intro, it's often guesswork. You know, they either have to rely on out of the box recommendations that ship with the databases and other services that they are using or it's a super manual process to go through and try to configure and tune this. And so with Optimize Live, we're making that one-click. And so we're continuously and consistently observing and watching the data that's flowing through these tools and we're serving back recommendations for the user. They can choose to let those recommendations automatically patch and deploy or they can retain some semblance of control over the recommendations and manually deploy them into their environment themselves. And we again, really believe that the user knows their application, they know the goals that they have, we don't. But we have a system that's smart enough to align with the business objectives and ultimately provide the relevant recommendations at that point. >> So the business objectives are an input from the application team and then your system is smart enough to adapt and adjust those. >> Application over application, right? And so the thresholds in any given organization across their different ecosystem of apps or environment could be different. The business objectives could be different. And so we don't want to predefine that for people. We want to give them the opportunity to build those thresholds in and then allow the machine learning to learn and to send recommendations within those bounds. >> And we're going to hear later from a customer who is hosting a Drupal, one of the largest Drupal host, is it? So it's all do it yourself across thousands of customers so it's very unpredictable. I want to make something clear though, as to where you fit in the ecosystem. You're not an observability platform, you leverage observability platforms, right? So talk about that and where you fit in into the ecosystem. >> Yeah so it's a great point. We, we're also you know, a series B startup and growing. We've made the choice to be very intentionally focused on the problems that we've solve and we've chosen to partner or integrate otherwise. And so we do get put into the APM category from time to time. We're really an intelligence platform. And that intelligence and insights that we're able to draw is because we, because of the core machine learning we've built over the years. And we also don't want organizations or users to have to switch from tools and investments that they've already made. And so we were never going to catch up to Datadog or Dynatrace or Splunk or AppDynamics or some of the other, and we're totally fine with that. They've got great market share and penetration and they do solve real problems. Instead, we felt like users would want a seamless integration into the tools they're already using. And so we view ourselves as kind of the Intel inside for that kind of a scenario. And it takes observability and APM data and insights that were somewhat reactive, they're visualized and somewhat reactive and we make those, we add that proactive nature onto it, the insights and ultimately the appropriate level of automation. >> So when I think Matt about cloud native and I go back to the sort of origins of CNCF, it was a, you know, handful of companies, and now you look at the participants, you know, make your eyes bleed. How do you address dealing with all those companies and what's the partnership strategy? >> Yeah it's so interesting because it's just that even at CNCF landscape has exploded. It was not too long ago where it was as smaller than the finOps Landscape today which by the way the FinOps pieces is also on a neck breaking, you know, growth curve. We, I do see although there are a lot of companies and a lot of tools, we're starting to see a significant amount of consistency or hardening of the tool chain with our customers and users. And so we've made strategic and intentional decisions on deep partnerships in some cases like OEM users of our technology and certainly, you know, intelligent and seamless integrations into a few. So, you know, we'll be announcing a really exciting partnership with AWS and specifically what they're doing with EKS, their Kubernetes distribution and services. We've got a deep partnership and integration with Datadog and then with Prometheus and specifically cloud provider, a few other cloud providers that are operating manage Prometheus environments. >> Okay so where do you want to take this thing? If it's not, you're not taking the observability guys head on, smart move, so many of those even entering the market now, but what is the vision? >> Yeah so we've had this debate a lot as well because it's super difficult to create a category. You know, on one hand, I have a lot of respect for founders and companies that do that, on the other hand from a market timing standpoint, you know, we fit into AIOps. That's really where we fit. You know we are, we've made a bet on the future of Kubernetes and what that's going to look like. And so from a containers and Kubernetes standpoint that's our bet. But we're an AIOps platform, we'll continue getting better at what, at the problems we solve with machine learning and we'll continue adding data inputs so we'll go beyond the application layer which is really where we play now. We'll add kind of whole cluster optimization capabilities across the full stack. And the way we'll get there is by continuing to add different data inputs that make sense across the different layers of the stack and it's exciting. We can stay vertically oriented on the problems that we're really good at solving but we become more applicable and compatible over time. >> So that's your next concentric circle. As the observability vendors expand their observation space you can just play right into that. The more data you get could be because you're purpose built to solving these types of problems. >> Yeah so you can imagine a world right now out of observability, we're taking things like telemetry data pretty quickly. You can imagine a world where we take traces and logs and other data inputs as that ecosystem continues to grow, it just feeds our own, you know, we are reliant on data. So. >> Excellent. Matt, thank you so much. Thanks for hoping on. >> Yeah, appreciate it. >> Okay. Keep it right there. In a moment, We're going to hear from a customer with a highly diverse and constantly changing environment that I mentioned earlier, they went through a major re-platforming with Kubernetes on AWS. You're watching theCube, your a leader in enterprise tech coverage. (bright music)
SUMMARY :
and the enabled portability to my intro if you will. and how many knobs you can turn to further address these problems. and the associated optimizations about the challenges that this addresses. And so that's the entire So I want you to sort and that was our focus. And so our biggest competitor is the VPA. So the business objectives are an input And so the thresholds in as to where you fit in the ecosystem. We've made the choice to be and I go back to the and certainly, you know, And the way we'll get there As the observability vendors and other data inputs as that Matt, thank you so much. We're going to hear from a customer
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
AWS | ORGANIZATION | 0.99+ |
Matt | PERSON | 0.99+ |
StormForge | ORGANIZATION | 0.99+ |
Matt Provo | PERSON | 0.99+ |
Dynatrace | ORGANIZATION | 0.99+ |
Datadog | ORGANIZATION | 0.99+ |
first | QUANTITY | 0.99+ |
Splunk | ORGANIZATION | 0.99+ |
thousands | QUANTITY | 0.99+ |
CNCF | ORGANIZATION | 0.99+ |
AppDynamics | ORGANIZATION | 0.99+ |
Kubernetes | TITLE | 0.98+ |
one-click | QUANTITY | 0.98+ |
four years | QUANTITY | 0.98+ |
first three | QUANTITY | 0.98+ |
two environments | QUANTITY | 0.98+ |
Prometheus | TITLE | 0.97+ |
EKS | ORGANIZATION | 0.97+ |
DevOps | TITLE | 0.97+ |
mid last decade | DATE | 0.97+ |
both | QUANTITY | 0.96+ |
Drupal | TITLE | 0.96+ |
Kool-Aid | ORGANIZATION | 0.93+ |
today | DATE | 0.91+ |
Enterprise | ORGANIZATION | 0.91+ |
second core module | QUANTITY | 0.9+ |
Optimize Live | TITLE | 0.85+ |
Datadog | TITLE | 0.84+ |
less than 1% | QUANTITY | 0.84+ |
Live | TITLE | 0.83+ |
Kubernetes | ORGANIZATION | 0.8+ |
of years ago | DATE | 0.8+ |
one | QUANTITY | 0.79+ |
less | QUANTITY | 0.76+ |
Intel | ORGANIZATION | 0.75+ |
CubeCon | EVENT | 0.69+ |
FinOps | TITLE | 0.65+ |
finOps Landscape | TITLE | 0.59+ |
double | QUANTITY | 0.58+ |
Optimize Live | ORGANIZATION | 0.57+ |
AIOps | ORGANIZATION | 0.56+ |
AIOps | TITLE | 0.54+ |
theCube | TITLE | 0.5+ |
Prometheus | ORGANIZATION | 0.49+ |
Nirvana | TITLE | 0.41+ |
Nirvana | ORGANIZATION | 0.27+ |
Webb Brown & Alex Thilen, Kubecost | AWS Startup Showcase S2 E1 | Open Cloud Innovations
>>Hi, everyone. Welcome to the cubes presentation of the eight of us startup showcase open cloud innovations. This is season two episode one of the ongoing series covering the exciting startups from ABC ecosystems today. Uh, episode one, steam is the open source community and open cloud innovations. I'm Sean for your host got two great guests, Webb brown CEO of coop costs and as Thielen, head of business development, coop quest, gentlemen, thanks for coming on the cube for the showcase 80, but startups. >>Thanks for having a Sean. Great to be back, uh, really excited for the discussion we have here. >>I keep alumni from many, many coupons go. You guys are in a hot area right now, monitoring and reducing the Kubernetes spend. Okay. So first of all, we know one thing for sure. Kubernetes is the hottest thing going on because of all the benefits. So take us through you guys. Macro view of this market. Kubernetes is growing, what's going on with the company. What is your company's role? >>Yeah, so we've definitely seen this growth firsthand with our customers in addition to the broader market. Um, you know, and I think we believe that that's really indicative of the value that Kubernetes provides, right? And a lot of that is just faster time to market more scalability, improved agility for developer teams and, you know, there's even more there, but it's a really exciting time for our company and also for the broader cloud native community. Um, so what that means for our company is, you know, we're, we're scaling up quickly to meet our users and support our users, every, you know, metric that our company's grown about four X over the last year, including our team. Um, and the reason that one's the most important is just because, you know, the, the more folks and the larger that our company is, the better that we can support our users and help them monitor and reduce those costs, which ultimately makes Kubernetes easier to use for customers and users out there on the market. >>Okay. So I want to get into why Kubernetes is costing so much. Obviously the growth is there, but before we get there, what is the background? What's the origination story? Where did coop costs come from? Obviously you guys have a great name costs. Qube you guys probably reduced costs and Kubernetes great name, but what's the origination story. How'd you guys get here? What HR you scratching? What problem are you solving? >>So yeah, John, you, you guessed it, uh, you know, oftentimes the, the name is a dead giveaway there where we're cost monitoring cost management solutions for Kubernetes and cloud native. Um, and backstory here is our founding team was at Google before starting the company. Um, we were working on infrastructure monitoring, um, both on internal infrastructure, as well as Google cloud. Um, we had a handful of our teammates join the Kubernetes effort, you know, early days. And, uh, we saw a lot of teams, you know, struggling with the problems we're solving. We were solving internally at Google and we're we're solving today. Um, and to speak to those problems a little bit, uh, you know, you, you, you touched on how just scale alone is making this come to the forefront, right. You know, there's now many billions of dollars being spent on CU, um, that is bringing this issue, uh, to make it a business critical questions that is being asked in lots of organizations. Um, you know, that combined with, you know, the dynamic nature and complexity of Kubernetes, um, makes it really hard to manage, um, you know, costs, uh, when you scale across a very large organization. Um, so teams turned to coop costs today, you know, thousands of them do, uh, to get monitoring in place, you know, including alerts, recurring reports and like dynamic management insights or automation. >>Yeah. I know we talked to CubeCon before Webb and I want to come back to the problem statement because when you have these emerging growth areas that are really relevant and enabling technologies, um, you move to the next point of failure. And so, so you scaling these abstraction layers. Now services are being turned on more and more keeping it as clusters are out there. So I have to ask you, what is the main cost driver problem that's happening in the cube space that you guys are addressing? Is it just sheer volume? Is it different classes of services? Is it like different things are kind of working together, different monitoring tools? Is it not a platform and take us through the, the problem area? What do you guys see this? >>Yeah, the number one problem area is still actually what, uh, the CNCF fin ops survey highlighted earlier this year, um, which is that approximately two thirds of companies still don't have kind of baseline to visibility into spend when they moved to Kubernetes. Um, so, you know, even if you had a really complex, you know, chargeback program in place, when you're building all your applications on BMS, you move to Kubernetes and most teams again, can't answer these really simple questions. Um, so we're able to give them that visibility in real time, so they can start breaking these problems down. Right. They can start to see that, okay, it's these, you know, the deployments are staple sets that are driving our costs or no, it's actually, you know, these workloads that are talking to, you know, S3 buckets and, you know, really driving, you know, egress costs. Um, so it's really about first and foremost, just getting the visibility, getting the eyes and ears. We're able to give that to teams in real time at the largest scale Kubernetes clusters in the world. Um, and again, most teams, when they first start working with us, don't have that visibility, not having that visibility can have a whole bunch of downstream impacts, um, including kind of not getting, you know, costs right. You know, performance, right. Et cetera. >>Well, let's get into that downstream benefit, uh, um, problems and or situations. But the first question I have just throw naysayer comment at you would be like, oh, wait, I have all this cost monitoring stuff already. What's different about Kubernetes. Why what's what's the problem I can are my other tool is going to work for me. How do you answer that one? >>Yeah. So, you know, I think first and foremost containers are very dynamic right there. They're often complex, often transient and consume variable cluster resources. And so as much as this enables teams to contract construct powerful solutions, um, the associated costs and actually tracking those, those different variables can be really difficult. And so that's why we see why a solution like food costs. That's purpose built for developers using Kubernetes is really necessary because some of those older, you know, traditional cloud cost optimization tools are just not as fit for, for this space specifically. >>Yeah. I think that's exactly right, Alex. And I would add to that just the way that software is being architected deployed and managed is fundamentally changing with Kubernetes, right? It is deeply impacting every part of scifi software delivery process. And through that, you know, decisions are getting made and, you know, engineers are ultimately being empowered, um, to make more, you know, costs impacting decisions. Um, and so we've seen, you know, organizations that get real time kind of built for Kubernetes are built for cloud native, um, benefit from that massively throughout their, their culture, um, you know, cost performance, et cetera. >>Uh, well, can you just give a quick example because I think that's a great point. The architectures are shifting, they're changing there's new things coming in, so it's not like you can use an old tool and just retrofit it. That's sometimes that's awkward. What specific things you see changing with Kubernetes that's that environments are leveraging that's good. >>Yeah. Yeah. Um, one would be all these Kubernetes primitives are concepts that didn't exist before. Right. So, um, you know, I'm not, you know, managing just a generic workload, I'm managing a staple set and, or, you know, three replica sets. Right. And so having a language that is very much tailored towards all of these Kubernetes concepts and abstractions, et cetera. Um, but then secondly, it was like, you know, we're seeing this very obvious, you know, push towards microservices where, you know, typically again, you're shipping faster, um, you know, teams are making more distributed or decentralized decisions, uh, where there's not one single point where you can kind of gate check everything. Um, and that's a great thing for innovation, right? We can move much faster. Um, but for some teams, um, you know, not using a tool like coop costs, that means sacrificing having a safety net in place, right. >>Or guard rails in place to really help manage and monitor this. And I would just say, lastly, you know, uh, a solution like coop costs because it's built for Kubernetes sits in your infrastructure, um, it can be deployed with a single helmet stall. You don't have to share any data remotely. Um, but because it's listening to your infrastructure, it can give you data in real time. Right. And so we're moving from this world where you can make real time automated decisions or manual decisions as opposed to waiting for a bill, you know, a day, two days or a week later, um, when it may be already too late, you know, to avoid, >>Or he got the extra costs and you know what, he wants that. And he got to fight for a refund. Oh yeah. I threw a switch or wasn't paying attention or human error or code because a lot of automation is going on. So I could see that as a benefit. I gotta, I gotta ask the question on, um, developer uptake, because develop, you mentioned a good point. There that's another key modern dynamic developers are in, in the moment making decisions on security, on policy, um, things to do in the CIC D pipeline. So if I'm a developer, how do I engage with Qube cost? Do I have to, can I just download something? Is it easy? How's the onboarding process for your customers? >>Yeah. Great, great question. Um, so, you know, first and foremost, I think this gets to the roots of our company and the roots of coop costs, which is, you know, born in open-source, everything we do is built on top of open source. Uh, so the answer is, you know, you can go out and install it in minutes. Like, you know, thousands of other teams have, um, it is, you know, the, the recommended route or preferred route on our side is, you know, a helm installed. Um, again, you don't have to share any data remotely. You can truly not lock down, you know, namespace eat grass, for example, on the coop cost namespace. Um, and yeah, and in minutes you'll have this visibility and can start to see, you know, really interesting metrics that, again, most teams, when we started working with them, either didn't have them in place at all, or they had a really rough estimate based on maybe even a coop cost Scruff on a dashboard that they installed. >>How does cube cost provide the visibility across the environment? How do you guys actually make it work? >>Yeah, so we, you know, sit in your infrastructure. Um, we have integrations with, um, for on-prem like custom pricing sheets, uh, with card providers will integrate with your actual billing data, um, so that we can, uh, listen for events in your infrastructure, say like a nude node coming up, or a new pod being scheduled, et cetera. Um, we take that information, join with your billing data, whether it's on-prem or in one of the big three cloud providers. And then again, we can, in real time tell you the cost of, you know, any dimension of your infrastructure, whether it's one of the backing, you know, virtual assets you're using, or one of the application dimensions like a label or annotation namespace, you know, pod container, you name it >>Awesome. Alex, what's your take on the landscape with, with the customers as they look the cost reductions. I mean, everyone loves cost reductions as a, certainly I love the safety net comment that Webb made, but at the end of the day, Kubernetes is not so much a cost driver. It's more of a, I want the modern apps faster. Right? So, so, so people who are buying Kubernetes usually aren't price sensitive, but they also don't want to get gouged either on mistakes. Where is the customer path here around Kubernetes cost management and reduction and a scale? >>Yeah. So I think one thing that we're looking forward to hearing this upcoming year, just like we did last year is continuing to work with the various tools that customers are already using and, you know, meeting those customers where they are. So some examples of that are, you know, working with like CICT tools out there. Like we have a great integration with armoring Spinnaker to help customers actually take the insights from coop costs and deploy those, um, in a more efficient manner. Um, we're also working with a lot of partners, like, you know, for fauna to help customers visualize our data and, you know, integrate with or rancher, which are management platforms for Kubernetes. And all of that I think is just to make cost come more to the forefront of the conversation when folks are using Kubernetes and provide that, that data to customers and all the various tools that they're using across the ecosystem. Um, so I think we really want to surface this and make costs more of a first-class citizen across, you know, the, the ecosystem and then the community partners. >>What's your strategy of the biz dev side. As you guys look at a growing ecosystem with CubeCon CNCF, you mentioned that earlier, um, the community is growing. It's always been growing fast. You know, the number of people entering in are amazing, but now that we start going, you know, the S curves kicking in, um, integration and interoperability and openness is always a key part of company success. What's Qube costs is vision on how you're going to do biz dev going forward. >>Absolutely. So, you know, our products opensource that is deeply important to our company, we're always going to continue to drive innovation on our open source product. Um, as Webb mentioned, you know, we have thousands of teams that are, that are using our product. And most of that is actually on the free, but something that we want to make sure continues to be available for the community and continue to bring that development for the community. And so I think a part of that is making sure that we're working with folks not just on the commercial side, but also those open source, um, types of products, right? So, you know, for Fanta is open source Spinnaker's are open source. I think a lot of the biz dev strategies just sticking to our roots and make sure that we continue to drive it a strong open source presence and product for, for our community of users, keep that >>And a, an open source and commercial and keep it stable. Well, I got to ask you, obviously, the wave is here. I always joke, uh, going back. I remember when the word Kubernetes was just kicked around pre uh, the OpenStack days many, many years ago. It's the luxury of being a old cube guy that I am 11 years doing the cube, um, all fun. But if we remember talking to him in the early days, is that with Kubernetes was, if, if it worked, the, the phrase was rising, tide floats all boats, I would say right now, the tides rising pretty well right now, you guys are in a good spot with the cube costs. Are there areas that you see coming where cost monitoring, um, is going to expand more? Where do you see the Kubernetes? Um, what's the aperture, if you will, of the, of the cost monitoring space at your end that you think you can address. >>Yeah, John, I think you're exactly right. This, uh, tide has risen and it just keeps riding rising, right? Like, um, you know, the, the sheer number of organizations we use C using Kubernetes at massive scale is just mind blowing at this point. Um, you know, what we see is this really natural pattern for teams to start using a solution like coop costs, uh, start with, again, either limited or no visibility, get that visibility in place, and then really develop an action plan from there. And that could again be, you know, different governance solutions like alerts or, you know, management reports or, you know, engineering team reports, et cetera. Um, but it's really about, you know, phase two of taking that information and really starting to do something with it. Right. Um, we, we are seeing and expect to see more teams turn to an increasing amount of, of automation to do that. Um, but ultimately that is, uh, very much after you get this baseline highly accurate, uh, visibility that you feel very comfortable making, potentially critical, very critical related to reliability, performance decisions within your infrastructure. >>Yeah. I think getting it right key, you mentioned baseline. Let me ask you a quick follow-up on that. How fast can companies get there when you say baseline, there's probably levels of baseline. Obviously all environments are different now. Not all one's the same, but what's just anecdotally you see, as that baseline, how fast we will get there, is there a certain minimum viable configuration or architecture? Just take us through your thoughts on that. >>Yeah. Great question. It definitely depends on organizational complexity and, you know, can depend on applicational application complexity as well. But I would say most importantly is, um, you know, the, the array of cost centers, departments, you know, complexity across the org as opposed to, you know, technological. Um, so I would say for, you know, less complex organizations, we've seen it happen in, you know, hours or, you know, a day less, et cetera. Um, because that's, you know, one or two or a smaller engineering games, they can share that visibility really quickly. And, um, you know, they may be familiar with Kubernetes and they just get it right away. Um, for larger organizations, we've seen it take kind of up 90 days where it's really about infusing this kind of into their DNA. When again, there may not have been a visibility or transparency here before. Um, again, I think the, the, the bulk of the time there is really about kind of the cultural element, um, and kind of awareness building, um, and just buy in throughout the organization. >>Awesome. Well, guys got a great product. Congratulations, final question for both of you, it's early days in Kubernetes, even though the tide is rising, keeps rising, more boats are coming in. Harbor is getting bigger, whatever, whatever metaphor you want to use, it's really going great. You guys are seeing customer adoption. We're seeing cloud native. I was told that my friends at dock or the container side is going crazy as well. Everything's going great in cloud native. What's the vision on the innovation? How do you guys continue to push the envelope on value in open source and in the commercial area? What's the vision? >>Yeah, I think there's, there's many areas here and I know Alex will have more to add here. Um, but you know, one area that I know is relevant to his world is just more, really interesting integrations, right? So he mentioned coop costs, insights, powering decisions, and say Spinnaker, right? I think more and more of this tool chain really coming together and really seeing the benefits of all this interoperability. Right. Um, so that I think combined with, uh, just more and more intelligence and automation being deployed again, that's only after the fact that teams are really comfortable with his decisions and the information and the decisions that are being made. Um, but I think that increasingly we see the community again, being ready to leverage this information and really powerful ways. Um, just because, you know, as teams scale, there's just a lot to manage. And so a team, you know, leveraging automation can, you know, supercharge them and in really impactful ways. >>Awesome, great integration integrations, Alex, expand on that. A whole different kind of set of business development integrations. When you have lots of tool chains, lots of platforms and tools kind of coming together, sharing data, working together, automating together. >>Well. Yeah, we, so I think it's going to be super important to keep a pulse on the new tools. Right. Make sure that we're on the forefront of what customers are using and just continuing to meet them where they are. And a lot of that honestly, is working with AWS too, right? Like they have great services and EKS and managed Prometheus's. Um, so we want to make sure that we continue to work with that team and support their services as that launched as well. >>Great stuff. I got a couple of minutes left. I felt I'll throw one more question in there since I got two great experts here. Um, just, you know, a little bit change of pace, more of an industry question. That's really no wrong answer, but I'd love to get your reaction to, um, the SAS conversation cloud has changed what used to be SAS. SAS was, oh yeah. Software as a service. Now that you have all these kinds of new kinds of you have automation, horizontally, scalable cloud and edge, you now have vertical machine learning. Data-driven insights. A lot of things in the stack are changing. So the question is what's the new SAS look like it's the same as the old SAS? Or is it a new kind of refactoring of what SAS is? What's your take on this? >>Yeah. Um, there's a web, please jump in here wherever. But in, in my view, um, it's a spectrum, right? There's there's customers that are on both ends of this. Some customers just want a fully hosted, fully managed product that wouldn't benefit from the luxury of not having to do any, any sort of infrastructure management or patching or anything like that. And they just want to consume a great product. Um, on the other hand, there's other customers that have more highly regulated industries or security requirements, and they're going to need things to deploy in their environment. Um, right now QP cost is, is self hosted. But I think in the future, we want to make sure that, you know, we, we have versions of our product available for customers across that entire spectrum. Um, so that, you know, if somebody wants the benefit of just not having to manage anything, they can use a fully self hosted sat or a fully multitenant managed SAS, or, you know, other customers can use a self hosted product. And then there's going to be customers that are in the middle, right, where there's certain components that are okay to be a SAS or hosted elsewhere. But then there's going to be components that are really important to keep in their own environment. So I think, uh, it's really across the board and it's going to depend on customer and customer, but it's important to make sure we have options for all of them. >>Great guys, we have SAS, same as the old SAS. What's the SAS playbook. Now >>I think it is such a deep and interesting question and one that, um, it's going to touch so many aspects of software and on our lives, I predict that we'll continue to see this, um, you know, tension or real trade-off across on the one hand convenience. And now on the other hand, security, privacy and control. Um, and I think, you know, like Alex mentioned, you know, different organizations are going to make different decisions here based on kind of their relative trade-offs. Um, I think it's going to be of epic proportions. I think, you know, we'll look back on this period and just say that, you know, this was one of the foundational questions of how to get this right. We ultimately view it as like, again, we want to offer choice, um, and make, uh, make every choice be great, but let our users, uh, pick the right one, given their profile on those, on those streets. >>I think, I think it's a great comment choice. And also you got now dimensions of implementations, right? Multitenant, custom regulated, secure. I want have all these controls. Um, it's great. No one, no one SaaS rules the world, so to speak. So it's again, great, great dynamic. But ultimately, if you want to leverage the data, is it horizontally addressable? MultiTech and again, this is a whole nother ball game we're watching this closely and you guys are in the middle of it with cube costs, as you guys are creating that baseline for customers. Uh, congratulations. Uh, great to see you where thanks for coming on. Appreciate it. Thank you so much for having us again. Okay. Great. Conservation aiders startup showcase open cloud innovators here. Open source is driving a lot of value as it goes. Commercial, going to the next generation. This is season two episode, one of the AWS startup series with the cube. Thanks for watching.
SUMMARY :
as Thielen, head of business development, coop quest, gentlemen, thanks for coming on the cube for the showcase 80, Great to be back, uh, really excited for the discussion we have here. So take us through you guys. Um, you know, and I think we believe that that's really indicative of the value Obviously you guys have a great name costs. Um, you know, that combined with, you know, the dynamic nature and complexity of Kubernetes, And so, so you scaling these abstraction layers. you know, even if you had a really complex, you know, chargeback program in place, when you're building all your applications But the first question I have just throw naysayer comment at you would be like, oh, wait, I have all this cost monitoring you know, traditional cloud cost optimization tools are just not as fit for, for this space specifically. Um, and so we've seen, you know, organizations that get What specific things you see changing with Kubernetes that's Um, but for some teams, um, you know, not using a tool like coop costs, And I would just say, lastly, you know, uh, a solution like coop costs because it's built for Kubernetes Or he got the extra costs and you know what, he wants that. Uh, so the answer is, you know, you can go out and install it in minutes. Yeah, so we, you know, sit in your infrastructure. comment that Webb made, but at the end of the day, Kubernetes is not so much a cost driver. So some examples of that are, you know, working with like CICT you know, the S curves kicking in, um, integration and interoperability So, you know, our products opensource that is deeply important to our company, I would say right now, the tides rising pretty well right now, you guys are in a good spot with the Um, you know, what we see is this really natural pattern How fast can companies get there when you say baseline, there's probably levels of baseline. you know, complexity across the org as opposed to, you know, technological. How do you guys continue Um, but you know, one area that I know is relevant to his world is just more, When you have lots of tool chains, lots of platforms and tools kind Um, so we want to make sure that we continue to work with that team and Um, just, you know, a little bit change of pace, more of an industry question. But I think in the future, we want to make sure that, you know, we, What's the SAS playbook. Um, and I think, you know, like Alex mentioned, you know, we're watching this closely and you guys are in the middle of it with cube costs, as you guys are creating
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
John | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Alex Thilen | PERSON | 0.99+ |
Webb Brown | PERSON | 0.99+ |
11 years | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
Sean | PERSON | 0.99+ |
thousands | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
Thielen | PERSON | 0.99+ |
Alex | PERSON | 0.99+ |
last year | DATE | 0.99+ |
eight | QUANTITY | 0.99+ |
Kubecost | PERSON | 0.99+ |
Webb | PERSON | 0.99+ |
90 days | QUANTITY | 0.99+ |
Webb brown | PERSON | 0.99+ |
ABC | ORGANIZATION | 0.99+ |
both | QUANTITY | 0.99+ |
first question | QUANTITY | 0.99+ |
CNCF | ORGANIZATION | 0.98+ |
Kubernetes | ORGANIZATION | 0.98+ |
CubeCon | ORGANIZATION | 0.98+ |
two great guests | QUANTITY | 0.98+ |
first | QUANTITY | 0.98+ |
both ends | QUANTITY | 0.97+ |
Kubernetes | TITLE | 0.97+ |
two great experts | QUANTITY | 0.96+ |
one more question | QUANTITY | 0.96+ |
a day | QUANTITY | 0.96+ |
single helmet | QUANTITY | 0.94+ |
earlier this year | DATE | 0.94+ |
today | DATE | 0.94+ |
secondly | QUANTITY | 0.94+ |
one thing | QUANTITY | 0.93+ |
S3 | COMMERCIAL_ITEM | 0.92+ |
Fanta | ORGANIZATION | 0.92+ |
Qube | ORGANIZATION | 0.91+ |
a week later | DATE | 0.91+ |
Kubernetes | PERSON | 0.91+ |
SAS | ORGANIZATION | 0.9+ |
season two episode | QUANTITY | 0.88+ |
approximately two thirds | QUANTITY | 0.87+ |
about four X | QUANTITY | 0.87+ |
coop | ORGANIZATION | 0.85+ |
three replica sets | QUANTITY | 0.85+ |
EKS | ORGANIZATION | 0.85+ |
billions of dollars | QUANTITY | 0.84+ |
80 | QUANTITY | 0.81+ |
two days | QUANTITY | 0.8+ |
single point | QUANTITY | 0.8+ |
one area | QUANTITY | 0.77+ |
season two | QUANTITY | 0.76+ |
BMS | TITLE | 0.76+ |
OpenStack | TITLE | 0.75+ |
Parasar Kodati & David Noy | KubeCon + CloudNativeCon NA 2021
>>mhm mhm >>Hey guys, welcome back to Los Angeles lisa martin. Coming to you live from cuba con and cloud native Con north America 2021. Very excited to be here. This is our third day of back to back coverage on the cube and we've got a couple of guests cube alumni joining me remotely. Please welcome parse our karate senior consultant, product marketing, Dell Technologies and David Noi VP product management at Dell Technologies. Gentlemen welcome back to the program. >>Thanks johnny >>so far so let's go ahead and start with you. Let's talk about what Dell EMC is offering to developers today in terms of unstructured data. >>Absolutely, it's great to be here. So let me start with the container storage interface. This is Q khan and a couple of years ago the container storage interface was still in beta and the storage vendors, we're very enthusiastically kind of building the plug in city of the different storage portfolio to offer enterprise grade features to developers are building applications of the Cuban this platform. And today if you look at the deli in storage portfolio, big block volumes. Nash shares s three object A P I S beyond their virtual volumes. However you're consuming storage, you have the plug ins that are required to run your applications with these enterprise Great feature speech right about snap sharks data replication, all available in the Cuban this layer and just this week at coupon we announced the container storage modules which is kind of the next step of productivity for developers beat you know uh in terms of observe ability of the storage metrics using tools like Prometheus visualizing it ravana authorization capabilities so that you know too bad moments can have better resource management of the storage that is being consumed um that so there are these multiple models were released. And if you look at unstructured data, this term may be a bit new for our kind of not very family for developers but basically the storage. Well there is a distinction that is being made you know, between primary storage and unstructured storage or unstructured data solutions And by unstructured we mean file and object storage. If you look at the cube contact nickel sessions, I was very glad to see that there is an entire stream for um machine learning and data so that speaks to how popular communities deployment models are getting when it comes to machine learning and artificial intelligence. Um even applications like genomics and media and entertainment and with the container storage interface uh and the container storage modules with the object storage portfolio that bill has, we offer the comprehensive unstructured data solutions for developers beat object or file. And the advantage the developers are getting is these you know, if you look at platforms like power scale and these areas, these are like the industry workhorses with the highest performance. And if you think of scale, you know, think of 250 nasnotes, you know with a single name space with NVIDIA gpu direct capabilities. All these capabilities developers can use um for you know, applications like machine learning or any competition intensive for data intensive applications that requires these nass uh scale of mass platforms. So so um that's that's what is new in terms of uh what we are offering, you have the storage heaters >>got a parcel. Thank you. David, let's bring you into the conversation now you've launched objects scale at VM World. Talk to us about that, what some of the key features and capabilities are and some of those big business benefits that customers are going to be able to achieve. >>Sure thing. So I really want to focus on three of the biggest benefits. This would be the fact that the product is actually based on kubernetes country, the scale of the product and then its ability to do global replication. So let me just touch on those in order. Mhm You said that the product is based on kubernetes and here we are cube concept. The perfect time to be talking about that. This product really caters to those who are looking for a flexible way to deploy object storage in containerized fashion, appeals to the devops folks and folks who like to automate things and call the communities a P I. S to make uh the actual deployment of the product. Very simple in turnkey and that's really what people turn to kubernetes for is the ability to spin things up when they need them and spend them down as they don't and make that all on commodity hardware and commodity, you know, the quantity pricing and the idea there is that I'm making it as simple and easy as possible. You're not going to get as much shadow I. T. You won't have people going off and putting things off into a public cloud. And so where security of an organization or control of the data that flows with an organization is important. Having something that's easy for developers to use in the same paradigm that they're used to is critical. Now I talked about scale and you know, if you have come to me two years ago I would have told you, you know, kubernetes, yeah, containers people are kicking it around and they're doing some interesting science experiments, I would say in the last year I started to see a lot of requests from customers um in the dozens, even 200 petabyte range as it relates to capacity for committees and specifically looking for C. S. I and cozy with this. This this is the the object storage implementation of the container storage interfaces. Uh So skin was definitely there and the idea of this product is to provide easy scalability from the terabytes range into the multi petabyte range and again it's that ease of use, ease of deployment because it is kubernetes basically because it's a KPI driven that makes that possible. So we're talking about going from a three night minimum to thousands of nodes. and this allows people to deploy the product either at the edge or in the data center um in the edge because you can get very small deployments in the data center to massive scale. So we want to provide something that covers the gamut. The last thing I talked about was replication. So let me just touch upon what I mean by that uh when people go and build these deployments, if you're building a deployment at the edge of an object scale product, you're probably taking in sensor data or some kind of information that you want to then send back to a data center for processing. So you make it simple to do bucket based replication. An object, sorry object storage based replication to move things to another location. And uh that can be used either for bringing data back for analytics from the edge, it can be used for availability. So making sure that you have data available across multiple data centers in the case that you have an outage. It could be even used for sharing data between developers in one site and another site. So we provide that level of flexibility overall. Um this is the next generation object store leveraging. Dell technologies number one position in object storage. So I'm pretty excited about >>and how David is object scale integrated with VM ware software. Stop give us that slice and dice. >>Yeah, and that's a good question. And so, you know, we're talking about this being a Kubernetes based product, you can deploy it on open shift or we integrate directly with VM ware cloud foundation and with Tansy, which is VM ware's container orchestration and management platform. I've seen the demo of the product myself from my team and they've showed it to be did all of the management of the product was actually done within the V sphere Ui, which is great. So easy to go and just enter the V sphere. You I installed the product very simply have it up and running and then go and do all of your management through that user interface or to automate it using the same api is that you used to through VM ware and the 10 Zoo uh platform. >>Thank you, paris are back to you. Security is a big theme here in kubernetes. It's also been a big theme here. We've been talking about it the last three days here at cop con. How does Dell EMC's unstructured portfolio offer that necessary cyber protection that developers need to have and bake that into what they're doing. So >>surely, you know, they talk about cybersecurity, you know, there are different layers of security right from, you know, smarter firewalls to you know how to manage privileged account access and so on. And what we are trying to do is to provide a layer of cyber defense, right at the asset that you're trying to protect, which is the data and this is where the ransom their defender solution is basically detecting any patterns of the compromise that might have happened and alerting the I. T. Um administration about this um possible um intrusions into their into the data by looking at the data access parents in real time. So that's a pretty big deal. Then we're actually putting all this, you know, observance on the primary data and that's what the power scale platform cybersecurity protection features offers. Now we've also extended this kind of detection mechanism for the object data framework on pcs platforms as well. So this is like an additional layer of security at the um layer of uh you know where the data is actually being read and written. Do that's the area, you know, in case of object here we're looking at the S. Three traffic and trying to find his parents in case of a file data atmosphere, looking at the file's access parents and so on. So and in relation to this we're also providing uh data isolation mechanism that is very critical in many cyber recovery processes with the smart absolution as well. So this is something that the developers are getting for like without having to worry about it because that is something implemented at the infrastructure layer itself. So they don't have to worry about you know trying to court it or develop their application to integrate these kinds of things because it's an it's embedded in the infrastructure at the one of the FBI level at the E C. S A P I level. So that's pretty um pretty differentiating in the industry in the country storage solutions. I'll get. >>Uh huh. Yeah. I mean look if you look at what a lot of the object storage players are doing as it relates to cyber security. They're they're playing off the fact that they've implemented object lock and basically using that to lockdown data. And that's that's good. I mean I'm glad that they're doing that and if the case that you were able to lock something down and someone wasn't able to bypass that in some way, that's fantastic. Or if they didn't already encrypted before I got locked down what parts are is referring to is a little bit more than that. It's actually the ability to look at user behavior and determined that something bad is happening. So this is about actually being able to do, you know, predictive analytics being able to go and figure out that you're under attack. There's anomalous behavior um and we're able to go and actually infer from that that something bad is happening and where we think it's happening and lock it down even even more securely than for example just saying hey we provide object like capabilities which is one of the responses that I've seen out there from object storage vendors >>can you share with us. Parts are a customer example like walk us through how this is actually being used and deployed and what some of those business outcomes are. >>Yes lisa. So in terms of container realization itself, they have a media and entertainment kind of customer story here. Um Swiss TXT um they have a platform as a service where they serve their customer base with a range of uh you know, media production and broadcasting solutions and they have containers this platform and part of this computerization is part of their services is they offer infrastructure as a service to you know, media producers who need a high performance storage, high performance computing and power skill And Iceland have been their local solutions to offer this And now that they have containerized their core platform. Well you see a sign interface for power skills, they are able to continue to deliver the infrastructure, high performance infrastructure and storage services to their customers through the A. P I. And it's great to see how fast they could, you know, re factor their application but yet continue to offer the high performance and degrees enterprise grade uh features of the power scale platform. So Swiss Txt and would love to share more. Keep it on the story. Yeah. Hyperlink. >>And where can folks go to learn more about objects scale and what you guys are announcing? Yes, particular. You are a website that you want to direct folks too. >>I would say that technologies dot com. And uh that's the best place to start. >>Yeah, I would go to the Delta product pages around objects should be publicly built. >>Excellent guys, thank you for joining me on the program today. Walking through what how Dell EMC is helping developers with respect to unstructured data, Talking to us about objects skill that you launched VM world, some of those big customer benefits and of course showing us the validation, the proof in the pudding with that customer story. We appreciate your insights. >>Thank you. Thank you lisa >>For my guests. I'm Lisa Martin. You're watching the Cube live from Los Angeles. We're coming to you from our coverage of coupon and cloud native on North America 21. Coming back. Stick around. Rather I should say we'll be back after a short break with our next guest.
SUMMARY :
Coming to you live from cuba con and cloud so far so let's go ahead and start with you. is kind of the next step of productivity for developers beat you know uh are and some of those big business benefits that customers are going to be able to achieve. centers in the case that you have an outage. and how David is object scale integrated with VM ware software. And so, you know, we're talking about this being a Kubernetes necessary cyber protection that developers need to have and bake that into what So they don't have to worry about you know trying So this is about actually being able to do, can you share with us. offer infrastructure as a service to you know, media producers And where can folks go to learn more about objects scale and what you guys are announcing? And uh that's the best place to start. EMC is helping developers with respect to unstructured data, Talking to us about objects skill that you launched Thank you lisa We're coming to you from our coverage of coupon and cloud native on North America 21.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
David | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Dell Technologies | ORGANIZATION | 0.99+ |
Los Angeles | LOCATION | 0.99+ |
David Noi | PERSON | 0.99+ |
FBI | ORGANIZATION | 0.99+ |
lisa martin | PERSON | 0.99+ |
David Noy | PERSON | 0.99+ |
johnny | PERSON | 0.99+ |
North Amer | LOCATION | 0.99+ |
three | QUANTITY | 0.99+ |
three night | QUANTITY | 0.99+ |
250 nasnotes | QUANTITY | 0.99+ |
Dell EMC | ORGANIZATION | 0.99+ |
two years ago | DATE | 0.99+ |
today | DATE | 0.99+ |
NVIDIA | ORGANIZATION | 0.99+ |
Dell | ORGANIZATION | 0.98+ |
last year | DATE | 0.98+ |
thousands | QUANTITY | 0.98+ |
Dell EMC | ORGANIZATION | 0.98+ |
Delta | ORGANIZATION | 0.97+ |
200 petabyte | QUANTITY | 0.97+ |
third day | QUANTITY | 0.97+ |
lisa | PERSON | 0.97+ |
one site | QUANTITY | 0.97+ |
one | QUANTITY | 0.96+ |
Prometheus | TITLE | 0.96+ |
dozens | QUANTITY | 0.95+ |
this week | DATE | 0.95+ |
VM World | ORGANIZATION | 0.93+ |
Swiss TXT | ORGANIZATION | 0.9+ |
single | QUANTITY | 0.9+ |
couple of years ago | DATE | 0.89+ |
Hyperlink | ORGANIZATION | 0.8+ |
V sphere Ui | TITLE | 0.8+ |
Kodati | PERSON | 0.8+ |
north America | LOCATION | 0.78+ |
CloudNativeCon NA 2021 | EVENT | 0.77+ |
Con | EVENT | 0.75+ |
Swiss Txt | ORGANIZATION | 0.75+ |
KubeCon + | EVENT | 0.75+ |
couple | QUANTITY | 0.74+ |
nodes | QUANTITY | 0.72+ |
three object | QUANTITY | 0.72+ |
Iceland | LOCATION | 0.71+ |
Nash | PERSON | 0.7+ |
Tansy | ORGANIZATION | 0.67+ |
petabyte | QUANTITY | 0.66+ |
VM | TITLE | 0.66+ |
cop | ORGANIZATION | 0.64+ |
Cuban | LOCATION | 0.64+ |
E | EVENT | 0.64+ |
paris | ORGANIZATION | 0.63+ |
con | EVENT | 0.63+ |
Parasar | ORGANIZATION | 0.62+ |
Cuban | OTHER | 0.62+ |
C. S. | PERSON | 0.6+ |
last three days | DATE | 0.58+ |
Cube | TITLE | 0.55+ |
guests | QUANTITY | 0.53+ |
S. Three | TITLE | 0.53+ |
2021 | EVENT | 0.51+ |
Kubernetes | TITLE | 0.49+ |
cuba | LOCATION | 0.49+ |
V | TITLE | 0.42+ |
10 Zoo | TITLE | 0.38+ |
Webb Brown | KubeCon + CloudNativeCon NA 2021
>> Welcome back to theCUBE's coverage of KubeCon + CloudNativeCon 21 live form Los Angeles. Lisa Martin, with Dave Nicholson. And we've got a CUBE alum back with us. Webb Brown is back. The co-founder and CEO of Kubecost. Welcome back! >> Thank you so much. It's great to be back. It's been right at two years, a lot's happening in our community and ecosystem as well as with our open source project and company. So awesome with that. >> Give the audience an overview in case they're not familiar with Kubecost. And then talk to us about this explosive growth that you've seen since we last saw you in person. >> Yeah, absolutely. So Kubecost provides cost management solutions purpose-built for teams throwing in Kubernetes and Cloud Native. Right? So everything we do is built on open source. All of our products can be installed in minutes. We give teams visibility into spend, then help them optimize it and govern it over time. So it's been a busy two years since we last talked, we have grown the team about, you know, 5 x, so like right around 20 people today. We now have thousands of mostly medium and large sized enterprises using the product. You know, that's north of a 10 x growth since we launched just before, you know, KubeCon San Diego, now managing billions of dollars of spin and, you know, I feel like, we're just getting started. So it's an incredibly exciting time for us as a company and also just great to be back in person with our friends in the community. >> This community is such a strong community. And it's great to see people back here. I agree. >> Absolutely, absolutely. >> So Kubecost, obviously you talk about cost optimization, but it's, you really, you're an insight engine in the sense that if you're looking at costs, you have to measure that against what you're getting for that cost. >> Absolutely. So what are some of the insights that your platform or that your tool set offers. >> Yeah, absolutely, so, you know, we think about our product is first and foremost, like visibility and monitoring and then insights and optimization and then governance. You know, if you talk to most teams today, they're still kind of getting that visibility, but once you do it quickly leads in how do we optimize? And then we're going to give you insights at every part of the stack, right? So like at the infrastructure layer, thinking about things like Spot and RIS and savings plans, et cetera. At the Kubernetes orchestration layer, thinking about things like auto scaling and, you know, setting requests and limits, et cetera, all the way up to like the application layer with all of that being purpose-built for, you know, Cloud Native Kubernetes. So the way we work as you deploy our product in your environment, anywhere you're running Kubernetes, 1.11 or above we'll run. And we're going to start dynamically generating these insights in minutes and they're real time. And again, they scaled to the largest Kubernetes clusters in the world. >> And you said, you've had a thousand or so customers in the medium to large enterprise. These are large organizations, probably brand names, I imagine we are familiar with that are leaning on Kubecost to help get that visibility that before they did not have the ability to get. >> Absolutely, absolutely. So definitely our users of our thousands of users, skews heavily towards, you know, medium and large side enterprise. Working with some amazing companies like Adobe, who, you know, just have such high scale and like complex and sophisticated infrastructure. So, you know, I think this is very natural in what we expect, which is like, as you start spending more resources, you know, missing visibility, having unoptimized infrastructure starts to be more costly. >> Absolutely. >> And we typically see as once that gets into like the multiple head count, right? And it starts to, you know, spend some, may make sense to spend some time optimizing and monitoring and, you know, putting the learning in place. So you can manage it more effectively as time goes on. >> Do you have any metrics or any X factor ranges of the costs that you've actually saved customers? >> Yeah. I mean, we've saved multiple customers in them, like many of millions of dollars at this point, >> So we're talking big. >> Really big. So yeah, we're now managing more than $2 billion of spin. So like some really big savings on a per customer base, but it's really common where we're saving, you know, north of 30%, sometimes up to 70% on your Kubernetes and related spin. And so we're giving you insights into your Kubernetes cluster and again, the full stack there, but also giving you visibility and insights into external things like external disk or cloud storage buckets or, you know, cloud sequel that, that sort of stuff, external cloud services. >> Taking those blinders off >> Exactly. And giving you that unified, you know, real time picture again, that accurately reflects everything that's going on in your system. >> So when these insights are produced or revealed, are the responses automated? or are they then manually applied? >> Yeah. Yeah. That's a great question. We support both and we support both in different ways By default, when you deploy Kubecost, and again it's, today it's Helm Install. It can be running in your cluster in, you know, minutes or less, it's deployed in read only mode. And by the way, you don't share any data externally, it's all in your local environment. So we started generating these insights, you know, right when you install in your environment. >> Let me ask you about, I'm sorry to interrupt, but when you say you're generating an insight, are you just giving an answer and guidance? or you're providing the reader background on what leads to that insight? >> Yeah. You know, is that a philosophical question of, do you need to provide the user rationale for the insight? >> Yeah, absolutely. And I think we're doing this today and we'll do more, but one example is, you know, if you just look at this notion of setting requests and limits for your applications in Kubernetes, you know, if you, in simple forms, if you set a request too high, you're potentially wasting money because the Kubernetes scheduler is presenting that resource for you. If you set it too low, you're at risk of being CPU throttled, right? So communicating that symbiotic relationship and the risk on either side really helps the team understand why do I need to strike this balance, right? It's not just cost it's performance and reliability as well. So absolutely given that background and again, out of the box we're read only, but we also have automation in our product with our cluster controller. So you can dynamically do things like right-size your infrastructure, or, you know, move workloads to Spot, et cetera. But we also have integrations with a bunch of tooling in this ecosystem. So like Prometheus native, you know, Alert Manager native, just launched an integration with Spinnaker and Armory where you can like dynamically at the time of deployment, you know, right size and have insights. So you can expect to see more from us there. But we very much think about automation is twofold. One, you know, building trust in Kubecost and our insights and adopting them over time. But then two is meeting you where you are with your existing tooling, whether it's your CICB pipeline, observability or, you know, existing kind of workflow automation system. >> Meeting customers where they are is, is critical these days. >> Absolutely. I think, especially in this market, right? where we have the potential to have so much interoperability and all these things working in harmony and also, you know, there's, there's a lot of booths back here, right? So we, you know, we have complex tech stacks and, you know, in certain cases we feel like when we bring you to our UI or API's or, you know, automation or COI's, we can do things more effective. But oftentimes when we bring that data to you, we can be more effective again, that's, you know, coming, bringing your data to Chronosphere or Prometheus or Grafana, you know, all of the tooling that you're already using on a daily, regular basis. >> Bringing that data into the tool is just another example of the value in data that the organizations can actually harness that value and unlock it. >> Webb: Yeah. >> There's so much potential there for them to be more competitive, for them to be able to develop products and services faster. >> Absolutely. Yeah, I think you're just seeing the coming of age with, you know, cost metrics into that equation. We now live in a world with Kubernetes as this amazing innovation platform where as an engineer, I can go spin up some pretty costly resources, really fast, and that's a great thing for innovation, right? But it also kind of pushes some of the accountability or awareness down to the individual >> Webb: IC who needs to be aware, you know, what, you know, things generally cost at a minimum in like a directional way, so they can make informed decisions again, when they think about this cost performance, reliability, trade-off. >> Lisa: Where are your customer conversations? Are your target users, DevOps folks? I was just wondering where finance might be in this whole game. >> Yeah, it's a great question. Given the fact that we are kind of open source first and started with open source, we, you know, 95% of the time when we start working with an infrastructure engineering team or dev ops team. They've already installed our product. They're already familiar with what we're doing, but then increasingly and increasingly fast, you know, finance is being brought into the equation and, you know, management is being brought into the equation. And I think it's a function of what we were talking about where, you know, 70% of teams grew their Kubernetes spend over the last year, you know, 20% of them more than doubled. So, you know, these are starting to be real, you know, expense items where finance is increasingly aware of what's going on. So yeah, they're coming into the picture, but it's simply thought that you starting with, and, and working with the infrastructure team, that's actually kind of putting some of these insights into action or hooking us into their pipelines or something. >> When you think of developers going out and grabbing resources, and you think of a, an insight tool that looks at controlling cost, that could seem like an inhibitor. But really if you're talking about how to efficiently use whatever resources you have to be able to have access to in terms of dollars, you could sell this to the developers on that basis. It's like, look, you have these 10 things that you want to be able to do. If you don't optimize using a tool like this, you're only going to be able to do 4 of them. >> Without a doubt. Yeah. And you know, us as our founding team, all engineers, you know, we were the ones getting those questions of, you know, how have we already spent, you know, our budget on just this project? We have these three others we want to do, right? Or why are costs going up as quickly as they are? You know, what are we spending on this application, instead of that kind of being a manual lift, like, let me go do a bunch of analysis or come back with answers. It's tools to where not only can management answer those questions themselves, but like engineering teams can make informed opportunity costs and optimizations decisions itself, whether it's tooling and automation doing it for them or them applying things, you know, directly. >> Lisa: So a lot of growth. You talked about the growth on employees, the growth in revenue, what lies ahead for Kubecost? What are some of the things that are coming on the horizon that you're really excited about? >> Yeah, we very much feel like we're just getting started you know, just like we feel this ecosystem and community is, right? Like there's been tons of progress all around, but like, wow, it's still early days. So, you know, we, we did raise, you know, five and a half million dollars from, you know, First Round who is an amazing group to work with at the end of last year. So by growing the engineering team were able to do a lot more. We got a bunch of really big things coming across all parts of our product. You can think about one thing we're really excited that's in limit availability right now is our first hosted solution. It's our first SaaS solution. And this is critically important to us in that we want to give teams the option to, if you want to own and control your data and never egress anything outside of your cluster, you can do that with our deploy product. You can do that with our open source. You can truly lock down namespace to egress and never send a byte out. Or if you'd like the convenience of us to manage it for you and be kind of stewards of your data, we're going to offer, you know, a great offering there too. So that's unlimited availability day. We're going to have a lot more announcements coming there, but we see those being at feature parity, you know, between like our enterprise offerings and our hosted solution and just, you know, a lot more coming with, you know, visibility, some more like GPU insights, you know, metrics coming quickly, a lot more with automation coming and then more integrations for governance. Again, kind of talked about Spinnaker and things like that. A lot more really interesting ones coming. >> So five and a half million raised in the last round of funding. Where are you going to be applying that? What are some of the growth engines that you want to tune with that money? >> Yeah, so, you know, first and foremost, it was really growing the engineering team, right? So we've, you know, like 4 x the engineering team in the last year, and just have an amazing group of engineers. We want to continue to do that. >> Webb: We're kind of super early on the like, you know, marketing and sales side. We're going to start thinking about that more and more, you know, our approach first off was like, we want to solve a really valuable problem and doing it in a way that is super compelling. And we think that when you do that, you know, good things happen. I think that's some of our Google background, which is like, you build a great search engine and like, you know, good things generally happen. So we're just super focused on, again, working with great users, you know, building great products that meet them where they are and solve problems that are really important to them. >> Lisa: Awesome. Well, congratulations on all the trajectory of success since we last saw you in person. >> Thank you. >> Great to have you back on the show, looking forward to, so folks can go to www.kubecost.com to learn more and see some of those announcements coming down the pike. >> Absolutely, yeah. >> Don't you make it two years before you come back. >> Webb: I would love to be back. I hope we're back bigger than ever, you know, next year, but it has been such a pleasure, you know, last time and this time, thank you so much for having me, you know, I love being part of the show and the community at large. >> It's a great community and we appreciate you sharing all your insights. >> Thank you so much. >> All right. For Dave Nicholson, I'm Lisa Martin coming to you live from Los Angeles. This is theCUBE's coverage of KubeCon and CloudNativeCon 21. We back with our next guest shortly. We'll see you there.
SUMMARY :
and CEO of Kubecost. Thank you so much. last saw you in person. of spin and, you know, I feel like, And it's great to see So Kubecost, obviously you or that your tool set offers. So the way we work as you And you said, you've had like Adobe, who, you know, And it starts to, you know, spend some, like many of millions of you know, north of 30%, that unified, you know, And by the way, you don't do you need to provide the at the time of deployment, you know, is critical these days. So we, you know, we have complex Bringing that data into the tool for them to be more competitive, the coming of age with, you know, aware, you know, what, you know, Lisa: Where are your over the last year, you know, and you think of a, you know, we were the ones Lisa: So a lot of growth. and just, you know, that you want to tune with that money? So we've, you know, like and like, you know, good we last saw you in person. Great to have you back on the show, years before you come back. you know, next year, but it and we appreciate you We'll see you there.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Lisa Martin | PERSON | 0.99+ |
Dave Nicholson | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
Los Angeles | LOCATION | 0.99+ |
95% | QUANTITY | 0.99+ |
20% | QUANTITY | 0.99+ |
Adobe | ORGANIZATION | 0.99+ |
4 | QUANTITY | 0.99+ |
Webb | PERSON | 0.99+ |
10 things | QUANTITY | 0.99+ |
thousands | QUANTITY | 0.99+ |
70% | QUANTITY | 0.99+ |
Webb Brown | PERSON | 0.99+ |
two years | QUANTITY | 0.99+ |
www.kubecost.com | OTHER | 0.99+ |
more than $2 billion | QUANTITY | 0.99+ |
KubeCon | EVENT | 0.99+ |
Prometheus | TITLE | 0.99+ |
two | QUANTITY | 0.99+ |
next year | DATE | 0.99+ |
Kubecost | ORGANIZATION | 0.99+ |
CloudNativeCon | EVENT | 0.99+ |
two years | QUANTITY | 0.99+ |
five and a half million dollars | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
last year | DATE | 0.98+ |
three | QUANTITY | 0.98+ |
One | QUANTITY | 0.98+ |
Kubernetes | TITLE | 0.98+ |
today | DATE | 0.98+ |
10 x | QUANTITY | 0.98+ |
one example | QUANTITY | 0.97+ |
one | QUANTITY | 0.97+ |
billions of dollars | QUANTITY | 0.97+ |
Armory | ORGANIZATION | 0.97+ |
first | QUANTITY | 0.97+ |
Chronosphere | TITLE | 0.97+ |
around 20 people | QUANTITY | 0.96+ |
first hosted solution | QUANTITY | 0.96+ |
five and a half milli | QUANTITY | 0.96+ |
KubeCon | ORGANIZATION | 0.96+ |
Cloud Native | TITLE | 0.95+ |
Alert Manager | TITLE | 0.94+ |
up to 70% | QUANTITY | 0.94+ |
theCUBE | ORGANIZATION | 0.93+ |
millions of dollars | QUANTITY | 0.89+ |
First Round | QUANTITY | 0.89+ |
CUBE | ORGANIZATION | 0.88+ |
Spinnaker | ORGANIZATION | 0.87+ |
21 | QUANTITY | 0.85+ |
a thousand | QUANTITY | 0.85+ |
San Diego | LOCATION | 0.84+ |
Grafana | TITLE | 0.83+ |
CICB | ORGANIZATION | 0.82+ |
end of last year | DATE | 0.81+ |
CloudNativeCon 21 | EVENT | 0.8+ |
30% | QUANTITY | 0.76+ |
5 x | QUANTITY | 0.73+ |
NA 2021 | EVENT | 0.73+ |
first SaaS solution | QUANTITY | 0.69+ |
twofold | QUANTITY | 0.63+ |
Pat Conte, Opsani | AWS Startup Showcase
(upbeat music) >> Hello and welcome to this CUBE conversation here presenting the "AWS Startup Showcase: "New Breakthroughs in DevOps, Data Analytics "and Cloud Management Tools" featuring Opsani for the cloud management and migration track here today, I'm your host John Furrier. Today, we're joined by Patrick Conte, Chief Commercial Officer, Opsani. Thanks for coming on. Appreciate you coming on. Future of AI operations. >> Thanks, John. Great to be here. Appreciate being with you. >> So congratulations on all your success being showcased here as part of the Startups Showcase, future of AI operations. You've got the cloud scale happening. A lot of new transitions in this quote digital transformation as cloud scales goes next generation. DevOps revolution as Emily Freeman pointed out in her keynote. What's the problem statement that you guys are focused on? Obviously, AI involves a lot of automation. I can imagine there's a data problem in there somewhere. What's the core problem that you guys are focused on? >> Yeah, it's interesting because there are a lot of companies that focus on trying to help other companies optimize what they're doing in the cloud, whether it's cost or whether it's performance or something else. We felt very strongly that AI was the way to do that. I've got a slide prepared, and maybe we can take a quick look at that, and that'll talk about the three elements or dimensions of the problem. So we think about cloud services and the challenge of delivering cloud services. You've really got three things that customers are trying to solve for. They're trying to solve for performance, they're trying to solve for the best performance, and, ultimately, scalability. I mean, applications are growing really quickly especially in this current timeframe with cloud services and whatnot. They're trying to keep costs under control because certainly, it can get way out of control in the cloud since you don't own the infrastructure, and more importantly than anything else which is why it's at the bottom sort of at the foundation of all this, is they want their applications to be a really a good experience for their customers. So our customer's customer is actually who we're trying to solve this problem for. So what we've done is we've built a platform that uses AI and machine learning to optimize, meaning tune, all of the key parameters of a cloud application. So those are things like the CPU usage, the memory usage, the number of replicas in a Kubernetes or container environment, those kinds of things. It seems like it would be simple just to grab some values and plug 'em in, but it's not. It's actually the combination of them has to be right. Otherwise, you get delays or faults or other problems with the application. >> Andrew, if you can bring that slide back up for a second. I want to just ask one quick question on the problem statement. You got expenditures, performance, customer experience kind of on the sides there. Do you see this tip a certain way depending upon use cases? I mean, is there one thing that jumps out at you, Patrick, from your customer's customer's standpoint? Obviously, customer experience is the outcome. That's the app, whatever. That's whatever we got going on there. >> Sure. >> But is there patterns 'cause you can have good performance, but then budget overruns. Or all of them could be failing. Talk about this dynamic with this triangle. >> Well, without AI, without machine learning, you can solve for one of these, only one, right? So if you want to solve for performance like you said, your costs may overrun, and you're probably not going to have control of the customer experience. If you want to solve for one of the others, you're going to have to sacrifice the other two. With machine learning though, we can actually balance that, and it isn't a perfect balance, and the question you asked is really a great one. Sometimes, you want to over-correct on something. Sometimes, scalability is more important than cost, but what we're going to do because of our machine learning capability, we're going to always make sure that you're never spending more than you should spend, so we're always going to make sure that you have the best cost for whatever the performance and reliability factors that you you want to have are. >> Yeah, I can imagine. Some people leave services on. Happened to us one time. An intern left one of the services on, and like where did that bill come from? So kind of looked back, we had to kind of fix that. There's a ton of action, but I got to ask you, what are customers looking for with you guys? I mean, as they look at Opsani, what you guys are offering, what's different than what other people might be proposing with optimization solutions? >> Sure. Well, why don't we bring up the second slide, and this'll illustrate some of the differences, and we can talk through some of this stuff as well. So really, the area that we play in is called AIOps, and that's sort of a new area, if you will, over the last few years, and really what it means is applying intelligence to your cloud operations, and those cloud operations could be development operations, or they could be production operations. And what this slide is really representing is in the upper slide, that's sort of the way customers experience their DevOps model today. Somebody says we need an application or we need a feature, the developers pull down something from get. They hack an early version of it. They run through some tests. They size it whatever way they know that it won't fail, and then they throw it over to the SREs to try to tune it before they shove it out into production, but nobody really sizes it properly. It's not optimized, and so it's not tuned either. When it goes into production, it's just the first combination of settings that work. So what happens is undoubtedly, there's some type of a problem, a fault or a delay, or you push new code, or there's a change in traffic. Something happens, and then, you've got to figure out what the heck. So what happens then is you use your tools. First thing you do is you over-provision everything. That's what everybody does, they over-provision and try to soak up the problem. But that doesn't solve it because now, your costs are going crazy. You've got to go back and find out and try as best you can to get root cause. You go back to the tests, and you're trying to find something in the test phase that might be an indicator. Eventually your developers have to hack a hot fix, and the conveyor belt sort of keeps on going. We've tested this model on every single customer that we've spoken to, and they've all said this is what they experience on a day-to-day basis. Now, if we can go back to the side, let's talk about the second part which is what we do and what makes us different. So on the bottom of this slide, you'll see it's really a shift-left model. What we do is we plug in in the production phase, and as I mentioned earlier, what we're doing is we're tuning all those cloud parameters. We're tuning the CPU, the memory, the Replicas, all those kinds of things. We're tuning them all in concert, and we're doing it at machine speed, so that's how the customer gets the best performance, the best reliability at the best cost. That's the way we're able to achieve that is because we're iterating this thing in machine speed, but there's one other place where we plug in and we help the whole concept of AIOps and DevOps, and that is we can plug in in the test phase as well. And so if you think about it, the DevOps guy can actually not have to over-provision before he throws it over to the SREs. He can actually optimize and find the right size of the application before he sends it through to the SREs, and what this does is collapses the timeframe because it means the SREs don't have to hunt for a working set of parameters. They get one from the DevOps guys when they send it over, and this is how the future of AIOps is being really affected by optimization and what we call autonomous optimization which means that it's happening without humans having to press a button on it. >> John: Andrew, bring that slide back up. I want to just ask another question. Tuning in concert thing is very interesting to me. So how does that work? Are you telegraphing information to the developer from the autonomous workload tuning engine piece? I mean, how does the developer know the right knobs or where does it get that provisioning information? I see the performance lag. I see where you're solving that problem. >> Sure. >> How does that work? >> Yeah, so actually, if we go to the next slide, I'll show you exactly how it works. Okay, so this slide represents the architecture of a typical application environment that we would find ourselves in, and inside the dotted line is the customer's application namespace. That's where the app is. And so, it's got a bunch of pods. It's got a horizontal pod. It's got something for replication, probably an HPA. And so, what we do is we install inside that namespace two small instances. One is a tuning pod which some people call a canary, and that tuning pod joins the rest of the pods, but it's not part of the application. It's actually separate, but it gets the same traffic. We also install somebody we call Servo which is basically an action engine. What Servo does is Servo takes the metrics from whatever the metric system is is collecting all those different settings and whatnot from the working application. It could be something like Prometheus. It could be an Envoy Sidecar, or more likely, it's something like AppDynamics, or we can even collect metrics off of Nginx which is at the front of the service. We can plug into anywhere where those metrics are. We can pull the metrics forward. Once we see the metrics, we send them to our backend. The Opsani SaaS service is our machine learning backend. That's where all the magic happens, and what happens then is that service sees the settings, sends a recommendation to Servo, Servo sends it to the tuning pod, and we tune until we find optimal. And so, that iteration typically takes about 20 steps. It depends on how big the application is and whatnot, how fast those steps take. It could be anywhere from seconds to minutes to 10 to 20 minutes per step, but typically within about 20 steps, we can find optimal, and then we'll come back and we'll say, "Here's optimal, and do you want to "promote this to production," and the customer says, "Yes, I want to promote it to production "because I'm saving a lot of money or because I've gotten "better performance or better reliability." Then, all he has to do is press a button, and all that stuff gets sent right to the production pods, and all of those settings get put into production, and now he's now he's actually saving the money. So that's basically how it works. >> It's kind of like when I want to go to the beach, I look at the weather.com, I check the forecast, and I decide whether I want to go or not. You're getting the data, so you're getting a good look at the information, and then putting that into a policy standpoint. I get that, makes total sense. Can I ask you, if you don't mind, expanding on the performance and reliability and the cost advantage? You mentioned cost. How is that impacting? Give us an example of some performance impact, reliability, and cost impacts. >> Well, let's talk about what those things mean because like a lot of people might have different ideas about what they think those mean. So from a cost standpoint, we're talking about cloud spend ultimately, but it's represented by the settings themselves, so I'm not talking about what deal you cut with AWS or Azure or Google. I'm talking about whatever deal you cut, we're going to save you 30, 50, 70% off of that. So it doesn't really matter what cost you negotiated. What we're talking about is right-sizing the settings for CPU and memory, Replica. Could be Java. It could be garbage collection, time ratios, or heap sizes or things like that. Those are all the kinds of things that we can tune. The thing is most of those settings have an unlimited number of values, and this is why machine learning is important because, if you think about it, even if they only had eight settings or eight values per setting, now you're talking about literally billions of combinations. So to find optimal, you've got to have machine speed to be able to do it, and you have to iterate very, very quickly to make it happen. So that's basically the thing, and that's really one of the things that makes us different from anybody else, and if you put that last slide back up, the architecture slide, for just a second, there's a couple of key words at the bottom of it that I want to want to focus on, continuous. So continuous really means that we're on all the time. We're not plug us in one time, make a change, and then walk away. We're actually always measuring and adjusting, and the reason why this is important is in the modern DevOps world, your traffic level is going to change. You're going to push new code. Things are going to happen that are going to change the basic nature of the software, and you have to be able to tune for those changes. So continuous is very important. Second thing is autonomous. This is designed to take pressure off of the SREs. It's not designed to replace them, but to take the pressure off of them having to check pager all the time and run in and make adjustments, or try to divine or find an adjustment that might be very, very difficult for them to do so. So we're doing it for them, and that scale means that we can solve this for, let's say, one big monolithic application, or we can solve it for literally hundreds of applications and thousands of microservices that make up those applications and tune them all at the same time. So the same platform can be used for all of those. You originally asked about the parameters and the settings. Did I answer the question there? >> You totally did. I mean, the tuning in concert. You mentioned early as a key point. I mean, you're basically tuning the engine. It's not so much negotiating a purchase SaaS discount. It's essentially cost overruns by the engine, either over burning or heating or whatever you want to call it. I mean, basically inefficiency. You're tuning the core engine. >> Exactly so. So the cost thing is I mentioned is due to right-sizing the settings and the number of Replicas. The performance is typically measured via latency, and the reliability is typically measured via error rates. And there's some other measures as well. We have a whole list of them that are in the application itself, but those are the kinds of things that we look for as results. When we do our tuning, we look for reducing error rates, or we look for holding error rates at zero, for example, even if we improve the performance or we improve the cost. So we're looking for the best result, the best combination result, and then a customer can decide if they want to do so to actually over-correct on something. We have the whole concept of guard rail, so if performance is the most important thing, or maybe some customers, cost is the most important thing, they can actually say, "Well, give us the best cost, "and give us the best performance and the best reliability, "but at this cost," and we can then use that as a service-level objective and tune around it. >> Yeah, it reminds me back in the old days when you had filtering white lists of black lists of addresses that can go through, say, a firewall or a device. You have billions of combinations now with machine learning. It's essentially scaling the same concept to unbelievable. These guardrails are now in place, and that's super cool and I think really relevant call-out point, Patrick, to kind of highlight that. At this kind of scale, you need machine learning, you need the AI to essentially identify quickly the patterns or combinations that are actually happening so a human doesn't have to waste their time that can be filled by basically a bot at that point. >> So John, there's just one other thing I want to mention around this, and that is one of the things that makes us different from other companies that do optimization. Basically, every other company in the optimization space creates a static recommendation, basically their recommendation engines, and what you get out of that is, let's say it's a manifest of changes, and you hand that to the SREs, and they put it into effect. Well, the fact of the matter is is that the traffic could have changed then. It could have spiked up, or it could have dropped below normal. You could have introduced a new feature or some other code change, and at that point in time, you've already instituted these changes. They may be completely out of date. That's why the continuous nature of what we do is important and different. >> It's funny, even the language that we're using here: network, garbage collection. I mean, you're talking about tuning an engine, am operating system. You're talking about stuff that's moving up the stack to the application layer, hence this new kind of eliminating of these kind of siloed waterfall, as you pointed out in your second slide, is kind of one integrated kind of operating environment. So when you have that or think about the data coming in, and you have to think about the automation just like self-correcting, error-correcting, tuning, garbage collection. These are words that we've kind of kicking around, but at the end of the day, it's an operating system. >> Well in the old days of automobiles, which I remember cause I'm I'm an old guy, if you wanted to tune your engine, you would probably rebuild your carburetor and turn some dials to get the air-oxygen-gas mix right. You'd re-gap your spark plugs. You'd probably make sure your points were right. There'd be four or five key things that you would do. You couldn't do them at the same time unless you had a magic wand. So we're the magic wand that basically, or in modern world, we're sort of that thing you plug in that tunes everything at once within that engine which is all now electronically controlled. So that's the big differences as you think about what we used to do manually, and now, can be done with automation. It can be done much, much faster without humans having to get their fingernails greasy, let's say. >> And I think the dynamic versus static is an interesting point. I want to bring up the SRE which has become a role that's becoming very prominent in the DevOps kind of plus world that's happening. You're seeing this new revolution. The role of the SRE is not just to be there to hold down and do the manual configuration. They had a scale. They're a developer, too. So I think this notion of offloading the SRE from doing manual tasks is another big, important point. Can you just react to that and share more about why the SRE role is so important and why automating that away through when you guys have is important? >> The SRE role is becoming more and more important, just as you said, and the reason is because somebody has to get that application ready for production. The DevOps guys don't do it. That's not their job. Their job is to get the code finished and send it through, and the SREs then have to make sure that that code will work, so they have to find a set of settings that will actually work in production. Once they find that set of settings, the first one they find that works, they'll push it through. It's not optimized at that point in time because they don't have time to try to find optimal, and if you think about it, the difference between a machine learning backend and an army of SREs that work 24-by-seven, we're talking about being able to do the work of many, many SREs that never get tired, that never need to go play video games, to unstress or whatever. We're working all the time. We're always measuring, adjusting. A lot of the companies we talked to do a once-a-month adjustment on their software. So they put an application out, and then they send in their SREs once a month to try to tune the application, and maybe they're using some of these other tools, or maybe they're using just their smarts, but they'll do that once a month. Well, gosh, they've pushed code probably four times during the month, and they probably had a bunch of different spikes and drops in traffic and other things that have happened. So we just want to help them spend their time on making sure that the application is ready for production. Want to make sure that all the other parts of the application are where they should be, and let us worry about tuning CPU, memory, Replica, job instances, and things like that so that they can work on making sure that application gets out and that it can scale, which is really important for them, for their companies to make money is for the apps to scale. >> Well, that's a great insight, Patrick. You mentioned you have a lot of great customers, and certainly if you have your customer base are early adopters, pioneers, and grow big companies because they have DevOps. They know that they're seeing a DevOps engineer and an SRE. Some of the other enterprises that are transforming think the DevOps engineer is the SRE person 'cause they're having to get transformed. So you guys are at the high end and getting now the new enterprises as they come on board to cloud scale. You have a huge uptake in Kubernetes, starting to see the standardization of microservices. People are getting it, so I got to ask you can you give us some examples of your customers, how they're organized, some case studies, who uses you guys, and why they love you? >> Sure. Well, let's bring up the next slide. We've got some customer examples here, and your viewers, our viewers, can probably figure out who these guys are. I can't tell them, but if they go on our website, they can sort of put two and two together, but the first one there is a major financial application SaaS provider, and in this particular case, they were having problems that they couldn't diagnose within the stack. Ultimately, they had to apply automation to it, and what we were able to do for them was give them a huge jump in reliability which was actually the biggest problem that they were having. We gave them 5,000 hours back a month in terms of the application. They were they're having pager duty alerts going off all the time. We actually gave them better performance. We gave them a 10% performance boost, and we dropped their cloud spend for that application by 72%. So in fact, it was an 80-plus % price performance or cost performance improvement that we gave them, and essentially, we helped them tune the entire stack. This was a hybrid environment, so this included VMs as well as more modern architecture. Today, I would say the overwhelming majority of our customers have moved off of the VMs and are in a containerized environment, and even more to the point, Kubernetes which we find just a very, very high percentage of our customers have moved to. So most of the work we're doing today with new customers is around that, and if we look at the second and third examples here, those are examples of that. In the second example, that's a company that develops websites. It's one of the big ones out in the marketplace that, let's say, if you were starting a new business and you wanted a website, they would develop that website for you. So their internal infrastructure is all brand new stuff. It's all Kubernetes, and what we were able to do for them is they were actually getting decent performance. We held their performance at their SLO. We achieved a 100% error-free scenario for them at runtime, and we dropped their cost by 80%. So for them, they needed us to hold-serve, if you will, on performance and reliability and get their costs under control because everything in that, that's a cloud native company. Everything there is cloud cost. So the interesting thing is it took us nine steps because nine of our iterations to actually get to optimal. So it was very, very quick, and there was no integration required. In the first case, we actually had to do a custom integration for an underlying platform that was used for CICD, but with the- >> John: Because of the hybrid, right? >> Patrick: Sorry? >> John: Because it was hybrid, right? >> Patrick: Yes, because it was hybrid, exactly. But within the second one, we just plugged right in, and we were able to tune the Kubernetes environment just as I showed in that architecture slide, and then the third one is one of the leading application performance monitoring companies on the market. They have a bunch of their own internal applications and those use a lot of cloud spend. They're actually running Kubernetes on top of VMs, but we don't have to worry about the VM layer. We just worry about the Kubernetes layer for them, and what we did for them was we gave them a 48% performance improvement in terms of latency and throughput. We dropped their error rates by 90% which is pretty substantial to say the least, and we gave them a 50% cost delta from where they had been. So this is the perfect example of actually being able to deliver on all three things which you can't always do. It has to be, sort of all applications are not created equal. This was one where we were able to actually deliver on all three of the key objectives. We were able to set them up in about 25 minutes from the time we got started, no extra integration, and needless to say, it was a big, happy moment for the developers to be able to go back to their bosses and say, "Hey, we have better performance, "better reliability. "Oh, by the way, we saved you half." >> So depending on the stack situation, you got VMs and Kubernetes on the one side, cloud-native, all Kubernetes, that's dream scenario obviously. Not many people like that. All the new stuff's going cloud-native, so that's ideal, and then the mixed ones, Kubernetes, but no VMs, right? >> Yeah, exactly. So Kubernetes with no VMs, no problem. Kubernetes on top of VMs, no problem, but we don't manage the VMs. We don't manage the underlay at all, in fact. And the other thing is we don't have to go back to the slide, but I think everybody will remember the slide that had the architecture, and on one side was our cloud instance. The only data that's going between the application and our cloud instance are the settings, so there's never any data. There's never any customer data, nothing for PCI, nothing for HIPPA, nothing for GDPR or any of those things. So no personal data, no health data. Nothing is passing back and forth. Just the settings of the containers. >> Patrick, while I got you here 'cause you're such a great, insightful guest, thank you for coming on and showcasing your company. Kubernetes real quick. How prevalent is this mainstream trend is because you're seeing such great examples of performance improvements. SLAs being met, SLOs being met. How real is Kubernetes for the mainstream enterprise as they're starting to use containers to tip their legacy and get into the cloud-native and certainly hybrid and soon to be multi-cloud environment? >> Yeah, I would not say it's dominant yet. Of container environments, I would say it's dominant now, but for all environments, it's not. I think the larger legacy companies are still going through that digital transformation, and so what we do is we catch them at that transformation point, and we can help them develop because as we remember from the AIOps slide, we can plug in at that test level and help them sort of pre-optimize as they're coming through. So we can actually help them be more efficient as they're transforming. The other side of it is the cloud-native companies. So you've got the legacy companies, brick and mortar, who are desperately trying to move to digitization. Then, you've got the ones that are born in the cloud. Most of them aren't on VMs at all. Most of them are on containers right from the get-go, but you do have some in the middle who have started to make a transition, and what they've done is they've taken their native VM environment and they've put Kubernetes on top of it so that way, they don't have to scuttle everything underneath it. >> Great. >> So I would say it's mixed at this point. >> Great business model, helping customers today, and being a bridge to the future. Real quick, what licensing models, how to buy, promotions you have for Amazon Web Services customers? How do people get involved? How do you guys charge? >> The product is licensed as a service, and the typical service is an annual. We license it by application, so let's just say you have an application, and it has 10 microservices. That would be a standard application. We'd have an annual cost for optimizing that application over the course of the year. We have a large application pack, if you will, for let's say applications of 20 services, something like that, and then we also have a platform, what we call Opsani platform, and that is for environments where the customer might have hundreds of applications and-or thousands of services, and we can plug into their deployment platform, something like a harness or Spinnaker or Jenkins or something like that, or we can plug into their their cloud Kubernetes orchestrator, and then we can actually discover the apps and optimize them. So we've got environments for both single apps and for many, many apps, and with the same platform. And yes, thanks for reminding me. We do have a promotion for for our AWS viewers. If you reference this presentation, and you look at the URL there which is opsani.com/awsstartupshowcase, can't forget that, you will, number one, get a free trial of our software. If you optimize one of your own applications, we're going to give you an Oculus set of goggles, the augmented reality goggles. And we have one other promotion for your viewers and for our joint customers here, and that is if you buy an annual license, you're going to get actually 15 months. So that's what we're putting on the table. It's actually a pretty good deal. The Oculus isn't contingent. That's a promotion. It's contingent on you actually optimizing one of your own services. So it's not a synthetic app. It's got to be one of your own apps, but that's what we've got on the table here, and I think it's a pretty good deal, and I hope your guys take us up on it. >> All right, great. Get Oculus Rift for optimizing one of your apps and 15 months for the price of 12. Patrick, thank you for coming on and sharing the future of AIOps with you guys. Great product, bridge to the future, solving a lot of problems. A lot of use cases there. Congratulations on your success. Thanks for coming on. >> Thank you so much. This has been excellent, and I really appreciate it. >> Hey, thanks for sharing. I'm John Furrier, your host with theCUBE. Thanks for watching. (upbeat music)
SUMMARY :
for the cloud management and Appreciate being with you. of the Startups Showcase, and that'll talk about the three elements kind of on the sides there. 'cause you can have good performance, and the question you asked An intern left one of the services on, and find the right size I mean, how does the and the customer says, and the cost advantage? and that's really one of the things I mean, the tuning in concert. So the cost thing is I mentioned is due to in the old days when you had and that is one of the things and you have to think about the automation So that's the big differences of offloading the SRE and the SREs then have to make sure and certainly if you So most of the work we're doing today "Oh, by the way, we saved you half." So depending on the stack situation, and our cloud instance are the settings, and get into the cloud-native that are born in the cloud. So I would say it's and being a bridge to the future. and the typical service is an annual. and 15 months for the price of 12. and I really appreciate it. I'm John Furrier, your host with theCUBE.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Emily Freeman | PERSON | 0.99+ |
Patrick | PERSON | 0.99+ |
John | PERSON | 0.99+ |
Andrew | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
Pat Conte | PERSON | 0.99+ |
10% | QUANTITY | 0.99+ |
50% | QUANTITY | 0.99+ |
Patrick Conte | PERSON | 0.99+ |
15 months | QUANTITY | 0.99+ |
second | QUANTITY | 0.99+ |
90% | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
thousands | QUANTITY | 0.99+ |
four | QUANTITY | 0.99+ |
nine steps | QUANTITY | 0.99+ |
30 | QUANTITY | 0.99+ |
Oculus | ORGANIZATION | 0.99+ |
100% | QUANTITY | 0.99+ |
72% | QUANTITY | 0.99+ |
48% | QUANTITY | 0.99+ |
10 microservices | QUANTITY | 0.99+ |
second part | QUANTITY | 0.99+ |
First | QUANTITY | 0.99+ |
second slide | QUANTITY | 0.99+ |
first case | QUANTITY | 0.99+ |
Today | DATE | 0.99+ |
Amazon Web Services | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.99+ | |
one | QUANTITY | 0.99+ |
20 services | QUANTITY | 0.99+ |
Prometheus | TITLE | 0.99+ |
second example | QUANTITY | 0.99+ |
second one | QUANTITY | 0.99+ |
five key | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
third one | QUANTITY | 0.99+ |
80-plus % | QUANTITY | 0.99+ |
eight settings | QUANTITY | 0.99+ |
Opsani | PERSON | 0.99+ |
third examples | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
services | QUANTITY | 0.99+ |
50 | QUANTITY | 0.99+ |
eight values | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
nine | QUANTITY | 0.98+ |
three elements | QUANTITY | 0.98+ |
Servo | ORGANIZATION | 0.98+ |
80% | QUANTITY | 0.98+ |
opsani.com/awsstartupshowcase | OTHER | 0.98+ |
first one | QUANTITY | 0.98+ |
two small instances | QUANTITY | 0.98+ |
10 | QUANTITY | 0.97+ |
three things | QUANTITY | 0.97+ |
once a month | QUANTITY | 0.97+ |
one time | QUANTITY | 0.97+ |
70% | QUANTITY | 0.97+ |
GDPR | TITLE | 0.97+ |
zero | QUANTITY | 0.97+ |
Servo | TITLE | 0.97+ |
about 20 steps | QUANTITY | 0.97+ |
12 | QUANTITY | 0.96+ |
Kubernetes | TITLE | 0.96+ |
four times | QUANTITY | 0.96+ |
Jordan Sher and Michael Fisher, OpsRamp | AWS Startup Showcase
(upbeat music) >> Hi, everyone. Welcome to today's session of theCUBE presentation of AWS Startup Showcase, the new breakthrough in DevOps, data analytics, cloud management tools, featuring OpsRamp for the cloud management migration track. I'm John Furrier, your hosts of theCUBE Today, we're joined by Jordan Sheer, vice president of corporate marketing and Michael Fisher, director of product management in OpsRamp. Gentlemen, thank you for joining us today for this topic of challenges of delivering availability for the modern enterprise. >> Thanks, John. >> Yeah, thanks for having us. >> Hey, so first of all, I have to congratulate you guys on the successful launch and growth of your company. You've been in the middle of the action of all this DevOps, microservices, cloud scale, and availability is the hottest topic right now. IT Ops, AI Ops, whoever you want to look at it, IT is automating a way in a lot of value. You guys are in the middle of it. Congratulations on that, and congratulations on being featured. Take a minute to explain what you guys do. What's the strategy? What's the vision? What's the platform. >> Yeah, I'll take that one. So I would just kind of take a step back and we look at the broader landscape of the ecosystem of tools that all sits in. There's a lot of promises and a lot of whats and features and functionality that are being announced. Three pillars of durability and all these tools are really trying to solve a fundamental problem we see in the market and this problem transcends the classic IT ops and it's really front and center, even in this modern DevOps market, this is the problem of availability. And so when we talk about availability, we don't just mean the four nines for an uptime metric, availability to the modern enterprise, is really about an application doing what it needs to do to serve the users in a way that works for the business. And I always like to have a classic example of an e-commerce site, right? So maybe you can get to an e-commerce sites online, but you can't add an item to a cart, right? Well, you can't do something that is a meaningful transaction for the business. And because of that, that experience is not available to you as a user and it's not available to the business because it didn't result in a positive outcome. So the promise of OpsRamp is really around this availability concept and the way we rationalize this as a three pillar formats. And so we think the three pillars of availability are the ability to observe data, this is the first piece of it all. And from a problem perspective, what we're really trying to say is do we have the right data at any given point in time to accurately diagnose, assess, and troubleshoot application behavior? And we see it as a huge problem with a lot of enterprises, because data that can be often siloed, too many tools, many teams, and each one has a slightly different understanding of application health. For example, the DevOps team may have a instance of Prometheus or they may have some other monitoring tool, or the IT team may have their own set, right? But when you have that kind of segmented view of the world, you're not really having the data in a central place to understand availability at the most holistic level, which is really from an end-user to that middleware, to the databases, to underlying microservices, which are really providing the end-user experience. So that observed problem is that first thing OpsRamp tries to solve. Secondly, this is the analyze phase, right? So analyze to us means are we giving the proper intelligence on top of the data to drive meaningful insights to this operator and user? And the promise here is that can we understand that baseline performance and potentially even mitigate future instance from happening? How often do we hear a cloud provider going down or some SaaS provider going down because of some microservice migration issue or some third party application or networking they're relying on? I can think of dozens on my head. So that's kind of the second piece. And then lastly is around this act. This is an area of a lot of investment for ops because we think this is the final pillar for nailing this availability problem. Because again, IT teams are not getting larger, they're getting smaller, right? Everyone's trying to do more with less. And so from a platform perspective, how do we enable teams to focus on the most business critical tasks, which are your cloud migrations, adopting microservices to run your modern applications, innovative projects. These are the things that IT and DevOps teams are tasked with. And maintaining availability is not something people want to do, that should be automated. And so when you think of automation, this is a big piece for us. So again, the key problem is how can we enable these IT or DevOps teams to focus on those business critical things, and automate it with the rest. And so this is the OpsRamp's three pillars of availability. >> John: Talk about the platform, if you don't mind. I know you've got a slide on this. I want to jump into it because this comes up a lot, availability's not just throughout uptime, because you know, uptime, five nine reliability is an old school concept. Now you have different kinds of services that might be up but slow, would cause some problems, as applications and this modern era have all these new sets of services. Can you go through and talked about the platform? >> Yeah, absolutely. So OpsRamp has a very... We address this availability problem pretty holistically, like I mentioned. From a platform perspective, there that two core lines that are comprising a product. One is this hybrid monitoring piece. This is that data layer. And the next one is event management, it's more of the we'll talk about that analysis. And so we treat the monitor as a direct feed into this event management. We're layering that on top, or layering machine learning and AI to augment the insights derived from that first pillar. And so this is where we see a really interesting intersection of data science and monitoring tools. We invest a lot in this area because there's a lot of meaningful problems to solve. In particular alert fatigue, or potentially root cause analysis, things that can take an operator or a developer a long time to do on their own, OpsRamp tries to augment that knowledge of your systems and applications so that you can get to the bottom of things faster and get on with your day. And so it's not just for the major outages, it's not just for the things that are on Twitter or CNN that's for daily things that can just distract you from the ability to do your job, which is to be a core innovator for a business. >> I will really say John, that we are already seeing some couple things here. Number one, we're already actually seeing fundamental transformations in the marketplace. Customers who have seen reduction in alert volumes of up to 95% in some cases, which is as you can imagine, that's completely transformational for these businesses. And number two, I think one of the promises of hybrid of observability working in tandem with event and incident management is the idea of finding unknown unknowns within your organization and being able to act upon them. All too many times nowadays, monitoring tools are there to just surface issues that you may know that you're looking for and then help you find it and then take action on them. But I think the idea of OpsRamp is that we really using that big data platform that Michael talks about is to really surface all the issues that you might not be able to see, identify the root cause, and then take action on those root causes. So in our world, application availability is a much more proactive activity where the IT operations team can actually be proactive about these incidents and then take action on them. >> Yes. Jordan, if you don't mind, I'm following up on that real quick. Talk about the difference uptime versus availability, because something could be up and reliable but not available and its services get flaky. Things may look like they're up and running. Can you just unpack that a little? >> So to me, I mean the really key aspect of availability that I think the old definition of uptime doesn't address is performance. That something can be up, but not performing, but still not really be available. And his e-commerce example, I think is a great one. Let's take, for example, you get on Amazon, right? The Amazon e-commerce experience is always available. And what that means is that at any given moment, when I want to click through the e-commerce experience, it performs. It's available. It's always there and I can buy it at any given time. If there's a latency issue, if the application has a lag, if it takes 30 seconds to really perform an activity on that application, in the alternative definition, that's not available anymore. Even though the application may be up, it's not performing, it's not providing a frictionless end customer experience, and it's not driving the business forward, and therefore it's not available. The definition of availability in OpsRamp is creating a meaningful customer experience that actually drives the business forward. So in that definition, if a service is up but it's latent, but it's not providing excellent customer experience that the business wants to promise to its end-user, it's not available. So that's really how we're redefining this whole notion of availability and we're urging our customers and people in the marketplace to do the same. Ask yourself the hard question, is your application available or is it just up? >> Yeah, and I think that the confluence of the business logic around what the outcome is, and I think this is the classic cliche, "Oh, it's all about outcomes." Here, you're saying that the outcome can be factored into the policy of the tech, meaning this is the experience we want for our users, our customers, and this is what we determined as acceptable and excellent. That's the new metric, so that's the new definition. You can almost flip the script. It feels like it's being flipped around. Is that the right way to think about it? >> Well, yeah, I think that's actually absolutely correct that an application needs to be business aware, especially in the modern day because all of the businesses that we work with, their applications are really the stock and trade of the business. And so if you create an application that is not business aware, that is just there for its own sake or is not performing according to the revenue goals or the targets of the business, then it's no longer available. >> I mean, it could be little things. It could be like an interface on the UI, it could be something really small or a microservice that's not getting to the database in time or some backup or some sort of high availability. Really interesting things could happen with microservices and DevOps, can you guys share some examples of what people might fall into from a trap standpoint or just from a bad architecture? What are some of the things that they might see in their environment that would say that they need help? >> Yeah, I can probably take that one. So there's a lot of, I call them symptoms of a bad availability experience. And I wouldn't even say it's a pure microservice specific thing. I would say it's really any application that's end-user phasing. I see similar pitfalls. One is a networking issue. I see the number one thing usually with these kinds of issues that networking or config changes that can cause environments to go down. And so when we talk to organizations get to the bottom of this is usually a config wasn't thought through thoroughly, or it was a QAed, they didn't have the proper controls in place. I would say that's probably the number one reasons I see applications go unavailable. I think that's some majority of DevOps teams that can empathize with that is someone did something and I didn't know, and it caused some applications servers go down and it causes cascading event of issues. That's like modern paradigm of issues. On old school days, it's a layer zero issue, someone unplugged something. Well, modern times it's someone pushed something I don't have an idea of what we're doing opposing a downstream effect it would have been and therefore my application went unavailable. So that's again, probably the number one pitfall. And again, I think the hardest problem in microservices still around networking, right? Enterprise level networking and connecting that with many data center applications. For example, Kubernetes, which is the provider or the opera orchestrator of any microservice is still getting to the level, many organizations are still getting a level of comfort with trusting production applications to run on it because one is a skill gap. There's not many large organizations have a huge Kubernetes application team, usually they're fairly small agile units. And so with that, there's a skill gaps, right? How do you network in Kubernetes? How do you persist in storage? How to make sure that your application has the proper security built into it, right? Because that these are all legacy problems kind of catching up with the modern environments, because just because you're modernizing, it doesn't mean these old problems go away. It just take a different form. >> Yeah. That's a great point. Modernization. You guys, can you guys talk about this modern application movement in context to how DevOps has risen really into providing value there? Certainly with cloud scale and how companies are dealing with the old legacy model of centralized IT or security teams who slow things down? Because one of the things that we're seeing in this market is speed, faster developer time to market, time to value. Especially if you're an e-commerce site, you're seeing potentially real-time impact. So you have the speed game on the application side that's actually good, being slowed down by lack of automation or just slow response to a policy or a change or an incident. I mean, this seems to be a big discussion. Can you guys share your thoughts on this and your reaction to that? >> I can tell you that one of the places that we are displacing, one of the markets that we are displacing is the legacy ITOM market, because it can't provide the speed that you're talking about, John. I think about a couple of specific examples. I won't necessarily name the providers, but there are several legacy item providers that for example, require an appliance. They require an appliance for you to administer IT operations management services. And that in and of itself is a much slower way of deploying item. Number two, they require this customized proof of value, proof of concept operation, where companies, enterprise organizations need to orchestrate the customization of the item platform for their use. You buy separate management packs that would integrate with different existing applications on your stack. To us, that's too slow. It means you have to make a bunch of decisions upfront about your item practice and then live with those decisions for years to come, especially with software licenses. So by even moving that entire operation to SaaS, which is what the OpsRamp platform has done, has accelerated the ability to drive availability for applications. Number two, and I'd like to pitch this over to Michael, because I think this is really fundamental to how OpsRamp is driving availability, is the use of artificial intelligence. So when we think about being proactive and we think about moving more quickly, it takes machine learning to do a lot of that work to be able to monitor alert streams and alert floods, especially with the smaller scale down IT teams that Michael has mentioned before. You need to harness the power of artificial intelligence to do some of that work. So those are two key ways that I see the platform driving additional speed, especially in a DevOps environment. And I'd love to hear as well from Michael, additional enhancements. >> Michael, if you don't mind, I'll add one thing. First of all, great call out there, Jordan. Yeah. So the legacy slow down, it's like say appliance or whatever that also impacts potentially the headroom on automation. So if you could also talk about the AI machine learning, AI piece, as well as how that impacts automation, because the end of the day automation is going to have to be lock step in with the AI. >> Yeah. And this kind of goes back to that OpsRamp three pillars of availability, right? So that's the what we do, but again, it's all goes back to the availability problem. But we see that observe, analyze, and act as a seamless flow, right? To have it under the same group or the same tent provides tremendous opportunity and value for our DevOps or IT Ops teams that trust the OpsRamp platform because I'm a big believer that garbage in, garbage out. Having the monitoring data in native or having this data native to your tool provides a lot of meaningful value for customers because they have their monitoring data, which is coming from the OpsRamp tool. They have the intelligence, which is being provided by their ops cube machine learning. And they have our process automation and workflow to feed off that directly. And so when I think of this modernization problem, I really think about modern DevOps teams and the problems they face, which is around doing more with less, that's kind of the paradigm of many teams, each one is trying to learn, how do I do security for Kubernetes? How do I observe my security in the Kubernetes' cluster? How do I make sure my CI/CD pipeline is set up in such a way that I don't need to monitor it, or I don't need to give it attention? And so having a really seamless flow from that observe, analyze, act enables those problems to be solved in a much more seamless way that I don't see many legacy providers be able to keep up with. >> Awesome. Jordan, if you don't mind, I'd love to get your definition of what modern availability means. >> Yeah. So, you know, as I've gone through a little bit previously, so modern availability to me is availability uptime. It's also performance, right? Is the app location marks set down by both the application team, but also by the business. And number three is it business aware. So a truly modern available application is being able, is driving an excellent customer experience according to the product roadmap, but it's also doing it in a way that moves the business forward. Right? And if your applications today are not meeting those benchmarks, if they're performing but they're not driving the business forward, if they're not performing, if they're not up, if they don't meet any one of those three core tenants, they're not truly available. And I think that what's most impactful to me about what the platform, what OpsRamp in particular does in today's environment is operating under that modern definition of available is more difficult than ever. It is more difficult because we are living in a hybrid, distributed, multi-cloud world with tons of software vendors that are being sold into these organizations today that are promising similar results. So when you're an IT operator, how do you drive availability in light of that kind of environment? You have reduced budget. You have greater complexity, you have more tools than ever, and yet your software is more impactful to the bottom line than ever before. It's in this environment that we took a hard look at what's going on in the world, and we say these operators need help driving availability. That's the germination of the OpsRamp platform. >> That's a great point. We're going to come into the culture. And the second Emily Freeman's keynote about the revolution in DevOps talks about this, multiple personas and multiple tools that drive specialism, specialties that actually don't help in the modern era. So I'm going to hold that for a second. We'll come to the cultural question in a minute. Michael, if you don't mind to pivot off that definition, what are the metrics? With all those tools out there, all these new things, what are the new metrics for modern availability? It's more than MTTR. >> Yeah. This whole metrics that I think people spend a lot of time on, I think it's actually people thinking in the wrong direction if you ask me. So I've seen a lot of work. People say that the red metrics, that rate error duration or its views, utilization, saturation errors, or it's these other more contrived application metrics. I think they're looking at a piece of the stack, they're not looking at the right things. Even things like mean time to resolve and critical and server response time, mean time to tech, those are all downstream indicators. I like to look at much more proactive signals. So things like app deck score, your application index, or application performance index, these are things that are much more end-user facing or even things like NPS score, right? This has never really been a classic metric for these operations teams, but what a NPS score shows you is are your users happy using your applications? Is your experience giving what they expect it to be? And usually when you ask these two questions, even if you ask the DevOps team do you know what your Atlas score is? And you use NPS score, but what are those, right? Because it's just never been in that conversation. Those have been more maybe on the business side or maybe on the product management side. But I think that as organizations modernize, we see a much more homogenous group forming among these DevOps and product units to answer these kinds of questions. That's something we focus a lot on OpsRamp it's not seeing the silo of DevOps product or Ops. We're each thinking of how do you have a better NPS and how do we drive a better app decks? Because those are our leading indicators of whether or not our applications available. >> So I want to ask you guys both before, again, back to the own cultural question I really want to get into, but from a customer standpoint, they're being bombarded with sales folks, "Hey, buy my tool. I got some monitoring over a year. I got AI ops. I got observability." I mean, there's a zillion venture back companies that just do observability, just monitoring, just AI Ops. As the modern error is here, what's going on in the psychology of the customer because they want to like clear the noise. We saw it in cybersecurity years ago. Right? They buy everything, and next thing you know, they're going to fog of tools. What's the current state of the customer? What do they need right now as to be positioned for the automation, for the edge, all these cool cloud-scale next gen opportunities? >> Yeah. So in my mind, it's basically three things, right? Customers, number one, they want a vision. They want a vision that understands their position in the enterprise organization and what the vision for application development is going to be moving forward. Number two, they don't want to be sold anymore. You're absolutely right. It's harder and harder to make a traditional enterprise sale nowadays. It's because there's a million vendors. They're just like us. They're trying to get people on the phone and it can be tough out there. And number three, they want to be able to validate on their own with their own time. So in light of that, we've introduced a free trial of our cloud monitoring. It's a lightweight version of the OpsRamp platform, but it is a hundred percent free right now. It is available for two weeks with an unlimited number of users and resource count. And you come in and you can get started on your own using preloaded infrastructure from us if you want, or you could bring your own infrastructure. And we can tell you that customers who onboard through the free trial can see insights on their infrastructure within 20 minutes of onboarding. And that experience in and of itself is a differentiator and it allows our customers to buy on their own terms and timelines. >> Sure. And that's a great point. We brought this up last quarter in the showcase, one of the VCs brought up and says he was an old school VC, kind of still in the game, but he was saying in the old days in shelf where you didn't know if it was going to be successful until like downstream, now it's SaaS. If a customer doesn't see the value immediately. It's there. I mean, there's no hiding. You cannot hide from the truth of value here in the modern era. That's a huge impact on how customers now are evaluating and making decisions. >> Absolutely. And you know, I don't think any customer out there wants to read it on the white paper on the state of enterprise IT anymore. We recognize that and so we are hyper-focused on driving value for our customers and prospects as fast as possible, and still providing them the control that they need to make decisions on their own terms. >> Michael, I've got to ask you, since you have the keys to the kingdom on the product management side, what's the priorities on your side for customers, obviously the pressure's there, you guys are doing great, customers try it out for free. They can get, see the value and then double down on it. That's the cloud way. That's what's DevOps all about. You have to prioritize the key things, what's going on with your world. >> Yeah. And I would say of course prod has their own perspective on this. Our number one goal right now is to accelerate that time to value. And so when we look at one who we're targeting, right? So there's DevOps user, this modern application of operator, what are their core concerns in the world? One is, again, that data problem. Are we bringing the right type of data to solve meaningful problems? And two, are we making insights out of that? So from my priority's perspective, we're really driving more focus on this time to value problem and reduced time to there's some key value metrics we have and I'll go to that, but it's all an effort to make sure that when they hit our platform and they use our platform, we're showing them their return on investment as fast as possible. And so, what a return on investment means (indistinct) can slightly vary, but we try to narrow focus on our key target persona and market and focused on them. So right now it definitely is on that modern DevOps team enterprise, looking to provide modern application availability. >> Awesome. Hey guys, for the last two minutes, I'd love to shift now to the culture. So Jordan, you mentioned that appliance, the item example, which is I think indicative of many scenarios in the legacy old world, old guard school, where there's a cultural shift where some people are pissed off, they're going to go and they slowing things down, right? So you see people that are unhappy, the sites having performance of an e-commerce sites, having five second delays or some impact to the business, and the developers are moving fast with DevOps. The DevOps has risen up now where it's driving the agenda. Kind of impacting the old school departments, whether it's security or IT, central groups that are responding in days and weeks to requests, not minutes. This is a huge cultural thing. What's your thoughts on this? >> I absolutely think it's true. I think the reason were options differ slightly on that is we do see the rise of DevOps culture and how it starts to take control and rest the customer experience back from the legacy providers within the organization, but we still see that there's value in having a foot in the old and a foot in the new, and it's why that term hybrid, we talked about hybrid observability is really important to us. It's true, DevOps culture has a lot of great reasons why it's taken over, right? Increases in speed, increases in quality, increases in innovation, all of that. And yet the enterprise is still heavily invested in the old way. And so what they are looking for is a platform to get them from the old way to the new way fast. And that's where we really shine. We say we can enable, we can work with the existing tool set that you have, and we can move you even more in the future of this new definition of availability. And we can get you that DevOps state of play even quicker. And so you don't have to make a heavy lift and you don't have to take a big gamble right now. You can still provide this kind of slow moving migration plan that you need to feel comfortable, and it doesn't force you to throw away a bunch of stuff. >> And if you guys can comment on whole day two operations, that's where the whole ops reliability thing comes in, right? This is kind of where we're at right now, Dev and Ops. Ops really driving the quality and reliability, availability and your definition. This is key, right? This is where we're starting to see the materialization of DevOps. >> It's why we have guys like Michael Fisher who are really driving our agenda forward, right? Because I think he represents the vision of the future that we all want to get to. And the platform that the product team in OpsRamp is building is there, right? But we also want to provide a path for day two, right? There are still some companies are living in day one and they want to get to day two. And so that's where we drive out here. >> And Michael, the platform with the things like containers really helps people get there. They don't have to kill the old to bring in the new, they can coexist. Can you quickly comment your reaction to that? >> Yeah, absolutely. And I talked to a lot of, I won't name any but large scale web companies, and they're actually balancing this today. They have some infrastructure or applications running on bare metal that somebody's got Kubernetes, and there's actually, it's not so much, everything has to go one direction. It actually is what makes the business, right? Even for migrating to the cloud, there has to be a compelling business reason to do so. And I think a lot of companies are realizing that for the application side as well. What runs where and how do we run it? Do we migrate a legacy monolith to a microservice? How fast do we do it? What's the business impact of doing it? These are all critical things that DevOps teams are engaged with on a daily basis as part of the core workflows, so that's my take on that. >> Guys. Great segment. Thanks for coming on and sharing that insight. Congratulates the OpsRamp, doing really extremely well, right in the right position on ramp for operations to be DevOps, whatever you want to call it, you guys are in the center of it with a platform. I think that's what people want, delivering on these availability, automation, AI. Congratulations and thanks for coming on theCUBE for the Showcase Summit. >> Thanks so much. >> Thank you so much, John. >> Okay, theCUBE's coverage of AWS showcase hottest startups in cloud. I'm John Furrier, your host. Thanks for watching. (relaxing music)
SUMMARY :
for the modern enterprise. and availability is the are the ability to observe data, of services that might be up from the ability to do your job, all the issues that you Talk about the difference and it's not driving the business forward, Is that the right way to think about it? because all of the businesses It could be like an interface on the UI, I see the number one thing usually I mean, this seems to be a big discussion. customization of the item platform So the legacy slow down, So that's the what we do, but again, I'd love to get your definition that moves the business forward. And the second Emily Freeman's keynote in the wrong direction if you ask me. for the automation, for the edge, of the OpsRamp platform, kind of still in the game, that they need to make on the product management side, that time to value. of many scenarios in the legacy in the future of this new Ops really driving the quality And the platform that the product team And Michael, the And I talked to a lot of, I won't name any for the Showcase Summit. I'm John Furrier, your
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Michael Fisher | PERSON | 0.99+ |
Jordan Sheer | PERSON | 0.99+ |
Michael | PERSON | 0.99+ |
John | PERSON | 0.99+ |
Jordan | PERSON | 0.99+ |
two weeks | QUANTITY | 0.99+ |
Emily Freeman | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
30 seconds | QUANTITY | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
second piece | QUANTITY | 0.99+ |
two questions | QUANTITY | 0.99+ |
last quarter | DATE | 0.99+ |
Jordan Sher | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
two key ways | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
one | QUANTITY | 0.99+ |
first piece | QUANTITY | 0.99+ |
CNN | ORGANIZATION | 0.99+ |
three things | QUANTITY | 0.99+ |
20 minutes | QUANTITY | 0.98+ |
Secondly | QUANTITY | 0.98+ |
two core lines | QUANTITY | 0.98+ |
first pillar | QUANTITY | 0.98+ |
three pillar | QUANTITY | 0.98+ |
Prometheus | TITLE | 0.98+ |
over a year | QUANTITY | 0.98+ |
hundred percent | QUANTITY | 0.98+ |
second | QUANTITY | 0.97+ |
Today | DATE | 0.97+ |
One | QUANTITY | 0.97+ |
up to 95% | QUANTITY | 0.97+ |
each one | QUANTITY | 0.97+ |
three core tenants | QUANTITY | 0.97+ |
OpsRamp | ORGANIZATION | 0.96+ |
dozens | QUANTITY | 0.96+ |
OpsRamp | TITLE | 0.96+ |
day two | QUANTITY | 0.96+ |
each | QUANTITY | 0.96+ |
two | QUANTITY | 0.96+ |
day one | QUANTITY | 0.95+ |
DevOps | ORGANIZATION | 0.95+ |
ORGANIZATION | 0.95+ | |
DevOps | TITLE | 0.95+ |
Kubernetes | TITLE | 0.94+ |
three pillars | QUANTITY | 0.94+ |
day | QUANTITY | 0.93+ |
Three pillars | QUANTITY | 0.93+ |
one direction | QUANTITY | 0.92+ |
Knox Anderson, Sysdig | CUBE Conversation
(soft electronic music) >> Welcome to this CUBE Conversation. I'm Lisa Martin. This conversation is part of our third AWS Startup Showcase for this year. I'm pleased to welcome Knox Anderson, the VP of Product Management at Sysdig. Knox, welcome to the program. >> Thanks for having me, Lisa. >> Talk to me a little bit about Sysdig, secure DevOps for containers, Kubernetes, and cloud. Give the audience an overview of what you guys do. >> So Sysdig is this secure DevOps platform that provides observability, security, and compliance functions for anyone that's adopting Kubernetes and Cloud. We really secure the entire lifecycle from source to production, so do things like scan your ISE for misconfiguration, monitor your runtime environments for threats and operational best practices. We provide a lot of capabilities around Prometheus Monitoring, as well, and then also let organizations perform incident response and compliance audits against these environments. >> So founded in 2013, talk to me about the gap in the market that you guys saw then and what some of the key challenges are that you saw for your customers. >> Yeah so we came to market around the same time as containers and Kubernetes and I'd say 2015 to 2018 we kept on saying it's the year of Kubernetes, it's the year of Kubernetes, it's the year of Kubernetes. And then really during the last year and a half in the COVID pandemic, Kubernetes has gone gangbusters. Every major cloud is seeing a huge adoption in their Kubernetes services so that's really our wedge into a lot of organizations. They're changing their platform to take advantages of containers and Kubernetes and you really have to rethink all of your security tooling, and that's when a company like Sysdig comes in. >> Talk to me about customers in terms of, especially in the last year and a half when things have been so dynamic, we've seen so much too, on the threat landscape front changing. Give me an example of a customer or two that you're really helped with solving some of their major challenges, here. >> Yeah, a great customer that we work with is SAP Concur and they kind of encompass a lot of the things that are nice about modern DevOps. So it's a DevOps team that's running a Kubernetes platform that thousands of developers are building their apps and deploying those onto. And they chose Sysdig because really it's not scalable to have every single data team ping that DevOps team and say what's the performance of my service, how is it responding, how can I get scanning integrated with that and so they use Sysdig as a platform that allows developers to easily onboard onto their Kubernetes clusters and then ensure that they're meeting compliance needs and FedRAMP needs for that platform that they deliver their core business apps on. >> Let's talk about the Sysdig's commitment to opensource on the Falco project. >> So Falco is a opensource project that we started at Sysdig, it's built on top of our core system core instrumentation. And so Falco meets a lot of your IDS or your file integrity monitoring requirements that you might have as you move to Kubernetes. And really, it's something we started at about 2016. In 2019, we donated that project to the CMCS which is the same governance body behind Kubernetes, Prometheus, and other kind of core building blocks of the climate of ecosystem. Since then, it's grown immensely. Companies like Shopify are using it to make sure that their PCI apps that they run Kubernetes are fully compliant. And so it's something that we are constantly contributing to the community also from even companies like AWS is a core contributor to the Falco project. And I'm really excited to see where it goes over the next year as Falco extends to also cover some cloud security use cases. >> What can you tell me about the relationship that Sysdig and AWS have? >> They've been a great partner. We internally run our SaaS on AWS so we're using AWS services to deliver our product to our customers. And then we've also really worked closely around how you can provide better security for services like Fargate. So we did working sessions with their engineering teams, learned what we could do to get the visibility that we need for tools like Falco and Sysdig to work seamlessly in Fargate environments. And last April we were able to kind of, AWS released that new functionality, Sysdig built on top of that, and we've already seen great adoption of customers using the Sysdig product on top of Fargate. >> Excellent. Well thank you very much, Knox, for stopping by theCUBE telling us about Sysdig, what you guys are doing ahead of the AWS Startup Showcase. We appreciate your time and your information. >> Thanks for having me. >> For Knox Anderson, I'm Lisa Martin. You're watching this CUBE Conversation. (soft electronic music)
SUMMARY :
I'm pleased to welcome Knox Anderson, Talk to me a little bit about Sysdig, We really secure the entire in the market that you and I'd say 2015 to 2018 in the last year and a that allows developers to easily onboard to opensource on the Falco project. that project to the CMCS get the visibility that we need ahead of the AWS Startup Showcase. (soft electronic music)
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Lisa Martin | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
2013 | DATE | 0.99+ |
2019 | DATE | 0.99+ |
Knox | PERSON | 0.99+ |
two | QUANTITY | 0.99+ |
Lisa | PERSON | 0.99+ |
2018 | DATE | 0.99+ |
Sysdig | ORGANIZATION | 0.99+ |
Knox Anderson | PERSON | 0.99+ |
2015 | DATE | 0.99+ |
Sysdig | PERSON | 0.99+ |
last April | DATE | 0.99+ |
next year | DATE | 0.99+ |
Shopify | ORGANIZATION | 0.98+ |
Kubernetes | TITLE | 0.98+ |
Falco | ORGANIZATION | 0.98+ |
thousands of developers | QUANTITY | 0.97+ |
COVID pandemic | EVENT | 0.97+ |
Sysdig | TITLE | 0.96+ |
CMCS | ORGANIZATION | 0.96+ |
Prometheus | TITLE | 0.96+ |
last year and a half | DATE | 0.94+ |
last year and a half | DATE | 0.89+ |
Startup Showcase | EVENT | 0.85+ |
FedRAMP | ORGANIZATION | 0.82+ |
this year | DATE | 0.79+ |
DevOps | TITLE | 0.78+ |
Cloud | TITLE | 0.77+ |
argate | ORGANIZATION | 0.76+ |
ISE | TITLE | 0.74+ |
2016 | DATE | 0.74+ |
single | QUANTITY | 0.73+ |
SAP Concur | ORGANIZATION | 0.72+ |
third | QUANTITY | 0.65+ |
about | DATE | 0.63+ |
CUBE | TITLE | 0.59+ |
Fargate | ORGANIZATION | 0.58+ |
Fargate | TITLE | 0.34+ |