Kubernetes on Any Infrastructure Top to Bottom Tutorials for Docker Enterprise Container Cloud
>>all right, We're five minutes after the hour. That's all aboard. Who's coming aboard? Welcome everyone to the tutorial track for our launchpad of them. So for the next couple of hours, we've got a SYRIZA videos and experts on hand to answer questions about our new product, Doctor Enterprise Container Cloud. Before we jump into the videos and the technology, I just want to introduce myself and my other emcee for the session. I'm Bill Milks. I run curriculum development for Mirant us on. And >>I'm Bruce Basil Matthews. I'm the Western regional Solutions architect for Moran Tissue esa and welcome to everyone to this lovely launchpad oven event. >>We're lucky to have you with us proof. At least somebody on the call knows something about your enterprise Computer club. Um, speaking of people that know about Dr Enterprise Container Cloud, make sure that you've got a window open to the chat for this session. We've got a number of our engineers available and on hand to answer your questions live as we go through these videos and disgusting problem. So that's us, I guess, for Dr Enterprise Container Cloud, this is Mirant asses brand new product for bootstrapping Doctor Enterprise Kubernetes clusters at scale Anything. The airport Abu's? >>No, just that I think that we're trying Thio. Uh, let's see. Hold on. I think that we're trying Teoh give you a foundation against which to give this stuff a go yourself. And that's really the key to this thing is to provide some, you know, many training and education in a very condensed period. So, >>yeah, that's exactly what you're going to see. The SYRIZA videos we have today. We're going to focus on your first steps with Dr Enterprise Container Cloud from installing it to bootstrapping your regional child clusters so that by the end of the tutorial content today, you're gonna be prepared to spin up your first documentary prize clusters using documented prize container class. So just a little bit of logistics for the session. We're going to run through these tutorials twice. We're gonna do one run through starting seven minutes ago up until I guess it will be ten fifteen Pacific time. Then we're gonna run through the whole thing again. So if you've got other colleagues that weren't able to join right at the top of the hour and would like to jump in from the beginning, ten. Fifteen Pacific time. We're gonna do the whole thing over again. So if you want to see the videos twice, you got public friends and colleagues that, you know you wanna pull in for a second chance to see this stuff, we're gonna do it all. All twice. Yeah, this session. Any any logistics I should add, Bruce that No, >>I think that's that's pretty much what we had to nail down here. But let's zoom dash into those, uh, feature films. >>Let's do Edmonds. And like I said, don't be shy. Feel free to ask questions in the chat or engineers and boosting myself are standing by to answer your questions. So let me just tee up the first video here and walk their cost. Yeah. Mhm. Yes. Sorry. And here we go. So our first video here is gonna be about installing the Doctor Enterprise Container Club Management cluster. So I like to think of the management cluster as like your mothership, right? This is what you're gonna use to deploy all those little child clusters that you're gonna use is like, Come on it as clusters downstream. So the management costs was always our first step. Let's jump in there >>now. We have to give this brief little pause >>with no good day video. Focus for this demo will be the initial bootstrap of the management cluster in the first regional clusters to support AWS deployments. The management cluster provides the core functionality, including identity management, authentication, infantry release version. The regional cluster provides the specific architecture provided in this case, eight of us and the Elsie um, components on the UCP Cluster Child cluster is the cluster or clusters being deployed and managed. The deployment is broken up into five phases. The first phase is preparing a big strap note on this dependencies on handling with download of the bridge struck tools. The second phase is obtaining America's license file. Third phase. Prepare the AWS credentials instead of the adduce environment. The fourth configuring the deployment, defining things like the machine types on the fifth phase. Run the bootstrap script and wait for the deployment to complete. Okay, so here we're sitting up the strap node, just checking that it's clean and clear and ready to go there. No credentials already set up on that particular note. Now we're just checking through AWS to make sure that the account we want to use we have the correct credentials on the correct roles set up and validating that there are no instances currently set up in easy to instance, not completely necessary, but just helps keep things clean and tidy when I am perspective. Right. So next step, we're just going to check that we can, from the bootstrap note, reach more antis, get to the repositories where the various components of the system are available. They're good. No areas here. Yeah, right now we're going to start sitting at the bootstrap note itself. So we're downloading the cars release, get get cars, script, and then next, we're going to run it. I'm in. Deploy it. Changing into that big struck folder. Just making see what's there. Right now we have no license file, so we're gonna get the license filed. Oh, okay. Get the license file through the more antis downloads site, signing up here, downloading that license file and putting it into the Carisbrook struck folder. Okay, Once we've done that, we can now go ahead with the rest of the deployment. See that the follow is there. Uh, huh? That's again checking that we can now reach E C two, which is extremely important for the deployment. Just validation steps as we move through the process. All right, The next big step is valid in all of our AWS credentials. So the first thing is, we need those route credentials which we're going to export on the command line. This is to create the necessary bootstrap user on AWS credentials for the completion off the deployment we're now running an AWS policy create. So it is part of that is creating our Food trucks script, creating the mystery policy files on top of AWS, Just generally preparing the environment using a cloud formation script you'll see in a second will give a new policy confirmations just waiting for it to complete. Yeah, and there is done. It's gonna have a look at the AWS console. You can see that we're creative completed. Now we can go and get the credentials that we created Today I am console. Go to that new user that's being created. We'll go to the section on security credentials and creating new keys. Download that information media Access key I D and the secret access key. We went, Yeah, usually then exported on the command line. Okay. Couple of things to Notre. Ensure that you're using the correct AWS region on ensure that in the conflict file you put the correct Am I in for that region? I'm sure you have it together in a second. Yes. Okay, that's the key. Secret X key. Right on. Let's kick it off. Yeah, So this process takes between thirty and forty five minutes. Handles all the AWS dependencies for you, and as we go through, the process will show you how you can track it. Andi will start to see things like the running instances being created on the west side. The first phase off this whole process happening in the background is the creation of a local kind based bootstrapped cluster on the bootstrap node that clusters then used to deploy and manage all the various instances and configurations within AWS. At the end of the process, that cluster is copied into the new cluster on AWS and then shut down that local cluster essentially moving itself over. Okay. Local clusters boat just waiting for the various objects to get ready. Standard communities objects here Okay, so we speed up this process a little bit just for demonstration purposes. Yeah. There we go. So first note is being built the best in host. Just jump box that will allow us access to the entire environment. Yeah, In a few seconds, we'll see those instances here in the US console on the right. Um, the failures that you're seeing around failed to get the I. P for Bastian is just the weight state while we wait for a W s to create the instance. Okay. Yes. Here, beauty there. Okay. Mhm. Okay. Yeah, yeah. Okay. On there. We got question. Host has been built on three instances for the management clusters have now been created. We're going through the process of preparing. Those nodes were now copying everything over. See that? The scaling up of controllers in the big Strap cluster? It's indicating that we're starting all of the controllers in the new question. Almost there. Yeah. Yeah, just waiting for key. Clark. Uh huh. Start to finish up. Yeah. No. What? Now we're shutting down control this on the local bootstrap node on preparing our I. D. C. Configuration. Fourth indication, soon as this is completed. Last phase will be to deploy stack light into the new cluster the last time Monitoring tool set way Go stack like to plan It has started. Mhm coming to the end of the deployment Mountain. Yeah, America. Final phase of the deployment. Onda, We are done. Okay, You'll see. At the end they're providing us the details of you. I log in so there's a keeper clogging. You can modify that initial default password is part of the configuration set up with one documentation way. Go Councils up way can log in. Yeah, yeah, thank you very much for watching. >>Excellent. So in that video are wonderful field CTO Shauna Vera bootstrapped up management costume for Dr Enterprise Container Cloud Bruce, where exactly does that leave us? So now we've got this management costume installed like what's next? >>So primarily the foundation for being able to deploy either regional clusters that will then allow you to support child clusters. Uh, comes into play the next piece of what we're going to show, I think with Sean O'Mara doing this is the child cluster capability, which allows you to then deploy your application services on the local cluster. That's being managed by the ah ah management cluster that we just created with the bootstrap. >>Right? So this cluster isn't yet for workloads. This is just for bootstrapping up the downstream clusters. Those or what we're gonna use for workings. >>Exactly. Yeah. And I just wanted to point out, since Sean O'Mara isn't around, toe, actually answer questions. I could listen to that guy. Read the phone book, and it would be interesting, but anyway, you can tell him I said that >>he's watching right now, Crusoe. Good. Um, cool. So and just to make sure I understood what Sean was describing their that bootstrap er knows that you, like, ran document fresh pretender Cloud from to begin with. That's actually creating a kind kubernetes deployment kubernetes and Docker deployment locally. That then hits the AWS a p i in this example that make those e c two instances, and it makes like a three manager kubernetes cluster there, and then it, like, copies itself over toe those communities managers. >>Yeah, and and that's sort of where the transition happens. You can actually see it. The output that when it says I'm pivoting, I'm pivoting from my local kind deployment of cluster AP, I toothy, uh, cluster, that's that's being created inside of AWS or, quite frankly, inside of open stack or inside of bare metal or inside of it. The targeting is, uh, abstracted. Yeah, but >>those air three environments that we're looking at right now, right? Us bare metal in open staff environments. So does that kind cluster on the bootstrap er go away afterwards. You don't need that afterwards. Yeah, that is just temporary. To get things bootstrapped, then you manage things from management cluster on aws in this example? >>Yeah. Yeah. The seed, uh, cloud that post the bootstrap is not required anymore. And there's no, uh, interplay between them after that. So that there's no dependencies on any of the clouds that get created thereafter. >>Yeah, that actually reminds me of how we bootstrapped doctor enterprise back in the day, be a temporary container that would bootstrap all the other containers. Go away. It's, uh, so sort of a similar, similar temporary transient bootstrapping model. Cool. Excellent. What will convict there? It looked like there wasn't a ton, right? It looked like you had to, like, set up some AWS parameters like credentials and region and stuff like that. But other than that, that looked like heavily script herbal like there wasn't a ton of point and click there. >>Yeah, very much so. It's pretty straightforward from a bootstrapping standpoint, The config file that that's generated the template is fairly straightforward and targeted towards of a small medium or large, um, deployment. And by editing that single file and then gathering license file and all of the things that Sean went through, um, that that it makes it fairly easy to script >>this. And if I understood correctly as well that three manager footprint for your management cluster, that's the minimum, right. We always insist on high availability for this management cluster because boy do not wanna see oh, >>right, right. And you know, there's all kinds of persistent data that needs to be available, regardless of whether one of the notes goes down or not. So we're taking care of all of that for you behind the scenes without you having toe worry about it as a developer. >>No, I think there's that's a theme that I think will come back to throughout the rest of this tutorial session today is there's a lot of there's a lot of expertise baked him to Dr Enterprise Container Cloud in terms of implementing best practices for you like the defaulter, just the best practices of how you should be managing these clusters, Miss Seymour. Examples of that is the day goes on. Any interesting questions you want to call out from the chap who's >>well, there was. Yeah, yeah, there was one that we had responded to earlier about the fact that it's a management cluster that then conduce oh, either the the regional cluster or a local child molester. The child clusters, in each case host the application services, >>right? So at this point, we've got, in some sense, like the simplest architectures for our documentary prize Container Cloud. We've got the management cluster, and we're gonna go straight with child cluster. In the next video, there's a more sophisticated architecture, which will also proper today that inserts another layer between those two regional clusters. If you need to manage regions like across a BS, reads across with these documents anything, >>yeah, that that local support for the child cluster makes it a lot easier for you to manage the individual clusters themselves and to take advantage of our observation. I'll support systems a stack light and things like that for each one of clusters locally, as opposed to having to centralize thumb >>eso. It's a couple of good questions. In the chat here, someone was asking for the instructions to do this themselves. I strongly encourage you to do so. That should be in the docks, which I think Dale helpfully thank you. Dale provided links for that's all publicly available right now. So just head on in, head on into the docks like the Dale provided here. You can follow this example yourself. All you need is a Mirante license for this and your AWS credentials. There was a question from many a hear about deploying this toe azure. Not at G. Not at this time. >>Yeah, although that is coming. That's going to be in a very near term release. >>I didn't wanna make promises for product, but I'm not too surprised that she's gonna be targeted. Very bracing. Cool. Okay. Any other thoughts on this one does. >>No, just that the fact that we're running through these individual pieces of the steps Well, I'm sure help you folks. If you go to the link that, uh, the gentleman had put into the chat, um, giving you the step by staff. Um, it makes it fairly straightforward to try this yourselves. >>E strongly encourage that, right? That's when you really start to internalize this stuff. OK, but before we move on to the next video, let's just make sure everyone has a clear picture in your mind of, like, where we are in the life cycle here creating this management cluster. Just stop me if I'm wrong. Who's creating this management cluster is like, you do that once, right? That's when your first setting up your doctor enterprise container cloud environment of system. What we're going to start seeing next is creating child clusters and this is what you're gonna be doing over and over and over again. When you need to create a cluster for this Deb team or, you know, this other team river it is that needs commodity. Doctor Enterprise clusters create these easy on half will. So this was once to set up Dr Enterprise Container Cloud Child clusters, which we're going to see next. We're gonna do over and over and over again. So let's go to that video and see just how straightforward it is to spin up a doctor enterprise cluster for work clothes as a child cluster. Undocumented brands contain >>Hello. In this demo, we will cover the deployment experience of creating a new child cluster, the scaling of the cluster and how to update the cluster. When a new version is available, we begin the process by logging onto the you I as a normal user called Mary. Let's go through the navigation of the U I so you can switch. Project Mary only has access to development. Get a list of the available projects that you have access to. What clusters have been deployed at the moment there. Nan Yes, this H Keys Associate ID for Mary into her team on the cloud credentials that allow you to create access the various clouds that you can deploy clusters to finally different releases that are available to us. We can switch from dark mode to light mode, depending on your preferences, Right? Let's now set up semester search keys for Mary so she can access the notes and machines again. Very simply, had Mississippi key give it a name, we copy and paste our public key into the upload key block. Or we can upload the key if we have the file available on our local machine. A simple process. So to create a new cluster, we define the cluster ad management nodes and add worker nodes to the cluster. Yeah, again, very simply, you go to the clusters tab. We hit the create cluster button. Give the cluster name. Yeah, Andi, select the provider. We only have access to AWS in this particular deployment, so we'll stick to AWS. What's like the region in this case? US West one release version five point seven is the current release Onda Attach. Mary's Key is necessary Key. We can then check the rest of the settings, confirming the provider Any kubernetes c r D r I p address information. We can change this. Should we wish to? We'll leave it default for now on. Then what components? A stack light I would like to deploy into my Custer. For this. I'm enabling stack light on logging on Aiken. Sit up the retention sizes Attention times on. Even at this stage, at any customer alerts for the watchdogs. E consider email alerting which I will need my smart host details and authentication details. Andi Slack Alerts. Now I'm defining the cluster. All that's happened is the cluster's been defined. I now need to add machines to that cluster. I'll begin by clicking the create machine button within the cluster definition. Oh, select manager, Select the number of machines. Three is the minimum. Select the instant size that I'd like to use from AWS and very importantly, ensure correct. Use the correct Am I for the region. I commend side on the route device size. There we go, my three machines obviously creating. I now need to add some workers to this custom. So I go through the same process this time once again, just selecting worker. I'll just add to once again, the AM is extremely important. Will fail if we don't pick the right, Am I for a boon to machine in this case and the deployment has started. We can go and check on the bold status are going back to the clusters screen on clicking on the little three dots on the right. We get the cluster info and the events, so the basic cluster info you'll see pending their listen cluster is still in the process of being built. We kick on, the events will get a list of actions that have been completed This part of the set up of the cluster. So you can see here we've created the VPC. We've created the sub nets on We've created the Internet gateway. It's unnecessary made of us and we have no warnings of the stage. Yeah, this will then run for a while. We have one minute past waken click through. We can check the status of the machine bulls as individuals so we can check the machine info, details of the machines that we've assigned, right? Mhm Onda. See any events pertaining to the machine areas like this one on normal? Yeah. Just watch asked. The community's components are waiting for the machines to start. Go back to Custer's. Okay, right. Because we're moving ahead now. We can see we have it in progress. Five minutes in new Matt Gateway on the stage. The machines have been built on assigned. I pick up the U. S. Thank you. Yeah. There we go. Machine has been created. See the event detail and the AWS. I'd for that machine. Mhm. No speeding things up a little bit. This whole process and to end takes about fifteen minutes. Run the clock forward, you'll notice is the machines continue to bold the in progress. We'll go from in progress to ready. A soon as we got ready on all three machines, the managers on both workers way could go on and we could see that now we reached the point where the cluster itself is being configured. Mhm, mhm. And then we go. Cluster has been deployed. So once the classes deployed, we can now never get around our environment. Okay, Are cooking into configure cluster We could modify their cluster. We could get the end points for alert alert manager on See here The griffon occupying and Prometheus are still building in the background but the cluster is available on you would be able to put workloads on it the stretch to download the cube conflict so that I can put workloads on it. It's again three little dots in the right for that particular cluster. If the download cube conflict give it my password, I now have the Q conflict file necessary so that I can access that cluster Mhm all right Now that the build is fully completed, we can check out cluster info on. We can see that Allow the satellite components have been built. All the storage is there, and we have access to the CPU. I So if we click into the cluster, we can access the UCP dashboard, right? Shit. Click the signing with Detroit button to use the SSO on. We give Mary's possible to use the name once again. Thing is, an unlicensed cluster way could license at this point. Or just skip it on. There. We have the UCP dashboard. You can see that has been up for a little while. We have some data on the dashboard going back to the console. We can now go to the griffon, a data just being automatically pre configured for us. We can switch and utilized a number of different dashboards that have already been instrumented within the cluster. So, for example, communities cluster information, the name spaces, deployments, nodes. Mhm. So we look at nodes. If we could get a view of the resource is utilization of Mrs Custer is very little running in it. Yeah. General dashboard of Cuba navies cluster one of this is configurable. You can modify these for your own needs, or add your own dashboards on de scoped to the cluster. So it is available to all users who have access to this specific cluster, all right to scale the cluster on to add a notice. A simple is the process of adding a mode to the cluster, assuming we've done that in the first place. So we go to the cluster, go into the details for the cluster we select, create machine. Once again, we need to be ensure that we put the correct am I in and any other functions we like. You can create different sized machines so it could be a larger node. Could be bigger disks and you'll see that worker has been added from the provisioning state on shortly. We will see the detail off that worker as a complete to remove a note from a cluster. Once again, we're going to the cluster. We select the node would like to remove. Okay, I just hit delete On that note. Worker nodes will be removed from the cluster using according and drawing method to ensure that your workouts are not affected. Updating a cluster. When an update is available in the menu for that particular cluster, the update button will become available. And it's a simple as clicking the button, validating which release you would like to update to. In this case, the next available releases five point seven point one. Here I'm kicking the update by in the background We will coordinate. Drain each node slowly go through the process of updating it. Andi update will complete depending on what the update is as quickly as possible. Girl, we go. The notes being rebuilt in this case impacted the manager node. So one of the manager nodes is in the process of being rebuilt. In fact, to in this case, one has completed already on In a few minutes we'll see that there are great has been completed. There we go. Great. Done. Yeah. If you work loads of both using proper cloud native community standards, there will be no impact. >>Excellent. So at this point, we've now got a cluster ready to start taking our communities of workloads. He started playing or APs to that costume. So watching that video, the thing that jumped out to me at first Waas like the inputs that go into defining this workload cost of it. All right, so we have to make sure we were using on appropriate am I for that kind of defines the substrate about what we're gonna be deploying our cluster on top of. But there's very little requirements. A so far as I could tell on top of that, am I? Because Docker enterprise Container Cloud is gonna bootstrap all the components that you need. That s all we have is kind of kind of really simple bunch box that we were deploying these things on top of so one thing that didn't get dug into too much in the video. But it's just sort of implied. Bruce, maybe you can comment on this is that release that Shawn had to choose for his, uh, for his cluster in creating it. And that release was also the thing we had to touch. Wanted to upgrade part cluster. So you have really sharp eyes. You could see at the end there that when you're doing the release upgrade enlisted out a stack of components docker, engine, kubernetes, calico, aled, different bits and pieces that go into, uh, go into one of these commodity clusters that deploy. And so, as far as I can tell in that case, that's what we mean by a release. In this sense, right? It's the validated stack off container ization and orchestration components that you know we've tested out and make sure it works well, introduction environments. >>Yeah, and and And that's really the focus of our effort is to ensure that any CVS in any of the stack are taken care of that there is a fixes air documented and up streamed to the open stack community source community, um, and and that, you know, then we test for the scaling ability and the reliability in high availability configuration for the clusters themselves. The hosts of your containers. Right. And I think one of the key, uh, you know, benefits that we provide is that ability to let you know, online, high. We've got an update for you, and it's fixes something that maybe you had asked us to fix. Uh, that all comes to you online as your managing your clusters, so you don't have to think about it. It just comes as part of the product. >>You just have to click on Yes. Please give me that update. Uh, not just the individual components, but again. It's that it's that validated stack, right? Not just, you know, component X, y and Z work. But they all work together effectively Scalable security, reliably cool. Um, yeah. So at that point, once we started creating that workload child cluster, of course, we bootstrapped good old universal control plane. Doctor Enterprise. On top of that, Sean had the classic comment there, you know? Yeah. Yeah. You'll see a little warnings and errors or whatever. When you're setting up, UCP don't handle, right, Just let it do its job, and it will converge all its components, you know, after just just a minute or two. But we saw in that video, we sped things up a little bit there just we didn't wait for, you know, progress fighters to complete. But really, in real life, that whole process is that anything so spend up one of those one of those fosters so quite quite quick. >>Yeah, and and I think the the thoroughness with which it goes through its process and re tries and re tries, uh, as you know, and it was evident when we went through the initial ah video of the bootstrapping as well that the processes themselves are self healing, as they are going through. So they will try and retry and wait for the event to complete properly on. And once it's completed properly, then it will go to the next step. >>Absolutely. And the worst thing you could do is panic at the first warning and start tearing things that don't don't do that. Just don't let it let it heal. Let take care of itself. And that's the beauty of these manage solutions is that they bake in a lot of subject matter expertise, right? The decisions that are getting made by those containers is they're bootstrapping themselves, reflect the expertise of the Mirant ISS crew that has been developing this content in these two is free for years and years now, over recognizing humanities. One cool thing there that I really appreciate it actually that it adds on top of Dr Enterprise is that automatic griffon a deployment as well. So, Dr Enterprises, I think everyone knows has had, like, some very high level of statistics baked into its dashboard for years and years now. But you know our customers always wanted a double click on that right to be able to go a little bit deeper. And Griffon are really addresses that it's built in dashboards. That's what's really nice to see. >>Yeah, uh, and all of the alerts and, uh, data are actually captured in a Prometheus database underlying that you have access to so that you are allowed to add new alerts that then go out to touch slack and say hi, You need to watch your disk space on this machine or those kinds of things. Um, and and this is especially helpful for folks who you know, want to manage the application service layer but don't necessarily want to manage the operations side of the house. So it gives them a tool set that they can easily say here, Can you watch these for us? And Miran tas can actually help do that with you, So >>yeah, yeah, I mean, that's just another example of baking in that expert knowledge, right? So you can leverage that without tons and tons of a long ah, long runway of learning about how to do that sort of thing. Just get out of the box right away. There was the other thing, actually, that you could sleep by really quickly if you weren't paying close attention. But Sean mentioned it on the video. And that was how When you use dark enterprise container cloud to scale your cluster, particularly pulling a worker out, it doesn't just like Territo worker down and forget about it. Right? Is using good communities best practices to cordon and drain the No. So you aren't gonna disrupt your workloads? You're going to just have a bunch of containers instantly. Excellent crash. You could really carefully manage the migration of workloads off that cluster has baked right in tow. How? How? Document? The brass container cloud is his handling cluster scale. >>Right? And And the kubernetes, uh, scaling methodology is is he adhered to with all of the proper techniques that ensure that it will tell you. Wait, you've got a container that actually needs three, uh, three, uh, instances of itself. And you don't want to take that out, because that node, it means you'll only be able to have to. And we can't do that. We can't allow that. >>Okay, Very cool. Further thoughts on this video. So should we go to the questions. >>Let's let's go to the questions >>that people have. Uh, there's one good one here, down near the bottom regarding whether an a p I is available to do this. So in all these demos were clicking through this web. You I Yes, this is all a p. I driven. You could do all of this. You know, automate all this away is part of the CSC change. Absolutely. Um, that's kind of the point, right? We want you to be ableto spin up. Come on. I keep calling them commodity clusters. What I mean by that is clusters that you can create and throw away. You know, easily and automatically. So everything you see in these demos eyes exposed to FBI? >>Yeah. In addition, through the standard Cube cuddle, Uh, cli as well. So if you're not a programmer, but you still want to do some scripting Thio, you know, set up things and deploy your applications and things. You can use this standard tool sets that are available to accomplish that. >>There is a good question on scale here. So, like, just how many clusters and what sort of scale of deployments come this kind of support our engineers report back here that we've done in practice up to a Zeman ia's like two hundred clusters. We've deployed on this with two hundred fifty nodes in a cluster. So were, you know, like like I said, hundreds, hundreds of notes, hundreds of clusters managed by documented press container fall and then those downstream clusters, of course, subject to the usual constraints for kubernetes, right? Like default constraints with something like one hundred pods for no or something like that. There's a few different limitations of how many pods you can run on a given cluster that comes to us not from Dr Enterprise Container Cloud, but just from the underlying kubernetes distribution. >>Yeah, E. I mean, I don't think that we constrain any of the capabilities that are available in the, uh, infrastructure deliveries, uh, service within the goober Netease framework. So were, you know, But we are, uh, adhering to the standards that we would want to set to make sure that we're not overloading a node or those kinds of things, >>right. Absolutely cool. Alright. So at this point, we've got kind of a two layered our protection when we are management cluster, but we deployed in the first video. Then we use that to deploy one child clustering work, classroom, uh, for more sophisticated deployments where we might want to manage child clusters across multiple regions. We're gonna add another layer into our architectural we're gonna add in regional cluster management. So this idea you're gonna have the single management cluster that we started within the first video. On the next video, we're gonna learn how to spin up a regional clusters, each one of which would manage, for example, a different AWS uh, US region. So let me just pull out the video for that bill. We'll check it out for me. Mhm. >>Hello. In this demo, we will cover the deployment of additional regional management. Cluster will include a brief architectures of you how to set up the management environment, prepare for the deployment deployment overview and then just to prove it, to play a regional child cluster. So, looking at the overall architecture, the management cluster provides all the core functionality, including identity management, authentication, inventory and release version. ING Regional Cluster provides the specific architecture provider in this case AWS on the LCN components on the D you speak Cluster for child cluster is the cluster or clusters being deployed and managed? Okay, so why do you need a regional cluster? Different platform architectures, for example aws who have been stack even bare metal to simplify connectivity across multiple regions handle complexities like VPNs or one way connectivity through firewalls, but also help clarify availability zones. Yeah. Here we have a view of the regional cluster and how it connects to the management cluster on their components, including items like the LCN cluster Manager we also Machine Manager were held. Mandel are managed as well as the actual provider logic. Mhm. Okay, we'll begin by logging on Is the default administrative user writer. Okay, once we're in there, we'll have a look at the available clusters making sure we switch to the default project which contains the administration clusters. Here we can see the cars management cluster, which is the master controller. And you see, it only has three nodes, three managers, no workers. Okay, if we look at another regional cluster similar to what we're going to deploy now, also only has three managers once again, no workers. But as a comparison, here's a child cluster This one has three managers, but also has additional workers associate it to the cluster. All right, we need to connect. Tell bootstrap note. Preferably the same note that used to create the original management plaster. It's just on AWS, but I still want to machine. All right. A few things we have to do to make sure the environment is ready. First thing we're going to see go into route. We'll go into our releases folder where we have the kozberg struck on. This was the original bootstrap used to build the original management cluster. Yeah, we're going to double check to make sure our cube con figures there once again, the one created after the original customers created just double check. That cute conflict is the correct one. Does point to the management cluster. We're just checking to make sure that we can reach the images that everything is working. A condom. No damages waken access to a swell. Yeah. Next we're gonna edit the machine definitions. What we're doing here is ensuring that for this cluster we have the right machine definitions, including items like the am I. So that's found under the templates AWS directory. We don't need to edit anything else here. But we could change items like the size of the machines attempts. We want to use that The key items to ensure where you changed the am I reference for the junta image is the one for the region in this case AWS region for utilizing this was no construct deployment. We have to make sure we're pointing in the correct open stack images. Yeah, okay. Set the correct and my save file. Now we need to get up credentials again. When we originally created the bootstrap cluster, we got credentials from eight of the U. S. If we hadn't done this, we would need to go through the u A. W s set up. So we're just exporting the AWS access key and I d. What's important is CAAs aws enabled equals. True. Now we're sitting the region for the new regional cluster. In this case, it's Frankfurt on exporting our cube conflict that we want to use for the management cluster. When we looked at earlier Yeah, now we're exporting that. Want to call the cluster region Is Frank Foods Socrates Frankfurt yet trying to use something descriptive It's easy to identify. Yeah, and then after this, we'll just run the bootstrap script, which will complete the deployment for us. Bootstrap of the regional cluster is quite a bit quicker than the initial management clusters. There are fewer components to be deployed. Um, but to make it watchable, we've spent it up. So we're preparing our bootstrap cluster on the local bootstrap node. Almost ready on. We started preparing the instances at W s and waiting for that bastard and no to get started. Please. The best you nerd Onda. We're also starting to build the actual management machines they're now provisioning on. We've reached the point where they're actually starting to deploy. Dr. Enterprise, this is probably the longest face. Yeah, seeing the second that all the nerds will go from the player deployed. Prepare, prepare. Yeah, You'll see their status changes updates. He was the first night ready. Second, just applying second already. Both my time. No waiting from home control. Let's become ready. Removing cluster the management cluster from the bootstrap instance into the new cluster running the date of the U. S. All my stay. Ah, now we're playing Stockland. Switch over is done on. Done. Now I will build a child cluster in the new region very, very quickly to find the cluster will pick. Our new credential has shown up. We'll just call it Frankfurt for simplicity a key and customs to find. That's the machine. That cluster stop with three managers. Set the correct Am I for the region? Yeah, Do the same to add workers. There we go test the building. Yeah. Total bill of time Should be about fifteen minutes. Concedes in progress. It's going to expect this up a little bit. Check the events. We've created all the dependencies, machine instances, machines, a boat shortly. We should have a working cluster in Frankfurt region. Now almost a one note is ready from management. Two in progress. Yeah, on we're done. Clusters up and running. Yeah. >>Excellent. So at this point, we've now got that three tier structure that we talked about before the video. We got that management cluster that we do strapped in the first video. Now we have in this example to different regional clustering one in Frankfurt, one of one management was two different aws regions. And sitting on that you can do Strap up all those Doctor enterprise costumes that we want for our work clothes. >>Yeah, that's the key to this is to be able to have co resident with your actual application service enabled clusters the management co resident with it so that you can, you know, quickly access that he observation Elson Surfboard services like the graph, Ana and that sort of thing for your particular region. A supposed to having to lug back into the home. What did you call it when we started >>the mothership? >>The mothership. Right. So we don't have to go back to the mother ship. We could get >>it locally. Yeah, when, like to that point of aggregating things under a single pane of glass? That's one thing that again kind of sailed by in the demo really quickly. But you'll notice all your different clusters were on that same cluster. Your pain on your doctor Enterprise Container Cloud management. Uh, court. Right. So both your child clusters for running workload and your regional clusters for bootstrapping. Those child clusters were all listed in the same place there. So it's just one pane of glass to go look for, for all of your clusters, >>right? And, uh, this is kind of an important point. I was, I was realizing, as we were going through this. All of the mechanics are actually identical between the bootstrapped cluster of the original services and the bootstrapped cluster of the regional services. It's the management layer of everything so that you only have managers, you don't have workers and that at the child cluster layer below the regional or the management cluster itself, that's where you have the worker nodes. And those are the ones that host the application services in that three tiered architecture that we've now defined >>and another, you know, detail for those that have sharp eyes. In that video, you'll notice when deploying a child clusters. There's not on Lee. A minimum of three managers for high availability management cluster. You must have at least two workers that's just required for workload failure. It's one of those down get out of work. They could potentially step in there, so your minimum foot point one of these child clusters is fine. Violence and scalable, obviously, from a >>That's right. >>Let's take a quick peek of the questions here, see if there's anything we want to call out, then we move on to our last want to my last video. There's another question here about, like where these clusters can live. So again, I know these examples are very aws heavy. Honestly, it's just easy to set up down on the other us. We could do things on bare metal and, uh, open stack departments on Prem. That's what all of this still works in exactly the same way. >>Yeah, the, uh, key to this, especially for the the, uh, child clusters, is the provision hers? Right? See you establish on AWS provision or you establish a bare metal provision or you establish a open stack provision. Or and eventually that list will include all of the other major players in the cloud arena. But you, by selecting the provision or within your management interface, that's where you decide where it's going to be hosted, where the child cluster is to be hosted. >>Speaking off all through a child clusters. Let's jump into our last video in the Siri's, where we'll see how to spin up a child cluster on bare metal. >>Hello. This demo will cover the process of defining bare metal hosts and then review the steps of defining and deploying a bare metal based doctor enterprise cluster. So why bare metal? Firstly, it eliminates hyper visor overhead with performance boost of up to thirty percent. Provides direct access to GP use, prioritize for high performance wear clothes like machine learning and AI, and supports high performance workloads like network functions, virtualization. It also provides a focus on on Prem workloads, simplifying and ensuring we don't need to create the complexity of adding another opera visor. Lay it between so continue on the theme Why Communities and bare metal again Hyper visor overhead. Well, no virtualization overhead. Direct access to hardware items like F p G A s G p us. We can be much more specific about resource is required on the nodes. No need to cater for additional overhead. Uh, we can handle utilization in the scheduling. Better Onda we increase the performances and simplicity of the entire environment as we don't need another virtualization layer. Yeah, In this section will define the BM hosts will create a new project will add the bare metal hosts, including the host name. I put my credentials I pay my address the Mac address on then provide a machine type label to determine what type of machine it is for later use. Okay, let's get started. So well again. Was the operator thing. We'll go and we'll create a project for our machines to be a member off helps with scoping for later on for security. I begin the process of adding machines to that project. Yeah. So the first thing we had to be in post, Yeah, many of the machine A name. Anything you want, que experimental zero one. Provide the IAP my user name type my password. Okay. On the Mac address for the common interface with the boot interface and then the i p m I i p address These machines will be at the time storage worker manager. He's a manager. Yeah, we're gonna add a number of other machines on will. Speed this up just so you could see what the process looks like in the future. Better discovery will be added to the product. Okay. Okay. Getting back there we have it are Six machines have been added, are busy being inspected, being added to the system. Let's have a look at the details of a single note. Yeah, you can see information on the set up of the node. Its capabilities? Yeah. As well as the inventory information about that particular machine. I see. Okay, let's go and create the cluster. Yeah, So we're going to deploy a bare metal child cluster. The process we're going to go through is pretty much the same as any other child cluster. So we'll credit custom. We'll give it a name, but if it were selecting bare metal on the region, we're going to select the version we want to apply. No way. We're going to add this search keys. If we hope we're going to give the load. Balancer host I p that we'd like to use out of dress range on update the address range that we want to use for the cluster. Check that the sea ideal blocks for the Cuban ladies and tunnels are what we want them to be. Enable disabled stack light. Yeah, and soothe stack light settings to find the cluster. And then, as for any other machine, we need to add machines to the cluster. Here. We're focused on building communities clusters, so we're gonna put the count of machines. You want managers? We're gonna pick the label type manager and create three machines is the manager for the Cuban eighties. Casting Okay thing. We're having workers to the same. It's a process. Just making sure that the worker label host level are I'm sorry. On when Wait for the machines to deploy. Let's go through the process of putting the operating system on the notes validating and operating system deploying doctor identifies Make sure that the cluster is up and running and ready to go. Okay, let's review the bold events waken See the machine info now populated with more information about the specifics of things like storage and of course, details of a cluster etcetera. Yeah, yeah, well, now watch the machines go through the various stages from prepared to deploy on what's the cluster build? And that brings us to the end of this particular demo. You can see the process is identical to that of building a normal child cluster we got our complaint is complete. >>All right, so there we have it, deploying a cluster to bare metal. Much the same is how we did for AWS. I guess maybe the biggest different stepwise there is there is that registration face first, right? So rather than just using AWS financials toe magically create PM's in the cloud. You got a point out all your bare metal servers to Dr Enterprise between the cloud and they really come in, I guess three profiles, right? You got your manager profile with a profile storage profile which has been labeled as allocate. Um, crossword cluster has appropriate, >>right? And And I think that the you know, the key differentiator here is that you have more physical control over what, uh, attributes that love your cat, by the way, uh, where you have the different attributes of a server of physical server. So you can, uh, ensure that the SSD configuration on the storage nodes is gonna be taken advantage of in the best way the GP use on the worker nodes and and that the management layer is going to have sufficient horsepower to, um, spin up to to scale up the the environments, as required. One of the things I wanted to mention, though, um, if I could get this out without the choking much better. Um, is that Ah, hey, mentioned the load balancer and I wanted to make sure in defining the load balancer and the load balancer ranges. Um, that is for the top of the the cluster itself. That's the operations of the management, uh, layer integrating with your systems internally to be able to access the the Cube Can figs. I I p address the, uh, in a centralized way. It's not the load balancer that's working within the kubernetes cluster that you are deploying. That's still cube proxy or service mesh, or however you're intending to do it. So, um, it's kind of an interesting step that your initial step in building this, um and we typically use things like metal L B or in gen X or that kind of thing is to establish that before we deploy this bear mental cluster so that it can ride on top of that for the tips and things. >>Very cool. So any other thoughts on what we've seen so far today? Bruce, we've gone through all the different layers. Doctor enterprise container clouds in these videos from our management are regional to our clusters on aws hand bear amount, Of course, with his dad is still available. Closing thoughts before we take just a very short break and run through these demos again. >>You know, I've been very exciting. Ah, doing the presentation with you. I'm really looking forward to doing it the second time, so that we because we've got a good rhythm going about this kind of thing. So I'm looking forward to doing that. But I think that the key elements of what we're trying to convey to the folks out there in the audience that I hope you've gotten out of it is that will that this is an easy enough process that if you follow the step by steps going through the documentation that's been put out in the chat, um, that you'll be able to give this a go yourself, Um, and you don't have to limit yourself toe having physical hardware on prim to try it. You could do it in a ws as we've shown you today. And if you've got some fancy use cases like, uh, you you need a Hadoop And and, uh, you know, cloud oriented ai stuff that providing a bare metal service helps you to get there very fast. So right. Thank you. It's been a pleasure. >>Yeah, thanks everyone for coming out. So, like I said we're going to take a very short, like, three minute break here. Uh, take the opportunity to let your colleagues know if they were in another session or they didn't quite make it to the beginning of this session. Or if you just want to see these demos again, we're going to kick off this demo. Siri's again in just three minutes at ten. Twenty five a. M. Pacific time where we will see all this great stuff again. Let's take a three minute break. I'll see you all back here in just two minutes now, you know. Okay, folks, that's the end of our extremely short break. We'll give people just maybe, like one more minute to trickle in if folks are interested in coming on in and jumping into our demo. Siri's again. Eso For those of you that are just joining us now I'm Bill Mills. I head up curriculum development for the training team here. Moran Tous on Joining me for this session of demos is Bruce. Don't you go ahead and introduce yourself doors, who is still on break? That's cool. We'll give Bruce a minute or two to get back while everyone else trickles back in. There he is. Hello, Bruce. >>How'd that go for you? Okay, >>Very well. So let's kick off our second session here. I e just interest will feel for you. Thio. Let it run over here. >>Alright. Hi. Bruce Matthews here. I'm the Western Regional Solutions architect for Marantz. Use A I'm the one with the gray hair and the glasses. Uh, the handsome one is Bill. So, uh, Bill, take it away. >>Excellent. So over the next hour or so, we've got a Siris of demos that's gonna walk you through your first steps with Dr Enterprise Container Cloud Doctor Enterprise Container Cloud is, of course, Miranda's brand new offering from bootstrapping kubernetes clusters in AWS bare metal open stack. And for the providers in the very near future. So we we've got, you know, just just over an hour left together on this session, uh, if you joined us at the top of the hour back at nine. A. M. Pacific, we went through these demos once already. Let's do them again for everyone else that was only able to jump in right now. Let's go. Our first video where we're gonna install Dr Enterprise container cloud for the very first time and use it to bootstrap management. Cluster Management Cluster, as I like to describe it, is our mother ship that's going to spin up all the other kubernetes clusters, Doctor Enterprise clusters that we're gonna run our workloads on. So I'm gonna do >>I'm so excited. I can hardly wait. >>Let's do it all right to share my video out here. Yeah, let's do it. >>Good day. The focus for this demo will be the initial bootstrap of the management cluster on the first regional clusters. To support AWS deployments, the management cluster provides the core functionality, including identity management, authentication, infantry release version. The regional cluster provides the specific architecture provided in this case AWS and the Elsom components on the UCP cluster Child cluster is the cluster or clusters being deployed and managed. The deployment is broken up into five phases. The first phase is preparing a bootstrap note on its dependencies on handling the download of the bridge struck tools. The second phase is obtaining America's license file. Third phase. Prepare the AWS credentials instead of the ideas environment, the fourth configuring the deployment, defining things like the machine types on the fifth phase, Run the bootstrap script and wait for the deployment to complete. Okay, so here we're sitting up the strap node. Just checking that it's clean and clear and ready to go there. No credentials already set up on that particular note. Now, we're just checking through aws to make sure that the account we want to use we have the correct credentials on the correct roles set up on validating that there are no instances currently set up in easy to instance, not completely necessary, but just helps keep things clean and tidy when I am perspective. Right. So next step, we're just gonna check that we can from the bootstrap note, reach more antis, get to the repositories where the various components of the system are available. They're good. No areas here. Yeah, right now we're going to start sitting at the bootstrap note itself. So we're downloading the cars release, get get cars, script, and then next we're going to run it. Yeah, I've been deployed changing into that big struck folder, just making see what's there right now we have no license file, so we're gonna get the license filed. Okay? Get the license file through more antis downloads site signing up here, downloading that license file and putting it into the Carisbrook struck folder. Okay, since we've done that, we can now go ahead with the rest of the deployment. Yeah, see what the follow is there? Uh huh. Once again, checking that we can now reach E C two, which is extremely important for the deployment. Just validation steps as we move through the process. Alright. Next big step is violating all of our AWS credentials. So the first thing is, we need those route credentials which we're going to export on the command line. This is to create the necessary bootstrap user on AWS credentials for the completion off the deployment we're now running in AWS policy create. So it is part of that is creating our food trucks script. Creating this through policy files onto the AWS, just generally preparing the environment using a cloud formation script, you'll see in a second, I'll give a new policy confirmations just waiting for it to complete. And there is done. It's gonna have a look at the AWS console. You can see that we're creative completed. Now we can go and get the credentials that we created. Good day. I am console. Go to the new user that's being created. We'll go to the section on security credentials and creating new keys. Download that information media access Key I. D and the secret access key, but usually then exported on the command line. Okay, Couple of things to Notre. Ensure that you're using the correct AWS region on ensure that in the conflict file you put the correct Am I in for that region? I'm sure you have it together in a second. Okay, thanks. Is key. So you could X key Right on. Let's kick it off. So this process takes between thirty and forty five minutes. Handles all the AWS dependencies for you. Um, as we go through, the process will show you how you can track it. Andi will start to see things like the running instances being created on the AWS side. The first phase off this whole process happening in the background is the creation of a local kind based bootstrapped cluster on the bootstrap node that clusters then used to deploy and manage all the various instances and configurations within AWS at the end of the process. That cluster is copied into the new cluster on AWS and then shut down that local cluster essentially moving itself over. Yeah, okay. Local clusters boat. Just waiting for the various objects to get ready. Standard communities objects here. Yeah, you mentioned Yeah. So we've speed up this process a little bit just for demonstration purposes. Okay, there we go. So first note is being built the bastion host just jump box that will allow us access to the entire environment. Yeah, In a few seconds, we'll see those instances here in the US console on the right. Um, the failures that you're seeing around failed to get the I. P for Bastian is just the weight state while we wait for AWS to create the instance. Okay. Yeah. Beauty there. Movies. Okay, sketch. Hello? Yeah, Okay. Okay. On. There we go. Question host has been built on three instances for the management clusters have now been created. Okay, We're going through the process of preparing. Those nodes were now copying everything over. See that scaling up of controllers in the big strapped cluster? It's indicating that we're starting all of the controllers in the new question. Almost there. Right? Okay. Just waiting for key. Clark. Uh huh. So finish up. Yeah. No. Now we're shutting down. Control this on the local bootstrap node on preparing our I. D. C configuration, fourth indication. So once this is completed, the last phase will be to deploy stack light into the new cluster, that glass on monitoring tool set, Then we go stack like deployment has started. Mhm. Coming to the end of the deployment mountain. Yeah, they were cut final phase of the deployment. And we are done. Yeah, you'll see. At the end, they're providing us the details of you. I log in. So there's a key Clark log in. Uh, you can modify that initial default possible is part of the configuration set up where they were in the documentation way. Go Councils up way can log in. Yeah. Yeah. Thank you very much for watching. >>All right, so at this point, what we have we got our management cluster spun up, ready to start creating work clusters. So just a couple of points to clarify there to make sure everyone caught that, uh, as advertised. That's darker. Enterprise container cloud management cluster. That's not rework loans. are gonna go right? That is the tool and you're gonna use to start spinning up downstream commodity documentary prize clusters for bootstrapping record too. >>And the seed host that were, uh, talking about the kind cluster dingy actually doesn't have to exist after the bootstrap succeeds eso It's sort of like, uh, copies head from the seed host Toothy targets in AWS spins it up it then boots the the actual clusters and then it goes away too, because it's no longer necessary >>so that bootstrapping know that there's not really any requirements, Hardly on that, right. It just has to be able to reach aws hit that Hit that a p I to spin up those easy to instances because, as you just said, it's just a kubernetes in docker cluster on that piece. Drop note is just gonna get torn down after the set up finishes on. You no longer need that. Everything you're gonna do, you're gonna drive from the single pane of glass provided to you by your management cluster Doctor enterprise Continue cloud. Another thing that I think is sort of interesting their eyes that the convict is fairly minimal. Really? You just need to provide it like aws regions. Um, am I? And that's what is going to spin up that spending that matter faster. >>Right? There is a mammal file in the bootstrap directory itself, and all of the necessary parameters that you would fill in have default set. But you have the option then of going in and defining a different Am I different for a different region, for example? Oh, are different. Size of instance from AWS. >>One thing that people often ask about is the cluster footprint. And so that example you saw they were spitting up a three manager, um, managing cluster as mandatory, right? No single manager set up at all. We want high availability for doctrine Enterprise Container Cloud management. Like so again, just to make sure everyone sort of on board with the life cycle stage that we're at right now. That's the very first thing you're going to do to set up Dr Enterprise Container Cloud. You're going to do it. Hopefully exactly once. Right now, you've got your management cluster running, and they're gonna use that to spend up all your other work clusters Day today has has needed How do we just have a quick look at the questions and then lets take a look at spinning up some of those child clusters. >>Okay, e think they've actually been answered? >>Yeah, for the most part. One thing I'll point out that came up again in the Dail, helpfully pointed out earlier in surgery, pointed out again, is that if you want to try any of the stuff yourself, it's all of the dogs. And so have a look at the chat. There's a links to instructions, so step by step instructions to do each and every thing we're doing here today yourself. I really encourage you to do that. Taking this out for a drive on your own really helps internalizing communicate these ideas after the after launch pad today, Please give this stuff try on your machines. Okay, So at this point, like I said, we've got our management cluster. We're not gonna run workloads there that we're going to start creating child clusters. That's where all of our work and we're gonna go. That's what we're gonna learn how to do in our next video. Cue that up for us. >>I so love Shawn's voice. >>Wasn't that all day? >>Yeah, I watched him read the phone book. >>All right, here we go. Let's now that we have our management cluster set up, let's create a first child work cluster. >>Hello. In this demo, we will cover the deployment experience of creating a new child cluster the scaling of the cluster on how to update the cluster. When a new version is available, we begin the process by logging onto the you I as a normal user called Mary. Let's go through the navigation of the u I. So you can switch Project Mary only has access to development. Uh huh. Get a list of the available projects that you have access to. What clusters have been deployed at the moment there. Man. Yes, this H keys, Associate ID for Mary into her team on the cloud credentials that allow you to create or access the various clouds that you can deploy clusters to finally different releases that are available to us. We can switch from dark mode to light mode, depending on your preferences. Right. Let's now set up some ssh keys for Mary so she can access the notes and machines again. Very simply, had Mississippi key give it a name. We copy and paste our public key into the upload key block. Or we can upload the key if we have the file available on our machine. A very simple process. So to create a new cluster, we define the cluster ad management nodes and add worker nodes to the cluster. Yeah, again, very simply, we got the clusters tab we had to create cluster button. Give the cluster name. Yeah, Andi, select the provider. We only have access to AWS in this particular deployment, so we'll stick to AWS. What's like the region in this case? US West one released version five point seven is the current release Onda Attach. Mary's Key is necessary key. We can then check the rest of the settings, confirming the provider any kubernetes c r D a r i p address information. We can change this. Should we wish to? We'll leave it default for now and then what components of stack light? I would like to deploy into my custom for this. I'm enabling stack light on logging, and I consider the retention sizes attention times on. Even at this stage, add any custom alerts for the watchdogs. Consider email alerting which I will need my smart host. Details and authentication details. Andi Slack Alerts. Now I'm defining the cluster. All that's happened is the cluster's been defined. I now need to add machines to that cluster. I'll begin by clicking the create machine button within the cluster definition. Oh, select manager, Select the number of machines. Three is the minimum. Select the instant size that I'd like to use from AWS and very importantly, ensure correct. Use the correct Am I for the region. I convinced side on the route. Device size. There we go. My three machines are busy creating. I now need to add some workers to this cluster. So I go through the same process this time once again, just selecting worker. I'll just add to once again the am I is extremely important. Will fail if we don't pick the right. Am I for a Clinton machine? In this case and the deployment has started, we can go and check on the bold status are going back to the clusters screen on clicking on the little three dots on the right. We get the cluster info and the events, so the basic cluster info you'll see pending their listen. Cluster is still in the process of being built. We kick on, the events will get a list of actions that have been completed This part of the set up of the cluster. So you can see here. We've created the VPC. We've created the sub nets on. We've created the Internet Gateway. It's unnecessary made of us. And we have no warnings of the stage. Okay, this will then run for a while. We have one minute past. We can click through. We can check the status of the machine balls as individuals so we can check the machine info, details of the machines that we've assigned mhm and see any events pertaining to the machine areas like this one on normal. Yeah. Just last. The community's components are waiting for the machines to start. Go back to customers. Okay, right. Because we're moving ahead now. We can see we have it in progress. Five minutes in new Matt Gateway. And at this stage, the machines have been built on assigned. I pick up the U S. Yeah, yeah, yeah. There we go. Machine has been created. See the event detail and the AWS. I'd for that machine. No speeding things up a little bit this whole process and to end takes about fifteen minutes. Run the clock forward, you'll notice is the machines continue to bold the in progress. We'll go from in progress to ready. A soon as we got ready on all three machines, the managers on both workers way could go on and we could see that now we reached the point where the cluster itself is being configured mhm and then we go. Cluster has been deployed. So once the classes deployed, we can now never get around. Our environment are looking into configure cluster. We could modify their cluster. We could get the end points for alert Alert Manager See here the griffon occupying and Prometheus are still building in the background but the cluster is available on You would be able to put workloads on it at this stage to download the cube conflict so that I can put workloads on it. It's again the three little dots in the right for that particular cluster. If the download cube conflict give it my password, I now have the Q conflict file necessary so that I can access that cluster. All right, Now that the build is fully completed, we can check out cluster info on. We can see that all the satellite components have been built. All the storage is there, and we have access to the CPU. I. So if we click into the cluster, we can access the UCP dashboard, click the signing with the clock button to use the SSO. We give Mary's possible to use the name once again. Thing is an unlicensed cluster way could license at this point. Or just skip it on. Do we have the UCP dashboard? You could see that has been up for a little while. We have some data on the dashboard going back to the console. We can now go to the griffon. A data just been automatically pre configured for us. We can switch and utilized a number of different dashboards that have already been instrumented within the cluster. So, for example, communities cluster information, the name spaces, deployments, nodes. Um, so we look at nodes. If we could get a view of the resource is utilization of Mrs Custer is very little running in it. Yeah, a general dashboard of Cuba Navies cluster. What If this is configurable, you can modify these for your own needs, or add your own dashboards on de scoped to the cluster. So it is available to all users who have access to this specific cluster. All right to scale the cluster on to add a No. This is simple. Is the process of adding a mode to the cluster, assuming we've done that in the first place. So we go to the cluster, go into the details for the cluster we select, create machine. Once again, we need to be ensure that we put the correct am I in and any other functions we like. You can create different sized machines so it could be a larger node. Could be bigger group disks and you'll see that worker has been added in the provisioning state. On shortly, we will see the detail off that worker as a complete to remove a note from a cluster. Once again, we're going to the cluster. We select the node we would like to remove. Okay, I just hit delete On that note. Worker nodes will be removed from the cluster using according and drawing method to ensure that your workloads are not affected. Updating a cluster. When an update is available in the menu for that particular cluster, the update button will become available. And it's a simple as clicking the button validating which release you would like to update to this case. This available releases five point seven point one give you I'm kicking the update back in the background. We will coordinate. Drain each node slowly, go through the process of updating it. Andi update will complete depending on what the update is as quickly as possible. Who we go. The notes being rebuilt in this case impacted the manager node. So one of the manager nodes is in the process of being rebuilt. In fact, to in this case, one has completed already. Yeah, and in a few minutes, we'll see that the upgrade has been completed. There we go. Great. Done. If you work loads of both using proper cloud native community standards, there will be no impact. >>All right, there. We haven't. We got our first workload cluster spun up and managed by Dr Enterprise Container Cloud. So I I loved Shawn's classic warning there. When you're spinning up an actual doctor enterprise deployment, you see little errors and warnings popping up. Just don't touch it. Just leave it alone and let Dr Enterprises self healing properties take care of all those very transient temporary glitches, resolve themselves and leave you with a functioning workload cluster within victims. >>And now, if you think about it that that video was not very long at all. And that's how long it would take you if someone came into you and said, Hey, can you spend up a kubernetes cluster for development development A. Over here, um, it literally would take you a few minutes to thio Accomplish that. And that was with a W s. Obviously, which is sort of, ah, transient resource in the cloud. But you could do exactly the same thing with resource is on Prem or resource is, um physical resource is and will be going through that later in the process. >>Yeah, absolutely one thing that is present in that demo, but that I like to highlight a little bit more because it just kind of glides by Is this notion of, ah, cluster release? So when Sean was creating that cluster, and also when when he was upgrading that cluster, he had to choose a release. What does that didn't really explain? What does that mean? Well, in Dr Enterprise Container Cloud, we have released numbers that capture the entire staff of container ization tools that will be deploying to that workload costume. So that's your version of kubernetes sed cor DNs calico. Doctor Engineer. All the different bits and pieces that not only work independently but are validated toe work together as a staff appropriate for production, humanities, adopted enterprise environments. >>Yep. From the bottom of the stack to the top, we actually test it for scale. Test it for CVS, test it for all of the various things that would, you know, result in issues with you running the application services. And I've got to tell you from having, you know, managed kubernetes deployments and things like that that if you're the one doing it yourself, it can get rather messy. Eso This makes it easy. >>Bruce, you were staying a second ago. They I'll take you at least fifteen minutes to install your release. Custer. Well, sure, but what would all the other bits and pieces you need toe? Not just It's not just about pressing the button to install it, right? It's making the right decision. About what components work? Well, our best tested toe be successful working together has a staff? Absolutely. We this release mechanism and Dr Enterprise Container Cloud. Let's just kind of package up that expert knowledge and make it available in a really straightforward, fashionable species. Uh, pre Confederate release numbers and Bruce is you're pointing out earlier. He's got delivered to us is updates kind of transparent period. When when? When Sean wanted toe update that cluster, he created little update. Custer Button appeared when an update was available. All you gotta do is click. It tells you what Here's your new stack of communities components. It goes ahead. And the straps those components for you? >>Yeah, it actually even displays at the top of the screen. Ah, little header That says you've got an update available. Do you want me to apply? It s o >>Absolutely. Another couple of cool things. I think that are easy to miss in that demo was I really like the on board Bafana that comes along with this stack. So we've been Prometheus Metrics and Dr Enterprise for years and years now. They're very high level. Maybe in in previous versions of Dr Enterprise having those detailed dashboards that Ravana provides, I think that's a great value out there. People always wanted to be ableto zoom in a little bit on that, uh, on those cluster metrics, you're gonna provides them out of the box for us. Yeah, >>that was Ah, really, uh, you know, the joining of the Miranda's and Dr teams together actually spawned us to be able to take the best of what Morantes had in the open stack environment for monitoring and logging and alerting and to do that integration in in a very short period of time so that now we've got it straight across the board for both the kubernetes world and the open stack world. Using the same tool sets >>warm. One other thing I wanna point out about that demo that I think there was some questions about our last go around was that demo was all about creating a managed workplace cluster. So the doctor enterprise Container Cloud managers were using those aws credentials provisioned it toe actually create new e c two instances installed Docker engine stalled. Doctor Enterprise. Remember all that stuff on top of those fresh new VM created and managed by Dr Enterprise contain the cloud. Nothing unique about that. AWS deployments do that on open staff doing on Parramatta stuff as well. Um, there's another flavor here, though in a way to do this for all of our long time doctor Enterprise customers that have been running Doctor Enterprise for years and years. Now, if you got existing UCP points existing doctor enterprise deployments, you plug those in to Dr Enterprise Container Cloud, uh, and use darker enterprise between the cloud to manage those pre existing Oh, working clusters. You don't always have to be strapping straight from Dr Enterprises. Plug in external clusters is bad. >>Yep, the the Cube config elements of the UCP environment. The bundling capability actually gives us a very straightforward methodology. And there's instructions on our website for exactly how thio, uh, bring in import and you see p cluster. Um so it it makes very convenient for our existing customers to take advantage of this new release. >>Absolutely cool. More thoughts on this wonders if we jump onto the next video. >>I think we should move press on >>time marches on here. So let's Let's carry on. So just to recap where we are right now, first video, we create a management cluster. That's what we're gonna use to create All our downstream were closed clusters, which is what we did in this video. Let's maybe the simplest architectures, because that's doing everything in one region on AWS pretty common use case because we want to be able to spin up workload clusters across many regions. And so to do that, we're gonna add a third layer in between the management and work cluster layers. That's gonna be our regional cluster managers. So this is gonna be, uh, our regional management cluster that exists per region that we're going to manage those regional managers will be than the ones responsible for spending part clusters across all these different regions. Let's see it in action in our next video. >>Hello. In this demo, we will cover the deployment of additional regional management. Cluster will include a brief architectural overview, how to set up the management environment, prepare for the deployment deployment overview, and then just to prove it, to play a regional child cluster. So looking at the overall architecture, the management cluster provides all the core functionality, including identity management, authentication, inventory and release version. ING Regional Cluster provides the specific architecture provider in this case, AWS on the L C M components on the d you speak cluster for child cluster is the cluster or clusters being deployed and managed? Okay, so why do you need original cluster? Different platform architectures, for example AWS open stack, even bare metal to simplify connectivity across multiple regions handle complexities like VPNs or one way connectivity through firewalls, but also help clarify availability zones. Yeah. Here we have a view of the regional cluster and how it connects to the management cluster on their components, including items like the LCN cluster Manager. We also machine manager. We're hell Mandel are managed as well as the actual provider logic. Okay, we'll begin by logging on Is the default administrative user writer. Okay, once we're in there, we'll have a look at the available clusters making sure we switch to the default project which contains the administration clusters. Here we can see the cars management cluster, which is the master controller. When you see it only has three nodes, three managers, no workers. Okay, if we look at another regional cluster, similar to what we're going to deploy now. Also only has three managers once again, no workers. But as a comparison is a child cluster. This one has three managers, but also has additional workers associate it to the cluster. Yeah, all right, we need to connect. Tell bootstrap note, preferably the same note that used to create the original management plaster. It's just on AWS, but I still want to machine Mhm. All right, A few things we have to do to make sure the environment is ready. First thing we're gonna pseudo into route. I mean, we'll go into our releases folder where we have the car's boot strap on. This was the original bootstrap used to build the original management cluster. We're going to double check to make sure our cube con figures there It's again. The one created after the original customers created just double check. That cute conflict is the correct one. Does point to the management cluster. We're just checking to make sure that we can reach the images that everything's working, condone, load our images waken access to a swell. Yeah, Next, we're gonna edit the machine definitions what we're doing here is ensuring that for this cluster we have the right machine definitions, including items like the am I So that's found under the templates AWS directory. We don't need to edit anything else here, but we could change items like the size of the machines attempts we want to use but the key items to ensure where changed the am I reference for the junta image is the one for the region in this case aws region of re utilizing. This was an open stack deployment. We have to make sure we're pointing in the correct open stack images. Yeah, yeah. Okay. Sit the correct Am I save the file? Yeah. We need to get up credentials again. When we originally created the bootstrap cluster, we got credentials made of the U. S. If we hadn't done this, we would need to go through the u A. W s set up. So we just exporting AWS access key and I d. What's important is Kaz aws enabled equals. True. Now we're sitting the region for the new regional cluster. In this case, it's Frankfurt on exporting our Q conflict that we want to use for the management cluster when we looked at earlier. Yeah, now we're exporting that. Want to call? The cluster region is Frankfurt's Socrates Frankfurt yet trying to use something descriptive? It's easy to identify. Yeah, and then after this, we'll just run the bootstrap script, which will complete the deployment for us. Bootstrap of the regional cluster is quite a bit quicker than the initial management clusters. There are fewer components to be deployed, but to make it watchable, we've spent it up. So we're preparing our bootstrap cluster on the local bootstrap node. Almost ready on. We started preparing the instances at us and waiting for the past, you know, to get started. Please the best your node, onda. We're also starting to build the actual management machines they're now provisioning on. We've reached the point where they're actually starting to deploy Dr Enterprise, he says. Probably the longest face we'll see in a second that all the nodes will go from the player deployed. Prepare, prepare Mhm. We'll see. Their status changes updates. It was the first word ready. Second, just applying second. Grady, both my time away from home control that's become ready. Removing cluster the management cluster from the bootstrap instance into the new cluster running a data for us? Yeah, almost a on. Now we're playing Stockland. Thanks. Whichever is done on Done. Now we'll build a child cluster in the new region very, very quickly. Find the cluster will pick our new credential have shown up. We'll just call it Frankfurt for simplicity. A key on customers to find. That's the machine. That cluster stop with three manages set the correct Am I for the region? Yeah, Same to add workers. There we go. That's the building. Yeah. Total bill of time. Should be about fifteen minutes. Concedes in progress. Can we expect this up a little bit? Check the events. We've created all the dependencies, machine instances, machines. A boat? Yeah. Shortly. We should have a working caster in the Frankfurt region. Now almost a one note is ready from management. Two in progress. On we're done. Trust us up and running. >>Excellent. There we have it. We've got our three layered doctor enterprise container cloud structure in place now with our management cluster in which we scrap everything else. Our regional clusters which manage individual aws regions and child clusters sitting over depends. >>Yeah, you can. You know you can actually see in the hierarchy the advantages that that presents for folks who have multiple locations where they'd like a geographic locations where they'd like to distribute their clusters so that you can access them or readily co resident with your development teams. Um and, uh, one of the other things I think that's really unique about it is that we provide that same operational support system capability throughout. So you've got stack light monitoring the stack light that's monitoring the stack light down to the actual child clusters that they have >>all through that single pane of glass that shows you all your different clusters, whether their workload cluster like what the child clusters or usual clusters from managing different regions. Cool. Alright, well, time marches on your folks. We've only got a few minutes left and I got one more video in our last video for the session. We're gonna walk through standing up a child cluster on bare metal. So so far, everything we've seen so far has been aws focus. Just because it's kind of easy to make that was on AWS. We don't want to leave you with the impression that that's all we do, we're covering AWS bare metal and open step deployments as well documented Craftsman Cloud. Let's see it in action with a bare metal child cluster. >>We are on the home stretch, >>right. >>Hello. This demo will cover the process of defining bare metal hosts and then review the steps of defining and deploying a bare metal based doctor enterprise cluster. Yeah, so why bare metal? Firstly, it eliminates hyper visor overhead with performance boost of up to thirty percent provides direct access to GP use, prioritize for high performance wear clothes like machine learning and AI, and support high performance workouts like network functions, virtualization. It also provides a focus on on Prem workloads, simplifying and ensuring we don't need to create the complexity of adding another hyper visor layer in between. So continuing on the theme Why communities and bare metal again Hyper visor overhead. Well, no virtualization overhead. Direct access to hardware items like F p g A s G p, us. We can be much more specific about resource is required on the nodes. No need to cater for additional overhead. We can handle utilization in the scheduling better Onda. We increase the performance and simplicity of the entire environment as we don't need another virtualization layer. Yeah, In this section will define the BM hosts will create a new project. Will add the bare metal hosts, including the host name. I put my credentials. I pay my address, Mac address on, then provide a machine type label to determine what type of machine it is. Related use. Okay, let's get started Certain Blufgan was the operator thing. We'll go and we'll create a project for our machines to be a member off. Helps with scoping for later on for security. I begin the process of adding machines to that project. Yeah. Yeah. So the first thing we had to be in post many of the machine a name. Anything you want? Yeah, in this case by mental zero one. Provide the IAP My user name. Type my password? Yeah. On the Mac address for the active, my interface with boot interface and then the i p m i P address. Yeah, these machines. We have the time storage worker manager. He's a manager. We're gonna add a number of other machines on will speed this up just so you could see what the process. Looks like in the future, better discovery will be added to the product. Okay, Okay. Getting back there. We haven't Are Six machines have been added. Are busy being inspected, being added to the system. Let's have a look at the details of a single note. Mhm. We can see information on the set up of the node. Its capabilities? Yeah. As well as the inventory information about that particular machine. Okay, it's going to create the cluster. Mhm. Okay, so we're going to deploy a bare metal child cluster. The process we're going to go through is pretty much the same as any other child cluster. So credit custom. We'll give it a name. Thank you. But he thought were selecting bare metal on the region. We're going to select the version we want to apply on. We're going to add this search keys. If we hope we're going to give the load. Balancer host I p that we'd like to use out of the dress range update the address range that we want to use for the cluster. Check that the sea idea blocks for the communities and tunnels are what we want them to be. Enable disabled stack light and said the stack light settings to find the cluster. And then, as for any other machine, we need to add machines to the cluster. Here we're focused on building communities clusters. So we're gonna put the count of machines. You want managers? We're gonna pick the label type manager on create three machines. Is a manager for the Cuban a disgusting? Yeah, they were having workers to the same. It's a process. Just making sure that the worker label host like you are so yes, on Duin wait for the machines to deploy. Let's go through the process of putting the operating system on the notes, validating that operating system. Deploying Docker enterprise on making sure that the cluster is up and running ready to go. Okay, let's review the bold events. We can see the machine info now populated with more information about the specifics of things like storage. Yeah, of course. Details of a cluster, etcetera. Yeah, Yeah. Okay. Well, now watch the machines go through the various stages from prepared to deploy on what's the cluster build, and that brings us to the end of this particular do my as you can see the process is identical to that of building a normal child cluster we got our complaint is complete. >>Here we have a child cluster on bare metal for folks that wanted to play the stuff on Prem. >>It's ah been an interesting journey taken from the mothership as we started out building ah management cluster and then populating it with a child cluster and then finally creating a regional cluster to spread the geographically the management of our clusters and finally to provide a platform for supporting, you know, ai needs and and big Data needs, uh, you know, thank goodness we're now able to put things like Hadoop on, uh, bare metal thio in containers were pretty exciting. >>Yeah, absolutely. So with this Doctor Enterprise container cloud platform. Hopefully this commoditized scooping clusters, doctor enterprise clusters that could be spun up and use quickly taking provisioning times. You know, from however many months to get new clusters spun up for our teams. Two minutes, right. We saw those clusters gets better. Just a couple of minutes. Excellent. All right, well, thank you, everyone, for joining us for our demo session for Dr Enterprise Container Cloud. Of course, there's many many more things to discuss about this and all of Miranda's products. If you'd like to learn more, if you'd like to get your hands dirty with all of this content, police see us a training don Miranda's dot com, where we can offer you workshops and a number of different formats on our entire line of products and hands on interactive fashion. Thanks, everyone. Enjoy the rest of the launchpad of that >>thank you all enjoy.
SUMMARY :
So for the next couple of hours, I'm the Western regional Solutions architect for Moran At least somebody on the call knows something about your enterprise Computer club. And that's really the key to this thing is to provide some, you know, many training clusters so that by the end of the tutorial content today, I think that's that's pretty much what we had to nail down here. So the management costs was always We have to give this brief little pause of the management cluster in the first regional clusters to support AWS deployments. So in that video are wonderful field CTO Shauna Vera bootstrapped So primarily the foundation for being able to deploy So this cluster isn't yet for workloads. Read the phone book, So and just to make sure I understood The output that when it says I'm pivoting, I'm pivoting from on the bootstrap er go away afterwards. So that there's no dependencies on any of the clouds that get created thereafter. Yeah, that actually reminds me of how we bootstrapped doctor enterprise back in the day, The config file that that's generated the template is fairly straightforward We always insist on high availability for this management cluster the scenes without you having toe worry about it as a developer. Examples of that is the day goes on. either the the regional cluster or a We've got the management cluster, and we're gonna go straight with child cluster. as opposed to having to centralize thumb So just head on in, head on into the docks like the Dale provided here. That's going to be in a very near term I didn't wanna make promises for product, but I'm not too surprised that she's gonna be targeted. No, just that the fact that we're running through these individual So let's go to that video and see just how We can check the status of the machine bulls as individuals so we can check the machine the thing that jumped out to me at first Waas like the inputs that go into defining Yeah, and and And that's really the focus of our effort is to ensure that So at that point, once we started creating that workload child cluster, of course, we bootstrapped good old of the bootstrapping as well that the processes themselves are self healing, And the worst thing you could do is panic at the first warning and start tearing things that don't that then go out to touch slack and say hi, You need to watch your disk But Sean mentioned it on the video. And And the kubernetes, uh, scaling methodology is is he adhered So should we go to the questions. Um, that's kind of the point, right? you know, set up things and deploy your applications and things. that comes to us not from Dr Enterprise Container Cloud, but just from the underlying kubernetes distribution. to the standards that we would want to set to make sure that we're not overloading On the next video, we're gonna learn how to spin up a Yeah, Do the same to add workers. We got that management cluster that we do strapped in the first video. Yeah, that's the key to this is to be able to have co resident with So we don't have to go back to the mother ship. So it's just one pane of glass to the bootstrapped cluster of the regional services. and another, you know, detail for those that have sharp eyes. Let's take a quick peek of the questions here, see if there's anything we want to call out, then we move on to our last want all of the other major players in the cloud arena. Let's jump into our last video in the Siri's, So the first thing we had to be in post, Yeah, many of the machine A name. Much the same is how we did for AWS. nodes and and that the management layer is going to have sufficient horsepower to, are regional to our clusters on aws hand bear amount, Of course, with his dad is still available. that's been put out in the chat, um, that you'll be able to give this a go yourself, Uh, take the opportunity to let your colleagues know if they were in another session I e just interest will feel for you. Use A I'm the one with the gray hair and the glasses. And for the providers in the very near future. I can hardly wait. Let's do it all right to share my video So the first thing is, we need those route credentials which we're going to export on the command That is the tool and you're gonna use to start spinning up downstream It just has to be able to reach aws hit that Hit that a p I to spin up those easy to instances because, and all of the necessary parameters that you would fill in have That's the very first thing you're going to Yeah, for the most part. Let's now that we have our management cluster set up, let's create a first We can check the status of the machine balls as individuals so we can check the glitches, resolve themselves and leave you with a functioning workload cluster within exactly the same thing with resource is on Prem or resource is, All the different bits and pieces And I've got to tell you from having, you know, managed kubernetes And the straps those components for you? Yeah, it actually even displays at the top of the screen. I really like the on board Bafana that comes along with this stack. the best of what Morantes had in the open stack environment for monitoring and logging So the doctor enterprise Container Cloud managers were Yep, the the Cube config elements of the UCP environment. More thoughts on this wonders if we jump onto the next video. Let's maybe the simplest architectures, of the regional cluster and how it connects to the management cluster on their components, There we have it. that we provide that same operational support system capability Just because it's kind of easy to make that was on AWS. Just making sure that the worker label host like you are so yes, It's ah been an interesting journey taken from the mothership Enjoy the rest of the launchpad
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Mary | PERSON | 0.99+ |
Sean | PERSON | 0.99+ |
Sean O'Mara | PERSON | 0.99+ |
Bruce | PERSON | 0.99+ |
Frankfurt | LOCATION | 0.99+ |
three machines | QUANTITY | 0.99+ |
Bill Milks | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
first video | QUANTITY | 0.99+ |
second phase | QUANTITY | 0.99+ |
Shawn | PERSON | 0.99+ |
first phase | QUANTITY | 0.99+ |
Three | QUANTITY | 0.99+ |
Two minutes | QUANTITY | 0.99+ |
three managers | QUANTITY | 0.99+ |
fifth phase | QUANTITY | 0.99+ |
Clark | PERSON | 0.99+ |
Bill Mills | PERSON | 0.99+ |
Dale | PERSON | 0.99+ |
Five minutes | QUANTITY | 0.99+ |
Nan | PERSON | 0.99+ |
second session | QUANTITY | 0.99+ |
Third phase | QUANTITY | 0.99+ |
Seymour | PERSON | 0.99+ |
Bruce Basil Matthews | PERSON | 0.99+ |
Moran Tous | PERSON | 0.99+ |
five minutes | QUANTITY | 0.99+ |
hundreds | QUANTITY | 0.99+ |
Miska Kaipiainen, Mirantis | Mirantis Launchpad 2020
>> Announcer: From around the globe, it's theCUBE, with digital coverage at Mirantis Launchpad 2020, brought to you by Mirantis. >> Welcome back. And I'm Stu Miniman, and this is theCUBE's coverage of Mirantis Launchpad 2020. Of course we're spending a lot of time talking about Kubernetes. We're going to be digging in talking about some of the important developer tooling that Mirantis is helping to proliferate in the market, solve some real important challenges in the space. So happy to welcome to the program Miska Kaipiainen. He is the senior director of engineering with Mirantis. Miska, thanks so much for joining us. Welcome to theCUBE. >> Thank you so much. >> All right, so Miska, I notice you've got on the Kontena sweatshirt. You were the founder of the company, did some tools. One of the tools that you and your team helped create was Lens. You and your team joined Mirantis, and recently Lens was pulled in. So maybe if you could just give us a little bit about your background. You do some coding yourself, the team that you have there, and let's tee up the conversation, 'cause it's that Lens piece that we're going to spend a bunch of time talking about. >> Yeah, so the background of what we did, basically Kontena, we started back in 2015, and we a the focus on creating technologies around the container orchestration technologies to basically to make developer tooling that are very easy to use for the developers. So during the years at Kontena, we did many different types of products, and maybe the most interesting product that we created was Lens. And now really when we joined Mirantis in January this year, so we have been able to work on Lens, and actually, since the Lens was made open source, fully open source in March this year, so it's been really kind of picking up, and now Mirantis acquired the whole technology, so we can really start investing even more in the development. >> All right, so let's talk specifically about Lens. As I teed up at the beginning, we're talking about managing multiple clusters. Gosh, and I think back to 2015. It was early on. Most people were still learning about Docker, Docker swarms, Kubernetes, Mesos. There were a lot of fights over how orchestration would be done. A little bit different discussion about what developers were doing, how they scaled out configurations, how they manage those. So help us understand kind of that core, what Lens does, and how the product has matured and expanded over those last five years. >> Yeah, so over the last five years, so originally Lens was developed for our internal product. So like Mesosphere and Docker, and they all have their own orchestration technologies even before Kubernetes. And we also started working on our own orchestration technology. And I'm a huge believer in when we are dealing with very complex technologies, so if you can visualize it and make it kind of more interesting to look at, so it will kind of help with the adoption, and it's kind of more acceptable to the market. And that's why we started doing Lens. And over the years, we turned Lens to work with Kubernetes environments, and nowadays really Lens is very much loved by the Kubernetes developers, who are those people who need to deal with the Kubernetes clusters on a daily basis. So they are not necessarily those ops people who are creating those clusters , but they are the people who actually use those clusters. >> Well, of course that that general adoption is something that, you know, super important. You have some stats you can share on, you talk about the love of developers. You said it's open source, it's available on GitHub, but how many people are using it? What are some of those usage stats? >> Yeah, so it was interesting. So when we released Lens open source under MIT license in March, so since then we have been getting, in half a year, we have been getting 8,000 stargazers on GitHub. That is kind of mind-blowing because we try to create projects and trying to create anything that would get a lot of traction in the past, but truly, it totally happened just now after years of trying. So it has been since the last six months, it's been just amazing the adopts and we have more than 50,000 users using Lens and the retention is great. People keep on coming back. So yeah, the numbers look very, very good for Lens, and we are just getting started. >> Yeah, well, it's something that this community definitely is huge growth, and anybody in this space remembers just the huge adoption of Docker, which of course the enterprise piece of Docker is now part of Mirantis. Inside those developers, help us understand a little bit more, what is it that has them really not only looking at the GitHubs, starring it, as you said, they're the stargazers. It's like a favorite, for those that aren't in the system. I've had a chance to look at some of the demos, and it seems rather straightforward. But if you could, just in your words, explain what it is that it solves for developers that otherwise they either had to do themselves or they had to cobble together a lot of different tools. We know developers out there. The wonderful thing is there's no shortage of tools to choose from. It's about the right tool that can do the right thing. >> Absolutely, absolutely. So Lens, we are calling it IDE for a reason. So we are talking about IDE for Kubernetes developers. And what does it mean actually is that we are taking all those necessary tools and technologies and packaging them, integrating them seamlessly together for the purpose of making it more easy for developers to deploy, operate, observe, inspect their workloads that are running on Kubernetes clusters. And I think the main benefits that Lens will provide for these developers is that if you're a newcomer in the Kubernetes ecosystem, so Lens gives you a very easy way to learn Kubernetes because it's so visual. And for more experienced users, it just radically improves the, let's say the speed of business and the way how you can perform things with your clusters. >> So one of the pieces that that Lens does is that multi-cluster management. So first of all, I believe, as you said, it's open source and can work with, is it any certified Kubernetes out there, whether it be from the public cloud, companies like VMware and Red Hat that have Kubernetes, of course, Mirantis has Kubernetes, too. And secondly, I think you teased out a little bit, but help help us understand a little bit. Multi-cluster management is something that the big players, you hear Azure and Google Cloud talking about how they look at managing not only other environments, but oh yeah, we can have other clusters and we can help you manage it. I think that's more on the ops side of things, as opposed to, as you said, this is really a developer tool set. >> Yeah, so of course, all the organizations, they want to most likely have some sort of centralized system where they can manage multiple clusters, and some companies provide systems for on-premises, and some public cloud vendors, they provide systems for provisioning those clusters on their own own systems. And then we have also the kind of multicloud management systems. Most of these technologies, they are really designed for the operations side, so how the IT administrations can manage these multiple clusters. So now if you look at the situation from the developer's perspective, they are now given access to certain number of clusters from different environments. And by the way, some of these clusters are also running on their local development environments on their laptops. So what Lens is doing is basically provides a unified user experience across all these clusters no matter what is the flavor of the Kubernetes. It can be the Minikube. It can be from AKS. It can be Mirantis Enterprise, Docker Enterprise offering, or whatever. So it kind of brings them all together and makes it very easy to navigate and go around and do your work. >> Yeah, well, that's, the promise of Kubernetes isn't that it just levels the playing field amongst everything. As I've talked to the founders of Kubernetes, people like Joe Beda said it's not a silver bullet. It's a thin layer. But that skillset is what's so important because there is a lot of difference between every platform they deal with. So as a developer, it's nice to have some tools that I can work across those environments. From a developer standpoint, I think it's on Windows, Linux, Mac, works across those environment. What do you hear from your customers? How are they using it? Is this something that they're like, oh hey, I can go make an adjustment on my mobile when I'm not necessarily in the office? Are we not quite there yet? >> Actually, it's kind of funny, because sometimes we hear these type of requests that we would like to have a mobile app version of Lens. I don't know how that would actually work in practice. So we haven't been doing anything on that front yet. I think still the most common use case is that developers, they are given access to clusters from somewhere and they are just desperately trying to find a kind of convenient way how to navigate around these different clusters and how to manage their workloads. And I think Lens is hitting the sweet spot in there with the ease of use. >> All right, so let me understand. It's been open sourced, yet Mirantis owns it. Is there a service or support? Does this tie into other products in the Mirantis portfolio? How do people get it? What do they need to, if anything, pay for it? And help us understand how this fits into the broader Mirantis story. >> Yes, so it's still kind of early days, so we just kind of announced that Lens is now part of Mirantis, let's say portfolio. So I must say that still the kind of main focus for us is around improving Lens and making it better for developers. So that's much more important than trying to think about the ways how potentially we could monetize this. So, but there are plans going ahead, going around for different ways how we can better support bigger enterprises who want to start using Lens in a big scale. >> Well, yeah, that's so important. Of course, developers, we need to lower the friction, help them adopt things fast. Miska, just get your general viewpoint, though. One of the big value propositions that Mirantis has is of course allowing enterprises to take advantage of these new types of solutions, especially today around Kubernetes. So help us understand from your standpoint the philosophy of what your team's helping to build and the customer engagements that you're having. >> Yes, so Mirantis, of course, has a broad portfolio of products, and many of those products, of course, are related to Kubernetes. And so we have many products which I'm also one of the leading development efforts around those. So some of the products are related to how to manage image repositories and registries. Some of them are related to how to handle the helm charts, which has basically become the defacto packaging format for Kubernetes applications. And we are kind of trying to bring all these different products and technologies together in a way that make it even more easy for developers then to access through Lens. So it's still a little bit work in progress, of course, since the Lens ecosystem is quite new, but we are on track there trying to make a beautiful one kind of experience for our customers. >> All right, well, final question I have for you. As you said, it's new there, but it gives a little taste as to feedback you're getting from the community. Anything we should be looking at on kind of the near to mid-term road map when it comes to Lens. >> Oh yeah, so we are just barely scratching the surface of the potential on what we can do with Lens. So one of the big features that we will be releasing still during this year in a couple of months time is going to be the extension API, which will allow all these cloud-native technology ecosystem vendors to bring their own technologies easily available and accessible through Lens. So it is possible for third parties to extend the user interface with their own kind of unique features and visualizations. And we are already actively working with certain partners to integrate their technologies through this extension API. So that's going to be huge. It's going to be game-changer. >> Well, the great thing about an open source project is people can go out, they can grab it now, they can give feedback, participate in the community. Miska, thank you so much for joining us and great to chat. >> Thank you for having me. Thank you. >> All right, stay with us for more coverage of Mirantis Launchpad 2020. I'm Stu Miniman and thank you for watching theCUBE. (bright music)
SUMMARY :
the globe, it's theCUBE, some of the important developer tooling One of the tools that you and maybe the most interesting product and how the product has matured Yeah, so over the last five years, Well, of course that So it has been since the last six months, that can do the right thing. and the way how you can perform and we can help you manage it. flavor of the Kubernetes. the promise of Kubernetes and how to manage their workloads. in the Mirantis portfolio? So I must say that still the and the customer engagements So some of the products are related to on kind of the near to mid-term road map of the potential on what and great to chat. Thank you for having me. and thank you for watching theCUBE.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Miska | PERSON | 0.99+ |
Miska Kaipiainen | PERSON | 0.99+ |
2015 | DATE | 0.99+ |
March | DATE | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
Joe Beda | PERSON | 0.99+ |
Mirantis | ORGANIZATION | 0.99+ |
Kontena | ORGANIZATION | 0.99+ |
more than 50,000 users | QUANTITY | 0.99+ |
Lens | TITLE | 0.98+ |
Linux | TITLE | 0.98+ |
one | QUANTITY | 0.98+ |
January this year | DATE | 0.98+ |
Windows | TITLE | 0.98+ |
March this year | DATE | 0.98+ |
Kubernetes | TITLE | 0.98+ |
today | DATE | 0.97+ |
half a year | QUANTITY | 0.97+ |
GitHub | ORGANIZATION | 0.96+ |
One | QUANTITY | 0.96+ |
Red Hat | ORGANIZATION | 0.96+ |
AKS | ORGANIZATION | 0.95+ |
theCUBE | ORGANIZATION | 0.94+ |
secondly | QUANTITY | 0.93+ |
Docker | TITLE | 0.92+ |
Mirantis Launchpad 2020 | TITLE | 0.92+ |
Mesosphere | TITLE | 0.91+ |
VMware | ORGANIZATION | 0.91+ |
GitHubs | ORGANIZATION | 0.9+ |
Azure | TITLE | 0.9+ |
8,000 stargazers | QUANTITY | 0.86+ |
last six months | DATE | 0.82+ |
this year | DATE | 0.82+ |
last five years | DATE | 0.8+ |
Lens | ORGANIZATION | 0.77+ |
MIT | ORGANIZATION | 0.77+ |
Kubernetes | ORGANIZATION | 0.76+ |
Mac | TITLE | 0.74+ |
first | QUANTITY | 0.71+ |
Google Cloud | TITLE | 0.71+ |
Shaun O'Meara, Mirantis | Mirantis Launchpad 2020
>> Narrator: From around the globe, its theCUBE with digital coverage of Mirantis Launchpad 2020 brought to you by Mirantis . >> Welcome back, I'm Stu Miniman and this is theCUBE coverage of Mirantis Launchpad 2020, really looking at how Mirantis Docker Enterprise are coming together, changes happening in the field and to help us dig into that customer and product discussion. Happy to welcome to the program, Shaun O'Meara. He is the global Field Chief Technology Officer with Mirantis coming to us from Germany. Shaun thanks so much for joining us. >> Thanks for having me. >> All right, so let's start with the customers. I always love talking to the Field CTOs you're out there. You're talking strategy, you're getting into some of the architecture, lots of customers, probably still, trying to figure out that whole cloud native containerization, Kubernetes and modernization piece. So when you talk to your customers, what are some of their biggest challenges they're facing and those main discussion points that bring them to talk to Mirantis. >> Very good question, I think you've just laid it out yourself in many ways. It's complexity our customers are dealing with more and more change, more and more options, and it's driving complexity in their environments, and they're looking for ways to deal with that complexity and to allow more and more access and reduce barriers to getting applications and getting tools to market. And if we look at it and we look at the way the world is going today, we have multiple cloud environments. We have every single developer on the face of the planet wants to use different tools, different ways to build applications that don't want to be dictated to. Now, if you turn that around and you look at what operators have to deal with, it's just more and more complexity. Ultimately, that complexity is growing and we're looking for ways to make it easier, simpler, and subsequently increase the speed of getting applications to market for our customers. >> Yeah, You know we talk a bit about some of the macro challenges that customers have. What talk you kind of teed up a little bit, the operators and the developers. I remember a couple of years ago, I had the opportunity to interview Solomon Hykes and of course the founder from Docker. And there was that talk of well, containerization, it's this wonderful thing for developers. And he's like, hold on Stu we actually, really started looking at this or the operators we want that unit of operation to be closer to the application. So it should be simpler, it used to be okay, how many different applications do they have on a server or VMs all over the place and containers I could really have this microservice or this application is a container. So there is some operational simplicity there, but how is that dynamic inside the customer? Of course, we've seen the growth and the importance and the embracing of developers, but there's still the DevOps adoption and we'd love to be able to say one of these years that, oh, we don't have silos anymore and everybody works together and we're all on the same page. >> Oh yeah, the reality is in the big enterprise companies and the companies that are building applications for market today, your big financial services companies, there's still a very clear separation between operators and developers. A lot of that is driven by legislation, a lot of that's driven by just old fashioned thinking in many ways, but developers are starting to have a lot more influence on what applications are used and the infrastructure. We just see with the rise of AWS, all the contenders to AWS in the form of Azure and Google. Developers are starting to have a lot more power over that decision, but they're still highly dependent on operators to deliver those platforms that they use, and to make sure that the platforms that they're running their applications on top of, are stable and run well in production situations. There's a big difference between building something on your laptop in one or two instances, and then trying to push it out to a massive scalable cloud platform. And I think those are the areas that we can have a lot of impact, and that's where we are building our tools for at the moment. >> Well, great. Let's dig into those tools a little bit, as I said, at the beginning, we're familiar, Mirantis had the Mirantis Cloud Platform for a few years, big embrace of Kubernetes and then Docker Enterprise, it comes into the mix. So help me understand a little bit, what is kind of the solution set to portfolio? How does Mirantis present that today? >> Yeah, well, it's been an interesting eight, nine months now of the whole process since with the Docker Enterprise business, a couple of key areas. So if we look at what MCP was, and MCP still here today apparently, it focuses on delivering all the components necessary to have an effective cloud platform. So lifecycle management, lifecycle management of all those underlying components, which in their own right is extremely complex set of software. What we focused on there was understanding in enterprise infrastructure, the right way to do that. As soon as you bring in from the Docker Enterprise business is that they have a scalable, large, well deployed container platform. And many thousands of users across the world in all sorts of different scales and production systems. We are merging that knowledge that we have around infrastructure, infrastructure management, and simplifying access to infrastructure with this platform that provides for all that application, hosting provides for all the control of containers, plus all the security components around the container lifecycle. And delivering in such a way that you can choose your underlying preference. So we're no longer looking to lock you in to say, you have to go on-prem, you have to go into cloud. We're saying, we'll give you the choice, but we'll also give you a standardized platform for your developers across all of those potential infrastructure environments, so I'll use it again, public cloud, private cloud, bare metal on-premise, or your options like the VMware of this world. By consolidating all of that into one platform, we're giving you that as a developer, the ability to write applications that'll run anyway and sorry, go on. >> No, please finish up, I've just got to follow, yeah. >> But that simplicity drives and like that's simple choice across all those platforms essentially drives speed. It takes away the typical barriers that we're seeing in our customers. We hear every reason, we love Docker Enterprise because it solves the problem of getting containers, it solves a problem of securing containers, but it takes four teams to deploy it. Same for the MCP, we're saying is we cannot do that in a day and provide other self sets. So you can deploy a brand new container cloud in minutes rather than days or weeks. And that's one of the biggest changes that we're bringing to the product. >> Yeah, so absolutely what we hear from customers that agility, that speed that you talk about is the imperative, especially talk about 2020, everybody has had to readjust often accelerating some of the plans they had to meet the realities of what we have today. What I want to understand is when you talk about that single platform being able to be in any environment, oftentimes there's a misnomer that it's about portability. Most customers we talked to, they're not moving things lots of places. They do want that operational consistency, wherever they go. At the same time, you mentioned the rise of AWS and the hyperscalers often when they're now going to have to manage multiple clusters, it's not that I choose one Kubernetes and I use it anywhere, but I might be using AKS Azure had an early version of it, of course, Amazon has a couple of options now for enterprises. So help us understand how the Mirantis solutions, fit with the clouds, leverage cloud services and if I have multiple clusters, you even mentioned VMware, I might have a VMware cluster, have something from Mirantis, have something from one of the hyperscalers. Is that what you're seeing from your customers today? And how do they and how do they want to manage that going forward? Because we understand this is still a maturing space. >> So I mean, that's exactly the point. What we're seeing from our customers is that they have policies to go cloud first. They still have a lot of infrastructure on-premise. The question is which cloud, which cloud suits their needs in which region. Now, all of a sudden you've got a risk management policy from an organization that says, well, I have to go to Azure and I have to go to AWS. That's using them as examples. The deployment and management of those two platforms is completely different. Just the learning curve for a developer who wants to focus on writing code, to build a platform on top of AWS is barely extensive. Yes, it's easy to get started, but if you really want to deal with the fine print of how to run some in production, it's not that simple. There are potentially a thousand different buttons, you can click when deploying an instance on Amazon. So what we're saying is, instead of you having to deal with that we're going to abstract that pain from you. We're going to say we'll deploy Docker Enterprise on top of Amazon, on top of Google, on top of Azure, on top of your VMware cluster, give you a consistent interface to that, consistent set of tools across all those platforms, still consuming those platforms as you would, but solve all those dependency problems. To set up a cube cluster on top of Amazon, I'm not talking about an AKS or something like that right now, but the sort of cube cluster means I have to set up load balances, I have to set up networks, I have to set up monitors, I have to set up the instances, I have to deploy Kubernetes, and then I'm only getting started. I still haven't integrated that to my corporate identity management. We're saying we'll bring all that to give them, we are bringing all that together in the form of Docker Enterprise container cloud. >> Yeah, definitely as you said, we need more simplicity here. The promise of cloud it's supposed to be simplicity and now of course we have the paradox of choice when it comes there. >> Yeah. >> One of the other things we've seen, rapid change a lot the last year or so is many of the offerings out there are now managed services. So as you said, I don't want to have to build all of those pieces. I want to just be able to go to somebody. What are you hearing from your customers? How does manage service fit into what Mirantis is doing? >> Great, well, what we're hearing from customers is they want the pain to go away. The answer to that could be delivered through software that's really easy to use and doesn't set up any barriers and gets them started fast, which is where we focus from a product perspective. Mirantis also has a strong manage services on so we've been doing manage services for some of the biggest enterprises in the world for MCP products for many years. We've brought those teams forward and we're now offering those same managed services on top of all of our platforms. So Docker Enterprise container cloud, we'll deploy it for you, we'll manage it for you. We'll handle all the dependencies around getting container cloud up and running within your organization, and then offer you that hands on service. So when you build clusters, when you want clusters that are much more longer lived, we can handle all the extra detail that goes around those. Short term, so if you just want quick clusters for your developers, easy access, you still have that as part of the service. So we're focusing on how fast can we get you started? How fast can you get up the cushions to market, not put any infrastructure barriers on the way, or where there are traditional infrastructure barriers find ways around us. That still acceptable to those enterprise operators who still have list as long as my arm, probably twice as long as my arm of fine print that they have to comply with for everything under the sun, the regulators, et cetera. >> Yeah, Shaun since you are based in Europe, I'm wondering if you can give us a little bit of the perspective on cloud adoption there, here in North America, discussion point has been for many years, just that massive movement to public cloud, of course governance the key issue in Europe and above also kind of the COVID impact, anecdotally there's lots of discussions of acceleration of public cloud. So what's the reality on the ground? How do your enterprise customers look at public cloud? How fast or slow are they moving and what is the 2020 impact? >> So interesting if you'd asked me this question, six months ago, seven months ago, pre COVID, I would have said public cloud is growing. People are still building some small private clouds for very unique use cases, looking at where our customers are now, all of a sudden there's a risk balance. So they're driving into public cloud, but they want those public clouds to be with European companies and European operators, or at least to have some level of security. You know, recently the European community canceled the privacy shield legislation that was in place between the US and Europe, which meant all of a sudden, a lot of companies in Europe had to look for other places to store their data, or had to deal with different rules around storing the data that they may have but in the US previously. What we're seeing customers saying is we have to go multicloud. The drive is no longer we can accept one vendor risk. We want to remove that risk, we will still have equipment on-premise. So on-prem equipment is still important to us, but as a backup to the public cloud, and as a way to secure our data and the mechanisms that we own and can touch and control. That's the operator's view. If we talk to developers, people writing applications, if they are not forced to, they will go public cloud almost every time. It's just easier for them. And that's really what we're, that's really the challenge that we're also trying to focus on here. >> Yeah, I'm curious, are there any European cloud providers that are rising to the top the big three have such a large megaphone that they kind of drown out a lot of discussion and understand that there's pockets and many local suppliers, and of course thousands of kind of cloud service providers out there, but any ones that are good partners of Mirantis or ones that you're hearing. >> There are a couple. >> Yeah. >> Sorry, there are a couple, I dunno if I can mention them here, but there's some great ones providing very unique businesses, places like the Netherlands, very unique, very focused business where they're taking advantage of specific laws within, well, the Netherlands and Germany, there's another company that we're working very closely with that feels that they can do a much more affordable, much more hands on service or cloud. So their cloud experience provide everything developers want, but at the same time handle those operator requirements and those enterprise requirements within Germany. So focusing on the GDPR laws, focusing on German technology laws, which are very complex, very much focused on privacy. And there are a few unique companies like that across Europe, I know of one in Italy, there's a company that focuses on providing cloud services to the EU government themselves, who we've worked with in the past. So yeah, but as you say, it's the big three, they're growing, they're dealing with those challenges. We see them as resources, we see them as partners to what we're trying to achieve. We certainly not trying to compete with them at that level. >> Absolutely, all right, Shaun final question I have for you, tell us what your customers see as the real differentiation, what draws them to Mirantis and what we should expect to see over the coming months? >> So I think choice is a key differentiator. We're offering choice, we're not trying to tell you you can only use one cloud platform or one cloud provider. And that's extremely important as one of the key differentiators. I've mentioned this many times, simplicity, driving simplicity at all levels, from the operator through to the developer, to the consumer of the cloud, let's make it easy. Let's truly reduce the friction to getting started, all right that's one of the really key focus areas for us and that's something we talk about all the time in every meeting and we question ourselves constantly is, does this make it easier? And then security is a major component for us. We really focus on security as part of our tool sets, providing that standardized platform and that standardized security across all of these environments, and ultimately reducing the complexity. >> Shaun O'Meara, thank you so much. Great to hear that the real customer interaction and what they're dealing with today. >> Sure, thank you very much. >> Be sure to check out the tracks for developers, for infrastructure as well as all the rest theCUBE interviews on the Mirantis Launchpad site of course powered by CUBE365. I'm Stu Miniman and thank you for watching theCUBE. (upbeat music)
SUMMARY :
to you by Mirantis . and to help us dig into of the architecture, of the planet wants to and of course the founder from Docker. all the contenders to AWS in Mirantis had the Mirantis the ability to write just got to follow, yeah. Same for the MCP, we're saying of the plans they had that they have policies to go cloud first. and now of course we have the paradox of the offerings out there that they have to comply with and above also kind of the COVID impact, or had to deal with different that are rising to the top So focusing on the GDPR laws, of the key differentiators. and what they're dealing with today. the tracks for developers,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Shaun | PERSON | 0.99+ |
Shaun O'Meara | PERSON | 0.99+ |
Europe | LOCATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
Italy | LOCATION | 0.99+ |
Germany | LOCATION | 0.99+ |
North America | LOCATION | 0.99+ |
one | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Netherlands | LOCATION | 0.99+ |
US | LOCATION | 0.99+ |
2020 | DATE | 0.99+ |
Mirantis | ORGANIZATION | 0.99+ |
two platforms | QUANTITY | 0.99+ |
Solomon Hykes | PERSON | 0.99+ |
eight | QUANTITY | 0.99+ |
six months ago | DATE | 0.99+ |
seven months ago | DATE | 0.99+ |
thousands | QUANTITY | 0.99+ |
one platform | QUANTITY | 0.99+ |
twice | QUANTITY | 0.99+ |
last year | DATE | 0.99+ |
GDPR | TITLE | 0.98+ |
ORGANIZATION | 0.98+ | |
today | DATE | 0.97+ |
four teams | QUANTITY | 0.97+ |
CUBE365 | ORGANIZATION | 0.97+ |
one cloud platform | QUANTITY | 0.97+ |
two instances | QUANTITY | 0.97+ |
EU government | ORGANIZATION | 0.96+ |
nine months | QUANTITY | 0.96+ |
European | OTHER | 0.96+ |
Mirantis Docker Enterprise | ORGANIZATION | 0.96+ |
single platform | QUANTITY | 0.96+ |
a day | QUANTITY | 0.96+ |
Docker Enterprise | ORGANIZATION | 0.94+ |
One | QUANTITY | 0.94+ |
one cloud provider | QUANTITY | 0.92+ |
Azure | TITLE | 0.91+ |
Docker | ORGANIZATION | 0.9+ |
couple of years ago | DATE | 0.88+ |
theCUBE | ORGANIZATION | 0.88+ |
Kubernetes | ORGANIZATION | 0.85+ |
first | QUANTITY | 0.83+ |
a couple | QUANTITY | 0.83+ |
a lot of companies | QUANTITY | 0.82+ |
thousands of users | QUANTITY | 0.77+ |
MCP | TITLE | 0.77+ |
Willem Du Plessis, Mirantis | Mirantis Launchpad 2020
>> Announcer: From around the globe, it's theCUBE, with digital coverage of Mirantis Launchpad 2020, brought to you by Mirantis. >> Welcome back I'm Stu Miniman, and this is theCUBE's coverage of Mirantis Launchpad 2020. Big event, multiple tracks powered by theCUBE365. Happy to welcome you to the program. We have a first time guest, Willem du Plessis. He's the Director of Customer Success and Operations with Mirantis. Willem, thanks so much for joining us. >> Hi Stu, thanks for having me. >> So customer success, of course, a big topic in the industry last few years. CX a is so important. Employee success and enabling that, but what, give us a little bit, your background and the purview that you and your team cover. >> Exactly, yeah, so everything under my umbrella would be basically post-sales. The whole customer experience after the point of a sale's been made so the whole account management, thereafter, the success of the accounts, as well as the health of the account, thereafter, that will be anything basically post-sales would be under my umbrella. >> Wonderful, well, the big piece is the shift. As we know, software went from shrink wrapped, and hardware talking about CapX to the cloud really ushered in OpX we're touching more subscription managed services and the like, so Mirantis has a subscription offering. Why don't you lay out for us the new pieces of this and how Mirantis puts together its offerings? >> Yeah, absolutely. So with the launch of our new product, Docker Enterprise Container Cloud, we're making two subscriptions available as well, named ProdCare, which is a 24/7 mission critical support offering and OpsCare, being a fully managed platform as a service subscription. Now, these offerings have been available on the Mirantis Cloud platform side of our business for quite some time, we've been very successful with them, so it's really excited making them available to our Docker Enterprise customers. So what we're trying to achieve with these accounts or with these subscriptions, rather, you know, 30% of the Fortune 100 companies are Mirantis customers, so we work on a day to day basis with their container and Kubernetes initiatives. So when we speak to these customers, there are really two trends that are becoming very clear, the first being the requirements of service providers or vendors being able to provide a true 24/7 experience. What I mean by that is not the ability to just react to an incident on a 24/7 basis. That's what I mean, what I mean is all of these companies would have operation centers spread across the globe. So it is at every hour of the day, it would be business as usual. And what these companies require is a, a partner or a service provider that can match that level, that way of operating. That is the first trend that we're noting. The second piece is really the, the evolution of the dev environment. The dev environment is no longer really seen as a secondary or a lower class citizen, if you want to call it, it's really become part of the whole DevOps pipeline, so it is really part of a mission critical process so that what customers, what we hear from our customers is that they require a real enterprise-grade subscription that they can cover this whole pipeline under and, you know, have the same quality of service from whether that is a dev or a production environment. So if you have a failure on your dev environment and your developer cannot push code, that is, is the same level of criticality than there would then they would be on if the failure was on the production environment. So this whole pipeline is decidedly seen as a mission critical component. And that's a great, that's really where ProdCare comes in. It is really this 24/7 mission critical follow the sun, enterprise-grade subscription that provides our customers with enhanced SLAs that, like I said, we've been running on the Mirantis Cloud platform side for quite some time, we've had some significant success with some really large companies. The second offering that we're making available with, like I said, is OpsCare. Now OpsCare is an ITIL-based managed service subscription, where we provide a platform as a service experience to a customer on their infrastructure of choice. So it is really irrelevant for us what your infrastructure is, whether that is on-prem or in the public cloud, as long as the product can support the infrastructure, you know, the subscription would be available for you and the experience would be very much the same. So what OpsCare, like I said, entails is, is this whole ITIL framework that would include, you know, the monitoring and managing of your alerts, the incident management process, the problem management process, as well as change management that would include the lifecycle management of the whole environment. And that would just enable our customers to run on the latest and greatest offer of our product at all times. And same as with ProdCare that's been available for our brass cloud platform customers for quite a while, and have seen some significant success with that, as well. >> Well, we definitely have seen that growth of the managed care offerings like you're talking about with OpsCare, you know, shift left is so important for companies to be able to focus on what's critically important. As you said, developers need to be enabled, it can't just be waiting for things or be, you know, relegated to, you know, have to wait in line or use something that's not optimal. What are some of those outcomes? What can companies do that they weren't able before? What are some of those successes that you're seeing with the managed care OpsCare solution? >> Yeah, so the real way we OpsCare really comes to its own is allowing the customer ability to focus on what is important to their business and spend less time on what we call, keep the lights on. What I mean by that is they're solely focused on developing the application, developing the workload and spend basically no time on managing the infrastructure and, you know, maintaining it, or, you know, providing, do whatever to, to keep the platform stable, because that is done by Mirantis, already. So for example, if we take 2020 year to date, all the platforms running under OpsCare has an availability number of above four nines, and that is a significant number. So that really just sets such a strong foundation for a customer to just have that sole focus on, on what is important to them and, you know, just sets that foundation for them to develop their workload, to develop their business, and achieve their goals. >> Well, what about when it comes to the managing and monitoring of the environment? What kind of metrics are your customers having? Help us understand what the customer still does themselves or the reporting they're getting and what Mirantis, I'm assuming there's probably a Tam involved for at least some of the larger accounts there. Help us understand that shared responsibility, if you would for these type of environments. >> Yeah, exactly. So the whole ITIL framework, as I explained earlier, incident management, problem management, change, all of that, this is wrapped around why a customer success manager that is, you know, brings a single level of ownership on an accountability, and just have a customer direct for a single point of contact as a business partner. So all this is all our customers, their primary KPI or metric that we look at is just the availability of the platform. That is the primary SLA and thereafter, all of the other things happening, you know, the success of the workload and so on, because there's a lot of things that makes the result of the workload, not just the platform or the infrastructure, it's the quality of the workload, and so on, and so forth. But the main metric our customer would be looking at is that availability number, you know, how available and how stable and accessible is the environment, and, you know, like I said, just removing that requirement for them to spend, basically, no time on the platform or the infrastructure, and just focusing on the workload. >> Yeah, when it comes to in the field, your field, your partners, that line between ProdCare and OpsCare, obviously, the trend is going towards, you know, the fully managed option, but what guidance do you have out there, or what trends do you seeing? Is it a certain size company, that tends to be trending that way? Are there certain verticals that may be are further ahead? What's the reality, today? What do you expect to see over the next kind of six, 12 months? >> Yeah, so most of the companies that we see that as, that is engaging with us on an OpsCare, or managed service engagement, you know, they have the ambitions to go down the block model and build, operate, transfer, you know, to take the operations over themselves, at some point, and we have that option available to them, if they wish to choose it further along the line. What we do find is, is that they, that they don't really, you know, exercise that later on. It is, we do find it is such a smooth integration with our customers, that they tend to stay on OpsCare and see the value. This is actually a money saver for them, if they could, just focus their efforts on building, you know, focusing their time on the workload on top of the platform. From a vertical perspective, it's really anything and everything. We have customers in the science and research, we have TELCOs, large manufacturing, manufacturing, a lot of large organizations. There's really the breadth of the verticals that we see that are utilizing OpsCare and not even to mention ProdCare, that's really everything in there, as well. So it is not a really a subscription that is, that is custom for one vertical. It is basically something that we, that any vertical can actually utilize and find a significant amount of value in. >> All right, well, what final words do you have that you want to leave everyone with today? >> Yeah, so over the last six to nine months, you know, we've invested a significant amount of resources in the Docker Enterprise support business and we just with one focus, and that is just to take the support business to the next level and improve or give the customers an optimal customer experience. So with the availability of all these new subscriptions, I'm really excited to engage with our Docker Enterprise customers with these new, enhanced SLAs and just be able to work with them on these, like I said, enhanced subscriptions and just see, just give them a better customer experience. So, I'm really looking forward to working with them on the subscriptions. >> Willem, thank you so much for all the updates and want to welcome everyone to be sure to check out all the rest of the tracks on the Launchpad 2020 event. I'm Stu Miniman and thank you for watching theCUBE. (soft electronic music)
SUMMARY :
the globe, it's theCUBE, Happy to welcome you to the program. that you and your team cover. so the whole account managed services and the So it is at every hour of the day, of the managed care offerings Yeah, so the real way we OpsCare really and monitoring of the environment? that makes the result of the workload, of the verticals that we see and that is just to take on the Launchpad 2020 event.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Willem du Plessis | PERSON | 0.99+ |
Stu | PERSON | 0.99+ |
Willem Du Plessis | PERSON | 0.99+ |
Willem | PERSON | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
OpsCare | TITLE | 0.99+ |
second piece | QUANTITY | 0.99+ |
30% | QUANTITY | 0.99+ |
Mirantis | ORGANIZATION | 0.99+ |
ProdCare | TITLE | 0.99+ |
first time | QUANTITY | 0.99+ |
six | QUANTITY | 0.98+ |
2020 year | DATE | 0.98+ |
first | QUANTITY | 0.98+ |
two trends | QUANTITY | 0.98+ |
first trend | QUANTITY | 0.98+ |
TELCOs | ORGANIZATION | 0.98+ |
Mirantis | EVENT | 0.98+ |
second offering | QUANTITY | 0.97+ |
Launchpad 2020 | EVENT | 0.97+ |
today | DATE | 0.97+ |
single point | QUANTITY | 0.95+ |
single | QUANTITY | 0.94+ |
theCUBE | ORGANIZATION | 0.93+ |
12 months | QUANTITY | 0.92+ |
two subscriptions | QUANTITY | 0.92+ |
one focus | QUANTITY | 0.92+ |
nine months | QUANTITY | 0.9+ |
one vertical | QUANTITY | 0.9+ |
OpsCare | ORGANIZATION | 0.89+ |
ProdCare | ORGANIZATION | 0.87+ |
OpX | ORGANIZATION | 0.84+ |
Docker Enterprise | ORGANIZATION | 0.81+ |
above four nines | QUANTITY | 0.8+ |
last few years | DATE | 0.78+ |
Mirantis | PERSON | 0.73+ |
100 companies | QUANTITY | 0.69+ |
Cloud | COMMERCIAL_ITEM | 0.69+ |
Launchpad 2020 | TITLE | 0.66+ |
CapX | TITLE | 0.63+ |
theCUBE365 | ORGANIZATION | 0.63+ |
Mirantis Launchpad 2020 | TITLE | 0.59+ |
Docker Enterprise Container | TITLE | 0.54+ |
Kubernetes | TITLE | 0.53+ |
Fortune | QUANTITY | 0.53+ |
CX | TITLE | 0.5+ |
Cloud | TITLE | 0.43+ |
DOCKER CLI FINAL
>>Hello, My name is John John Sheikh from Iran Tous. Welcome to our session on new extensions for doctors CLI as we all know, containers air everywhere. Kubernetes is coming on strong and the CNC F cloud landscape slide has become a marvel to behold its complexities about to surpass that of the photo. Letha dies used to fabricate the old intel to 86 and future generations of the diagram will be built out and up into multiple dimensions using extreme ultraviolet lithography. Meanwhile, complexity is exploding and uncertainty about tools, platform details, processes and the economic viability of our companies in changing and challenging times is also increasing. Mirant ous, as you've already heard today, believes that achieving speed is critical and that speed results from balancing choice with simplicity and security. You've heard about Dr Enterprise Container Cloud, a new framework built on kubernetes, the less you deploy compliant, secure by default. Cooper nineties clusters on any infrastructure, providing a seamless self service capable cloud experience to developers. Get clusters fast, Justus, you need them, Update them seamlessly. Scale them is needed all while keeping workloads running smoothly. And you've heard how Dr Enterprise Container Cloud also provides all the day one and Day two and observe ability, tools, the integration AP ICE and Top Down Security, Identity and Secrets management to run operations efficiently. You've also heard about Lens, an open source i D for kubernetes. Aimed at speeding up the most banding, tightest inner loop of kubernetes application development. Lens beautifully meets the needs of a new class of developers who need to deal with multiple kubernetes clusters. Multiple absent project sufficiently developers who find themselves getting bogged down and seal I only coop CTL work flows and context switches into and out of them. But what about Dr Developers? They're working with the same core technologies all the time. They're accessing many of the same amenities, including Docker, engine Enterprise, Docker, Trusted registry and so on. Sure, their outer loop might be different. For example, they might be orchestrating on swarm. Many companies are our future of Swarm session talks about the ongoing appeal of swarm and Miranda's commitment to maintaining and extending the capabilities of swarm Going forward. Dr Enterprise Container Cloud can, of course, deployed doctor enterprise clusters with 100% swarm orchestration on computes just Aziza Leah's. It can provide kubernetes orchestration or mixed swarming kubernetes clusters. The problem for Dr Dev's is that nobody's given them an easy way to use kubernetes without a learning curve and without getting familiar with new tools and work flows, many of which involved buoys and are somewhat tedious for people who live on the command line and like it that way until now. In a few moments you'll meet my colleagues Chris Price and Laura Powell, who enact a little skit to introduce and demonstrate our new extended docker CLI plug in for kubernetes. That plug in offers seamless new functionality, enabling easy context management between the doctor Command Line and Dr Enterprise Clusters deployed by Dr Enterprise Container Cloud. We hope it will help Dev's work faster, help them adapt decay. TSA's they and their organizations manage platform coexistence or transition. Here's Chris and Laura, or, as we like to call them, developer A and B. >>Have you seen the new release of Docker Enterprise Container Cloud? I'm already finding it easier to manage my collection of UCP clusters. >>I'm glad it's helping you. It's great we can manage multiple clusters, but the user interface is a little bit cumbersome. >>Why is that? >>Well, if I want to use docker cli with a cluster, I need to download a client bundle from UCP and use it to create a contact. I like that. I can see what's going on, but it takes a lot of steps. >>Let me guess. Are these the steps? First you have to navigate to the web. You i for docker Enterprise Container Cloud. You need to enter your user name and password. And since the cluster you want to access is part of the demo project, you need to change projects. Then you have to choose a cluster. So you choose the first demo cluster here. Now you need to visit the U C p u I for that cluster. You can use the link in the top right corner of the page. Is that about right? >>Uh yep. >>And this takes you to the UCP you. I log in page now you can enter your user name and password again, but since you've already signed in with key cloak, you can use that instead. So that's good. Finally, you've made it to the landing page. Now you want to download a client bundle what you can do by visiting your user profile, you'll generate a new bundle called Demo and download it. Now that you have the bundle on your local machine, you can import it to create a doctor context. First, let's take a look at the context already on your machine. I can see you have the default context here. Let's import the bundle and call it demo. If we look at our context again, you can see that the demo context has been created. Now you can use the context and you'll be able to interact with your UCP cluster. Let's take a look to see if any stacks are running in the cluster. I can see you have a stack called my stack >>in >>the default name space running on Kubernetes. We can verify that by checking the UCP you I and there it iss my stack in the default name space running on Kubernetes. Let's try removing the stack just so we could be sure we're dealing with the right cluster and it disappears. As you can see. It's easy to use the Docker cli once you've created a context, but it takes quite a bit of effort to create one in the first place. Imagine? >>Yes. Imagine if you had 10 or 20 or 50 clusters toe work with. It's a management nightmare. >>Haven't you heard of the doctor Enterprise Container Cloud cli Plug in? >>No, >>I think you're going to like it. Let me show you how it works. It's already integrated with the docker cli You start off by setting it up with your container cloud Instance, all you need to get started is the base. You are all of your container cloud Instance and your user name and password. I'll set up my clothes right now. I have to enter my user name and password this one time only. And now I'm all set up. >>But what does it actually dio? >>Well, we can list all of our clusters. And as you can see, I've got the cluster demo one in the demo project and the cluster demo to in the Demo project Taking a look at the web. You I These were the same clusters we're seeing there. >>Let me check. Looks good to me. >>Now we can select one of these clusters, but let's take a look at our context before and after so we can understand how the plug in manages a context for us. As you can see, I just have my default contact stored right now, but I can easily get a context for one of our clusters. Let's try demo to the plug in says it's created a context called Container Cloud for me and it's pointing at the demo to cluster. Let's see what our context look like now and there's the container cloud context ready to go. >>That's great. But are you saying once you've run the plug in the doctor, cli just works with that cluster? >>Sure. Let me show you. I've got a doctor stack right here and it deploys WordPress. Well, the play it to kubernetes for you. Head over to the U C P u I for the cluster so you can verify for yourself. Are you ready? >>Yes. >>First I need to make sure I'm using the context >>and >>then I can deploy. And now we just have to wait for the deployment to complete. It's as easy as ever. >>You weren't lying. Can you deploy the same stack to swarm on my other clusters? >>Of course. And that should also show you how easy it is to switch between clusters. First, let's just confirm that our stack has reported as running. I've got a stack called WordPress demo in the default name space running on Kubernetes to deploy to the other cluster. First I need to select it that updates the container cloud context so I don't even need to switch contexts, since I'm already using that one. If I check again for running stacks, you can see that our WordPress stack is gone. Bring up the UCP you I on your other cluster so you can verify the deployment. >>I'm ready. >>I'll start the deployment now. It should be appearing any moment. >>I see the services starting up. That's great. It seems a lot easier than managing context manually. But how do I know which cluster I'm currently using? >>Well, you could just list your clusters like So do you see how this one has an asterisk next to its name? That means it's the currently selected cluster >>I'm sold. Where can I get the plug in? >>Just go to get hub dot com slash miran tous slash container dash cloud dash cli and follow the instructions
SUMMARY :
built on kubernetes, the less you deploy compliant, secure by default. Have you seen the new release of Docker Enterprise Container Cloud? but the user interface is a little bit cumbersome. I can see what's going on, but it takes a lot of steps. Then you have to choose a cluster. what you can do by visiting your user profile, you'll generate the UCP you I and there it iss my stack It's a management nightmare. Let me show you how it works. I've got the cluster demo one in the demo project and the cluster demo to in Looks good to at the demo to cluster. But are you saying once you've run the plug in the doctor, Head over to the U C P u I for the cluster so you can verify for yourself. And now we just have to wait for the deployment to complete. Can you deploy the same stack to swarm And that should also show you how easy it is to switch between clusters. I'll start the deployment now. I see the services starting up. Where can I get the plug in?
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Laura Powell | PERSON | 0.99+ |
Chris | PERSON | 0.99+ |
Chris Price | PERSON | 0.99+ |
John John Sheikh | PERSON | 0.99+ |
Laura | PERSON | 0.99+ |
100% | QUANTITY | 0.99+ |
10 | QUANTITY | 0.99+ |
20 | QUANTITY | 0.99+ |
First | QUANTITY | 0.99+ |
Aziza Leah | PERSON | 0.97+ |
50 clusters | QUANTITY | 0.97+ |
docker Enterprise Container Cloud | TITLE | 0.95+ |
Kubernetes | TITLE | 0.94+ |
86 | QUANTITY | 0.94+ |
WordPress | ORGANIZATION | 0.93+ |
today | DATE | 0.92+ |
one | QUANTITY | 0.91+ |
one time | QUANTITY | 0.9+ |
Docker Enterprise Container Cloud | TITLE | 0.89+ |
Dr Enterprise Container Cloud | TITLE | 0.88+ |
first demo cluster | QUANTITY | 0.88+ |
Miranda | PERSON | 0.85+ |
Iran Tous | ORGANIZATION | 0.84+ |
intel | ORGANIZATION | 0.84+ |
Lens | TITLE | 0.83+ |
TSA | ORGANIZATION | 0.83+ |
Cooper nineties | ORGANIZATION | 0.81+ |
Day two | QUANTITY | 0.78+ |
Dr | TITLE | 0.73+ |
Docker | ORGANIZATION | 0.73+ |
first place | QUANTITY | 0.71+ |
WordPress | TITLE | 0.71+ |
Enterprise Container Cloud | COMMERCIAL_ITEM | 0.65+ |
Letha | PERSON | 0.59+ |
Cloud | COMMERCIAL_ITEM | 0.58+ |
DOCKER CLI | TITLE | 0.57+ |
Mirant | TITLE | 0.57+ |
day | QUANTITY | 0.55+ |
Trusted | ORGANIZATION | 0.51+ |
Dr Enterprise Clusters | TITLE | 0.47+ |
Dr Enterprise | TITLE | 0.46+ |
Cloud | TITLE | 0.43+ |
Dr | PERSON | 0.42+ |
Enterprise | COMMERCIAL_ITEM | 0.33+ |
Swarm | ORGANIZATION | 0.33+ |
SEAGATE AI FINAL
>>C G technology is focused on data where we have long believed that data is in our DNA. We help maximize humanity's potential by delivering world class, precision engineered data solutions developed through sustainable and profitable partnerships. Included in our offerings are hard disk drives. As I'm sure many of you know, ah, hard drive consists of a slider also known as a drive head or transducer attached to a head gimbal assembly. I had stack assembly made up of multiple head gimbal assemblies and a drive enclosure with one or more platters, or just that the head stacked assembles into. And while the concept hasn't changed, hard drive technology has progressed well beyond the initial five megabytes, 500 quarter inch drives that Seagate first produced. And, I think 1983. We have just announced in 18 terabytes 3.5 inch drive with nine flatters on a single head stack assembly with dual head stack assemblies this calendar year, the complexity of these drives further than need to incorporate Edge analytics at operation sites, so G Edward stemming established the concept of continual improvement and everything that we do, especially in product development and operations and at the end of World War Two, he embarked on a mission with support from the US government to help Japan recover from its four time losses. He established the concept of continual improvement and statistical process control to the leaders of prominent organizations within Japan. And because of this, he was honored by the Japanese emperor with the second order of the sacred treasure for his teachings, the only non Japanese to receive this honor in hundreds of years. Japan's quality control is now world famous, as many of you may know, and based on my own experience and product development, it is clear that they made a major impact on Japan's recovery after the war at Sea Gate. The work that we've been doing and adopting new technologies has been our mantra at continual improvement. As part of this effort, we embarked on the adoption of new technologies in our global operations, which includes establishing machine learning and artificial intelligence at the edge and in doing so, continue to adopt our technical capabilities within data science and data engineering. >>So I'm a principal engineer and member of the Operations and Technology Advanced Analytics Group. We are a service organization for those organizations who need to make sense of the data that they have and in doing so, perhaps introduce a different way to create an analyzed new data. Making sense of the data that organizations have is a key aspect of the work that data scientist and engineers do. So I'm a project manager for an initiative adopting artificial intelligence methodologies for C Gate manufacturing, which is the reason why I'm talking to you today. I thought I'd start by first talking about what we do at Sea Gate and follow that with a brief on artificial intelligence and its role in manufacturing. And I'd like them to discuss how AI and machine Learning is being used at Sea Gate in developing Edge analytics, where Dr Enterprise and Cooper Netease automates deployment, scaling and management of container raised applications. So finally, I like to discuss where we are headed with this initiative and where Mirant is has a major role in case some of you are not conversant in machine learning, artificial intelligence and difference outside some definitions. To cite one source, machine learning is the scientific study of algorithms and statistical bottles without computer systems use to effectively perform a specific task without using explicit instructions, relying on patterns and inference Instead, thus, being seen as a subset of narrow artificial intelligence were analytics and decision making take place. The intent of machine learning is to use basic algorithms to perform different functions, such as classify images to type classified emails into spam and not spam, and predict weather. The idea and this is where the concept of narrow artificial intelligence comes in, is to make decisions of a preset type basically let a machine learn from itself. These types of machine learning includes supervised learning, unsupervised learning and reinforcement learning and in supervised learning. The system learns from previous examples that are provided, such as images of dogs that are labeled by type in unsupervised learning. The algorithms are left to themselves to find answers. For example, a Siris of images of dogs can be used to group them into categories by association that's color, length of coat, length of snout and so on. So in the last slide, I mentioned narrow a I a few times, and to explain it is common to describe in terms of two categories general and narrow or weak. So Many of us were first exposed to General Ai in popular science fiction movies like 2000 and One, A Space Odyssey and Terminator General Ai is a I that can successfully perform any intellectual task that a human can. And if you ask you Lawn Musk or Stephen Hawking, this is how they view the future with General Ai. If we're not careful on how it is implemented, so most of us hope that is more like this is friendly and helpful. Um, like Wally. The reality is that machines today are not only capable of weak or narrow, a I AI that is focused on a narrow, specific task like understanding, speech or finding objects and images. Alexa and Google Home are becoming very popular, and they can be found in many homes. Their narrow task is to recognize human speech and answer limited questions or perform simple tasks like raising the temperature in your home or ordering a pizza as long as you have already defined the order. Narrow. AI is also very useful for recognizing objects in images and even counting people as they go in and out of stores. As you can see in this example, so artificial intelligence supplies, machine learning analytics inference and other techniques which can be used to solve actual problems. The two examples here particle detection, an image anomaly detection have the potential to adopt edge analytics during the manufacturing process. Ah, common problem in clean rooms is spikes in particle count from particle detectors. With this application, we can provide context to particle events by monitoring the area around the machine and detecting when foreign objects like gloves enter areas where they should not. Image Anomaly detection historically has been accomplished at sea gate by operators in clean rooms, viewing each image one at a time for anomalies, creating models of various anomalies through machine learning. Methodologies can be used to run comparative analyses in a production environment where outliers can be detected through influence in an automated real Time analytics scenario. So anomaly detection is also frequently used in machine learning to find patterns or unusual events in our data. How do you know what you don't know? It's really what you ask, and the first step in anomaly detection is to use an algorithm to find patterns or relationships in your data. In this case, we're looking at hundreds of variables and finding relationships between them. We can then look at a subset of variables and determine how they are behaving in relation to each other. We use this baseline to define normal behavior and generate a model of it. In this case, we're building a model with three variables. We can then run this model against new data. Observations that do not fit in the model are defined as anomalies, and anomalies can be good or bad. It takes a subject matter expert to determine how to classify the anomalies on classify classification could be scrapped or okay to use. For example, the subject matter expert is assisting the machine to learn the rules. We then update the model with the classifications anomalies and start running again, and we can see that there are few that generate these models. Now. Secret factories generate hundreds of thousands of images every day. Many of these require human toe, look at them and make a decision. This is dull and steak prone work that is ideal for artificial intelligence. The initiative that I am project managing is intended to offer a solution that matches the continual increased complexity of the products we manufacture and that minimizes the need for manual inspection. The Edge Rx Smart manufacturing reference architecture er, is the initiative both how meat and I are working on and sorry to say that Hamid isn't here today. But as I said, you may have guessed. Our goal is to introduce early defect detection in every stage of our manufacturing process through a machine learning and real time analytics through inference. And in doing so, we will improve overall product quality, enjoy higher yields with lesser defects and produce higher Ma Jin's. Because this was entirely new. We established partnerships with H B within video and with Docker and Amaranthus two years ago to develop the capability that we now have as we deploy edge Rx to our operation sites in four continents from a hardware. Since H P. E. And in video has been an able partner in helping us develop an architecture that we have standardized on and on the software stack side doctor has been instrumental in helping us manage a very complex project with a steep learning curve for all concerned. To further clarify efforts to enable more a i N M l in factories. Theobald active was to determine an economical edge Compute that would access the latest AI NML technology using a standardized platform across all factories. This objective included providing an upgrade path that scales while minimizing disruption to existing factory systems and burden on factory information systems. Resource is the two parts to the compute solution are shown in the diagram, and the gateway device connects to see gates, existing factory information systems, architecture ER and does inference calculations. The second part is a training device for creating and updating models. All factories will need the Gateway device and the Compute Cluster on site, and to this day it remains to be seen if the training devices needed in other locations. But we do know that one devices capable of supporting multiple factories simultaneously there are also options for training on cloud based Resource is the stream storing appliance consists of a kubernetes cluster with GPU and CPU worker notes, as well as master notes and docker trusted registries. The GPU nodes are hardware based using H B E l 4000 edge lines, the balance our virtual machines and for machine learning. We've standardized on both the H B E. Apollo 6500 and the NVIDIA G X one, each with eight in video V 100 GP use. And, incidentally, the same technology enables augmented and virtual reality. Hardware is only one part of the equation. Our software stack consists of Docker Enterprise and Cooper Netease. As I mentioned previously, we've deployed these clusters at all of our operations sites with specific use. Case is planned for each site. Moran Tous has had a major impact on our ability to develop this capability by offering a stable platform in universal control plane that provides us, with the necessary metrics to determine the health of the Kubernetes cluster and the use of Dr Trusted Registry to maintain a secure repository for containers. And they have been an exceptional partner in our efforts to deploy clusters at multiple sites. At this point in our deployment efforts, we are on prem, but we are exploring cloud service options that include Miranda's next generation Docker enterprise offering that includes stack light in conjunction with multi cluster management. And to me, the concept of federation of multi cluster management is a requirement in our case because of the global nature of our business where our operation sites are on four continents. So Stack Light provides the hook of each cluster that banks multi cluster management and effective solution. Open source has been a major part of Project Athena, and there has been a debate about using Dr CE versus Dr Enterprise. And that decision was actually easy, given the advantages that Dr Enterprise would offer, especially during a nearly phase of development. Cooper Netease was a natural addition to the software stack and has been widely accepted. But we have also been a work to adopt such open source as rabbit and to messaging tensorflow and tensor rt, to name three good lab for developments and a number of others. As you see here, is well, and most of our programming programming has been in python. The results of our efforts so far have been excellent. We are seeing a six month return on investment from just one of seven clusters where the hardware and software cost approached close to $1 million. The performance on this cluster is now over three million images processed per day for their adoption has been growing, but the biggest challenge we've seen has been handling a steep learning curve. Installing and maintaining complex Cooper needs clusters in data centers that are not used to managing the unique aspect of clusters like this. And because of this, we have been considering adopting a control plane in the cloud with Kubernetes as the service supported by Miranda's. Even without considering, Kubernetes is a service. The concept of federation or multi cluster management has to be on her road map, especially considering the global nature of our company. Thank you.
SUMMARY :
at the end of World War Two, he embarked on a mission with support from the US government to help and the first step in anomaly detection is to use an algorithm to find patterns
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Seagate | ORGANIZATION | 0.99+ |
hundreds of years | QUANTITY | 0.99+ |
two parts | QUANTITY | 0.99+ |
python | TITLE | 0.99+ |
six month | QUANTITY | 0.99+ |
World War Two | EVENT | 0.99+ |
C Gate | ORGANIZATION | 0.99+ |
one | QUANTITY | 0.99+ |
Stephen Hawking | PERSON | 0.99+ |
Sea Gate | ORGANIZATION | 0.99+ |
Japan | LOCATION | 0.99+ |
Lawn Musk | PERSON | 0.99+ |
Terminator | TITLE | 0.99+ |
1983 | DATE | 0.99+ |
one part | QUANTITY | 0.99+ |
two examples | QUANTITY | 0.99+ |
A Space Odyssey | TITLE | 0.99+ |
five megabytes | QUANTITY | 0.99+ |
3.5 inch | QUANTITY | 0.99+ |
second part | QUANTITY | 0.99+ |
18 terabytes | QUANTITY | 0.99+ |
first step | QUANTITY | 0.99+ |
hundreds | QUANTITY | 0.99+ |
both | QUANTITY | 0.98+ |
NVIDIA | ORGANIZATION | 0.98+ |
over three million images | QUANTITY | 0.98+ |
first | QUANTITY | 0.98+ |
each site | QUANTITY | 0.98+ |
H B E. Apollo 6500 | COMMERCIAL_ITEM | 0.98+ |
each cluster | QUANTITY | 0.98+ |
each image | QUANTITY | 0.98+ |
one source | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
G X one | COMMERCIAL_ITEM | 0.98+ |
Cooper | PERSON | 0.98+ |
second order | QUANTITY | 0.98+ |
Japan | ORGANIZATION | 0.98+ |
Hamid | PERSON | 0.97+ |
Dr Enterprise | ORGANIZATION | 0.97+ |
Cooper Netease | ORGANIZATION | 0.97+ |
each | QUANTITY | 0.97+ |
One | TITLE | 0.97+ |
Theobald | PERSON | 0.97+ |
nine flatters | QUANTITY | 0.97+ |
one devices | QUANTITY | 0.96+ |
Siris | TITLE | 0.96+ |
hundreds of thousands of images | QUANTITY | 0.96+ |
Docker Enterprise | ORGANIZATION | 0.95+ |
Docker | ORGANIZATION | 0.95+ |
seven clusters | QUANTITY | 0.95+ |
two years ago | DATE | 0.95+ |
US government | ORGANIZATION | 0.95+ |
Mirant | ORGANIZATION | 0.95+ |
Operations and Technology Advanced Analytics Group | ORGANIZATION | 0.94+ |
four time losses | QUANTITY | 0.94+ |
Wally | PERSON | 0.94+ |
Japanese | OTHER | 0.93+ |
two categories | QUANTITY | 0.93+ |
H B E l 4000 | COMMERCIAL_ITEM | 0.9+ |
H B | ORGANIZATION | 0.9+ |
three variables | QUANTITY | 0.9+ |
General Ai | TITLE | 0.87+ |
G Edward | PERSON | 0.87+ |
Google Home | COMMERCIAL_ITEM | 0.87+ |
$1 million | QUANTITY | 0.85+ |
Miranda | ORGANIZATION | 0.85+ |
Sea Gate | LOCATION | 0.85+ |
Alexa | TITLE | 0.85+ |
500 quarter inch drives | QUANTITY | 0.84+ |
Kubernetes | TITLE | 0.83+ |
single head | QUANTITY | 0.83+ |
eight | QUANTITY | 0.83+ |
Dr | TITLE | 0.82+ |
variables | QUANTITY | 0.81+ |
this calendar year | DATE | 0.78+ |
H P. E. | ORGANIZATION | 0.78+ |
2000 | DATE | 0.73+ |
Project Athena | ORGANIZATION | 0.72+ |
Rx Smart | COMMERCIAL_ITEM | 0.69+ |
dual | QUANTITY | 0.68+ |
V 100 | COMMERCIAL_ITEM | 0.65+ |
close | QUANTITY | 0.65+ |
four continents | QUANTITY | 0.64+ |
GP | QUANTITY | 0.62+ |
Adrian and Adam Keynote v4 fixed audio blip added slide
>>Welcome everyone. Good morning. Good evening to all of you around the world. I am so excited to welcome you to launch bad our annual conference for customers, for partners, for our own colleagues here at Mirandes. This is meant to be a forum for learning, for sharing for discovery. One of openness. We're incredibly excited. Do you have you here with us? I want to take a few minutes this morning and opened the conference and share with you first and foremost where we're going as a company. What is our vision then? I also want to share with you on update on what we have been up to you for the past year. Especially with two important acquisitions, Doc Enterprise and then container and lens. And what are some of the latest developments at Mirandes? And then I'll close also with an exciting announcement that we have today, which we hope is going to be interesting and valuable for all of you. But let me start with our mission. What are we here to Dio? It's very simple. We want to help you the ship code faster. This is something that we're very excited about, something that we have achieved for many of you around the world. And we just want thio double down on. We feel this is a mission that's very much worthwhile and relevant and important to you. Now, how do we do that? How do we help you ship code faster? There are three things we believe in. We believe in this world of cloud. Um, choice is incredibly important. We all know that developers want to use the latest tools. We all know that cloud technology is evolving very quickly and new innovations appear, um, very, very quickly, and we want to make them available to you. So choice is very important. At the same time, consuming choice can be difficult. So our mission is to make choice simple for you to give developers and operators simplicity and then finally underpinning everything that we dio is security. These are the three big things that we invest in and that we believe that choice, simplicity and security and the foundation technology that we're betting on to make that happen for you is kubernetes many of you, many of our customers use kubernetes from your aunties today and they use it at scale. And this is something we want to double down on the fundamental benefit. The our key promise we want to deliver for you is Speed. And we feel this is very relevant and important and and valuable in the world that we are in today. So you might also be interested in what have been our priorities since we acquired Doc Enterprise. What has happened for the past year at Miranda's And there are three very important things we focused on as a company. The first one is customer success. Um, when we acquired Doc Enterprise, the first thing we did is listen to you connect with the most important customers and find out what was your sentiment. What did you like? What were you concerned about? What needed to improve? How can we create more value and a better experience for you? So, customers success has been a top of our list of priorities ever since. And here is what we've heard here is what you've told us. You've told us that you very much appreciated the technology that you got a lot of value out of the technology, but that at the same time, there are some things that we can do better. Specifically, you wanted better. Sele's better support experience. You also wanted more clarity on the road map. You also wanted to have a deeper alignment and a deeper relationship between your needs and your requirements and our our technical development that keep people in our development organization are most important engineers. So those three things are were very, very important to you and they were very important to us here. So we've taken that to heart and over the past 12 months, we believe, as a team, we have dramatically improved the customer support experience. We introduced new SLS with prod care. We've rolled out a roadmap to many many of our customers. We've taken your requirements of the consideration and we've built better and deeper relationships with so many of you. And the evidence for that that we've actually made some progress is in a significant increase off the work clothes and in usage of all platforms. I was so fortunate that we were able to build better and stronger relationships and take you to the next level of growth for companies like Visa like soc T general, like nationwide, like Bosch, like Axa X l like GlaxoSmithKline, like standard and Poor's, like Apple A TNT. So many, many off you, Many of all customers around the world, I believe over the past 12 months have experienced better, better, better support strong s L. A s a deeper relationship and a lot more clarity on our roadmap and our vision forward. The second very big priority for us over the last year has been product innovation. This is something that we are very excited about that we've invested. Most of our resource is in, and we've delivered some strong proof points. Doc Enterprise 3.1 has been the first release that we have shipped. Um, as Mirant is as the unified company, Um, it's had some big innovative features or Windows support or a I and machine learning use cases and a significant number off improvements in stability and scalability earlier this year. We're very excited to have a quiet lens and container team, which is by far the most popular kubernetes. I'd, um, in the world today and every day, 600 new users are starting to use lens to manage the community's clusters to deploy applications on top of communities and to dramatically simplify the experience for communities for operators and developers alike. That is a very big step forward for us as a company. And then finally, this week at this conference, we announcing our latest product, which we believe is a huge step forward for Doc Enterprise and which we call Doc Enterprise, Container Cloud, and you will hear a lot more about that during this conference. The third vector of development, the third priority for us as a company over the past year was to become mawr and Mawr developer centric. As we've seen over the past 10 years, developers really move the world forward. They create innovation, they create new software. And while our platform is often managed and run and maybe even purchased by RT architects and operators and I T departments, the actual end users are developers. And we made it our mission a za company, to become closer and closer to developers to better understand their needs and to make our technology as easy and fast to consume as possible for developers. So as a company, we're becoming more and more developers centric, really. The two core products which fit together extremely well to make that happen, or lens, which is targeted squarely at a new breed off kubernetes developers sitting on the desktop and managing communities, environments and the applications on top on any cloud platform anywhere and then DACA enterprise contain a cloud which is a new and radically innovative, contain a platform which we're bringing to market this week. So with this a za background, what is the fundamental problem which we solve for you, for our customers? What is it that we feel are are your pain points that can help you resolve? We see too very, very big trends in the world today, which you are experiencing. On one side, we see the power of cloud emerging with more features mawr innovation, more capabilities coming to market every day. But with those new features and new innovations, there is also an exponential growth in cloud complexity and that cloud complexity is becoming increasingly difficult to navigate for developers and operators alike. And at the same time, we see the pace of change in the economy continuing to accelerate on bits in the economy and in the technology as well. So when you put these two things together on one hand, you have MAWR and Mawr complexity. On the other hand, you have fast and faster change. This makes for a very, very daunting task for enterprises, developers and operators to actually keep up and move with speed. And this is exactly the central problem that we want to solve for you. We want to empower you to move with speed in the middle off rising complexity and change and do it successfully and with confidence. So with that in mind, we are announcing this week at LAUNCHPAD a big and new concept to take the company forward and take you with us to create value for you. And we call this your cloud everywhere, which empowers you to ship code faster. Dr. Enterprise Container Cloud is a lynch bit off your cloud everywhere. It's a radical and new container platform, which gives you our customers a consistent experience on public clouds and private clouds alike, which enables you to ship code faster on any infrastructure, anywhere with a cohesive cloud fabric that meets your security standards that offers a choice or private and public clouds and offer you a offers you a simple, an extremely easy and powerful to use experience. for developers. All of this is, um, underpinned by kubernetes as the foundation technology we're betting on forward to help you achieve your goals at the same time. Lens kubernetes e. It's also very, very well into the real cloud. Every concept, and it's a second very strong linchpin to take us forward because it creates the developing experience. It supports developers directly on their desktop, enabling them Thio manage communities workloads to test, develop and run communities applications on any infrastructure anywhere. So Doc, Enterprise, Container, Cloud and Lens complement each other perfectly. So I'm very, very excited to share this with you today and opened the conference for you. And with this I want to turn it over to my colleague Adam Parker, who runs product development at Mirandes to share a lot more detail about Doc Enterprise Container Cloud. Why we're excited about it. Why we feel is a radical step forward to you and why we feel it can add so much value to your developers and operators who want to embrace the latest kubernetes technology and the latest container technology on any platform anywhere. I look forward to connecting with you during the conference and we should all the best. Bye bye. >>Thanks, Adrian. My name is Adam Parco, and I am vice president of engineering and product development at Mirant ISS. I'm extremely excited to be here today And to present to you Dr Enterprise Container Cloud Doc Enterprise Container Cloud is a major leap forward. It Turpal charges are platform. It is your cloud everywhere. It has been completely designed and built around helping you to ship code faster. The world is moving incredibly quick. We have seen unpredictable and rapid changes. It is the goal of Docker Enterprise Container Cloud to help navigate this insanity by focusing on speed and efficiency. To do this requires three major pillars choice, simplicity and security. The less time between a line of code being written and that line of code running in production the better. When you decrease that cycle, time developers are more productive, efficient and happy. The code is higher, quality contains less defects, and when bugs are found are fixed quicker and more easily. And in turn, your customers get more value sooner and more often. Increasing speed and improving developer efficiency is paramount. To do this, you need to be able to cycle through coding, running, testing, releasing and monitoring all without friction. We enabled us by offering containers as a service through a consistent, cloudlike experience. Developers can log into Dr Enterprise Container Cloud and, through self service, create a cluster No I T. Tickets. No industry specific experience required. Need a place to run. A workload simply created nothing quicker than that. The clusters air presented consistently no matter where they're created, integrate your pipelines and start deploying secure images everywhere. Instantly. You can't have cloud speed if you start to get bogged down by managing, so we offer fully automated lifecycle management. Let's jump into the details of how we achieve cloud speed. The first is cloud choice developers. Operators add mons users they all want. In fact, mandate choice choice is extremely important in efficiency, speed and ultimately the value created. You have cloud choice throughout the full stack. Choice allows developers and operators to use the tooling and services their most familiar with most efficient with or perhaps simply allows them to integrate with any existing tools and services already in use, allowing them to integrate and move on. Doc Enterprise Container Cloud isn't constructive. It's open and flexible. The next important choice we offer is an orchestration. We hear time and time again from our customers that they love swarm. That's simply enough for the majority of their applications. And that just works that they have skills and knowledge to effectively use it. They don't need to be or find coop experts to get immediate value, so we will absolutely continue to offer this choice and orchestration. Our existing customers could rest assure their workloads will continue to run. Great as always. On the other hand, we can't ignore the popularity that growth, the enthusiasm and community ecosystem that has exploded with communities. So we will also be including a fully conforming, tested and certified kubernetes going down the stock. You can't have choice or speed without your choice and operating system. This ties back to developer efficiency. We want developers to be able to leverage their operating system of choice, were initially supporting full stack lifecycle management for a bun, too, with other operating systems like red hat to follow shortly. Lastly, all the way down at the bottom of stack is your choice in infrastructure choice and infrastructure is in our DNA. We have always promoted no locking and flexibility to run where needed initially were supporting open stock AWS and full life cycle management of bare metal. We also have a road map for VM Ware and other public cloud providers. We know there's no single solution for the unique and complex requirements our customers have. This is why we're doubling down on being the most open platform. We want you to truly make this your cloud. If done wrong, all this choice at speed could have been extremely complex. This is where cloud simplification comes in. We offer a simple and consistent as a service cloud experience, from installation to day to ops clusters Air created using a single pane of glass no matter where they're created, giving a simple and consistent interface. Clusters can be created on bare metal and private data centers and, of course, on public cloud applications will always have specific operating requirements. For example, data protection, security, cost efficiency edge or leveraging specific services on public infrastructure. Being able to create a cluster on the infrastructure that makes the most sense while maintaining a consistent experience is incredibly powerful to developers and operators. This helps developers move quick by being able to leverage the infra and services of their choice and operators by leveraging, available, compute with the most efficient and for available. Now that we have users self creating clusters, we need centralized management to support this increase in scale. Doc Enterprise Container cloud use is the single pane of glass for observe ability and management of all your clusters. We have day to ops covered to keep things simple and new. Moving fast from this single pane of glass, you can manage the full stack lifecycle of your clusters from the infra up, including Dr Enterprise, as well as the fully automated deployment and management of all components deployed through it. What I'm most excited about is Doc Enterprise Container Cloud as a service. What do I mean by as a service doctor? Enterprise continue. Cloud is fully self managed and continuously delivered. It is always up to date, always security patched, always available new features and capabilities pushed often and directly to you truly as a service experience anywhere you want, it run. Security is of utmost importance to Miranda's and our customers. Security can't be an afterthought, and it can't be added later with Doctor and a price continued cloud, we're maintaining our leadership and security. We're doing this by leveraging the proven security and Dr Enterprise. Dr. Enterprise has the best and the most complete security certifications and compliance, such as Stig Oscar, How and Phipps 1 $40 to thes security certifications allows us to run in the world's most secure locations. We are proud and honored to have some of the most security conscious customers in the world from all industries into. She's like insurance, finance, health care as well as public, federal and government agencies. With Dr Enterprise Container Cloud. We put security as our top concern, but importantly, we do it with speed. You can't move fast with security in the way so they solve this. We've added what we're calling invisible security security enabled by default and configured for you as part of the platform. Dr Price Container Cloud is multi tenant with granular are back throughout. In conjunction with Doc Enterprise, Docker Trusted Registry and Dr Content Trust. We have a complete end to end secured software supply chain Onley run the images that have gone through the appropriate channels that you have authorized to run on the most secure container engine in the >>industry. >>Lastly, I want to quickly touch on scale. Today. Cluster sprawl is a very real thing. There are test clusters, staging clusters and, of course, production clusters. There's also different availability zones, different business units and so on. There's clusters everywhere. These clusters are also running all over the place. We have customers running Doc Enterprise on premise there, embracing public cloud and not just one cloud that might also have some bare metal. So cloud sprawl is also a very real thing. All these clusters on all these clouds is a maintenance and observe ability. Nightmare. This is a huge friction point to scaling Dr Price. Container Cloud solves these issues, lets you scale quicker and more easily. Little recap. What's new. We've added multi cluster management. Deploy and attach all your clusters wherever they are. Multi cloud, including public private and bare metal. Deploy your clusters to any infra self service cluster creation. No more I T. Tickets to get resources. Incredible speed. Automated Full stack Lifecycle management, including Dr Enterprise Container, cloud itself as a service from the in for up centralized observe ability with a single pane of glass for your clusters, their health, your APs and most importantly to our existing doc enterprise customers. You can, of course, add your existing D clusters to Dr Enterprise Container Cloud and start leveraging the many benefits it offers immediately. So that's it. Thank you so much for attending today's keynote. This was very much just a high level introduction to our exciting release. There is so much more to learn about and try out. I hope you are as excited as I am to get started today with Doc Enterprise. Continue, Cloud, please attend the tutorial tracks up Next is Miska, with the world's most popular Kubernetes E Lens. Thanks again, and I hope you enjoy the rest of our conference.
SUMMARY :
look forward to connecting with you during the conference and we should all the best. We want you to truly make this your cloud. This is a huge friction point to scaling Dr Price.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Adrian | PERSON | 0.99+ |
Bosch | ORGANIZATION | 0.99+ |
Adam Parco | PERSON | 0.99+ |
Adam Parker | PERSON | 0.99+ |
GlaxoSmithKline | ORGANIZATION | 0.99+ |
first | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Visa | ORGANIZATION | 0.99+ |
Adam | PERSON | 0.99+ |
standard and Poor's | ORGANIZATION | 0.99+ |
second | QUANTITY | 0.99+ |
Mirant | ORGANIZATION | 0.99+ |
first release | QUANTITY | 0.99+ |
600 new users | QUANTITY | 0.99+ |
last year | DATE | 0.99+ |
Mirandes | ORGANIZATION | 0.98+ |
three | QUANTITY | 0.98+ |
two things | QUANTITY | 0.98+ |
LAUNCHPAD | ORGANIZATION | 0.98+ |
Mawr | ORGANIZATION | 0.98+ |
Dr Enterprise | ORGANIZATION | 0.98+ |
today | DATE | 0.98+ |
this week | DATE | 0.97+ |
Today | DATE | 0.97+ |
Mirant ISS | ORGANIZATION | 0.97+ |
Doc Enterprise | ORGANIZATION | 0.97+ |
third vector | QUANTITY | 0.97+ |
third priority | QUANTITY | 0.97+ |
first one | QUANTITY | 0.97+ |
two important acquisitions | QUANTITY | 0.97+ |
Windows | TITLE | 0.96+ |
two core products | QUANTITY | 0.96+ |
Axa X l | ORGANIZATION | 0.96+ |
one cloud | QUANTITY | 0.96+ |
Miranda | ORGANIZATION | 0.96+ |
three things | QUANTITY | 0.96+ |
one side | QUANTITY | 0.96+ |
mawr | ORGANIZATION | 0.96+ |
Apple A TNT | ORGANIZATION | 0.95+ |
Miska | PERSON | 0.94+ |
single pane | QUANTITY | 0.93+ |
single solution | QUANTITY | 0.92+ |
Doc | ORGANIZATION | 0.91+ |
Dr. Enterprise | ORGANIZATION | 0.9+ |
past year | DATE | 0.9+ |
How and Phipps | ORGANIZATION | 0.89+ |
past year | DATE | 0.89+ |
Lens | ORGANIZATION | 0.88+ |
this morning | DATE | 0.87+ |
Container | ORGANIZATION | 0.87+ |
earlier this year | DATE | 0.85+ |
Doc Enterprise 3.1 | TITLE | 0.85+ |
Dr Content Trust | ORGANIZATION | 0.84+ |
Doc Enterprise | TITLE | 0.84+ |
Stig Oscar | ORGANIZATION | 0.84+ |
Docker Enterprise Container Cloud | TITLE | 0.83+ |
Dr Price | ORGANIZATION | 0.82+ |
soc T general | ORGANIZATION | 0.82+ |
Container Cloud | ORGANIZATION | 0.81+ |
Doc Enterprise Container Cloud | TITLE | 0.81+ |
Enterprise | ORGANIZATION | 0.79+ |
three major pillars | QUANTITY | 0.78+ |
Enterprise Container Cloud Doc Enterprise Container Cloud | TITLE | 0.78+ |
Container | TITLE | 0.77+ |
one hand | QUANTITY | 0.76+ |
months | DATE | 0.75+ |
$40 | QUANTITY | 0.74+ |
RT | ORGANIZATION | 0.73+ |
Dr Price Container | ORGANIZATION | 0.72+ |
Dio | ORGANIZATION | 0.7+ |
Sele | PERSON | 0.7+ |
Miska's keynote v3 ghosting fix
>>Hello. I miss Caribbean, the principal Off Lens Open Source project and senior director off Engineering at Mirandes. I'm excited to be here today at launch back 2020 Virtual conference. I will be your guide, helping you to navigate the rough waters off opportunities and containers and show you the way how to take full advantage off this great new technology with help off lens. The Coburn Edie's idea. It's happening all around us. Containers and Coburn ET is everywhere. Every day, hundreds of thousands off people create new clusters. Develop containerized application on they deploy those applications on top of Cuban Edie's. It has become the golden standard for container orchestration. How did we get here? The industry has been very creative and innovative in ways how to burn it is has been marketed with the help off develops movement, empowering individual development teams leveraging 12 factor model on infrastructure. As a code principles, we have created the need for a system that is able to obstruct everything. That's one a single system to rule them all. Cooper needs has become this system. It has become the operating system for cloud. But hey, people say Coburn Ages is difficult and complex. Absolutely many people on organizations are struggling to adopt kubernetes at scale terrorist complexity on complexity on top of complexity. On top off this, you might need to unlearn some of the things you have used to do in the past. Having had chance to speak with hundreds off, Cooper needs users on operators, from beginners to ninja level hackers. I feel Coburn Edie's is not too difficult or complex. People will get this perception on Lee when they are using primitive or were limited tools for job, or if they have failed to address the needs off all different stakeholders. By using proper quality tools and products, we can truly harness the power of communities on radically improved the speed of business To get there. In my mind, we have deserved at least two important stakeholders. First, mhm. We have hopes and idea means who want to use system for centralized kubernetes cluster creation operations and management in a listen take care a lot about underlying infrastructure, security and conformance. The industry has been serving teas people very well. He has an amazing products for this segment. Dr Enterprise Container Cloud. It's a great example off such a product. Secondly, we have developers who are, in fact the consumers off. The clusters provided by the ops and I T at means they are the people who actually access the clusters on daily basis. Take deploy, run, managed, debug, inspect on observed the workloads running on top of communities. The availability and quality off tools and products for this segment has been lacking. See, very luckily, that's not the case anymore. And that's the focus off my talk today to take away. I want you to have from this simple unless we have quality tools and products for both off these important stakeholders, we might not get all the benefits we were looking for. Docker Enterprise Container Cloud. We'll get you on top and when combined with the product, I'm about to talk. Next. We'll take you where you wanna be. I'm so excited about this lens. The Cooper needs I D. I. D stands for integrated development environment. We could call it in the credit operations environment as well, but let's stick with I D for a little bit longer. No, If you would be doing non virtual conference, I would be as asking how many off you have heard or actually tried using less >>before. It's okay, Let's make make it interactive. We can still do it all right. I'm probably I would see something to 20% of people raising their hands. To be honest, I'm amazed how many people have started using lens already. It's been out on Lee for just six months or so. Lens combines all a sense of tools and technologies >>required for streamlining cloud native applicants and development on Day two operations. It's all you need to take control off Coburn. Edie's clusters on workloads running on top, for example, you might have find hard time trying to understand what is really going on in your clusters with lens. You will have complete situational awareness off all your clusters on work clothes, and you will understand what's going on on quickly. Take actions if needed. Lenses designed for developers who need to work with Cooper needs on a daily basis. If you have somebody who is just getting started, lens will lower the barrier of entry because it will let you explore your clusters on workloads very easy. Take action to try out different things on diesel eyes, everything in a way that makes sense on provides full context. If you are very experienced ninja level heavy user, you will get things done fast. In essence, by using lens, you will become more productive on the quality off life is improved a lot lenses. A stand alone desktop application for Mac OS Windows and Linux operating systems. It's free and fully open. Source under Emmett license. If you want to get started, simply download the lens application from Lens website and start adding your clusters. Now you might wonder. How does lend play together with Mirandes >>offering sheep code faster at Mirandes, we want to convert open source innovation in the customer value. We want to be best in the world. At this. We want to increase developer velocity to continuously deliver code faster for public and private clouds. And in order to do that, we want to put capable person in the center. We want to invest in products and technologies that will improve the developer productivity that speed sheep gold faster. To have speed, we got to get right amount off simplicity. Choice on security simplicity does not mean less features. It means amazing usability on developer experience for using complex on feature rich systems Under the hood. Security means invisible security, something that is built into the system from >>beginning on its part of its DNA, something that is automatically applied to the underlying infrastructure and software running on top without need for developers to worry about too much choice. It's include chance. You should be able to choose the parts you want to use, for example, choice of the infrastructure, cloud providers or even host operating system running on your machines. Everything in here comes to life with talker in the price container cloud. Combined with lens, it's the end to end solution for harnessing the power of kubernetes and radically improving the speed of business. >>All right, I hope you got the idea how lens will play together with Mirandes offering on a highly law. Now I'd like to talk more about lens features in detail. Let's kick off with multi cluster management. Unlike multi cluster management systems designed for hopes and ideas, New people peace is the Monte Cluster management from the developers point of view, take a nap. Any number >>of cabernet, these clusters to provide quick and easy way to switch cluster context on access workloads Running on top thes clusters may be the ones provide provided by their hopes and ideas mean people, but they might be clusters running locally, used in some other projects or use for hobby purposes. As an example, the clusters are added simply simply by importing the cube conflict file and selecting the cluster context. Once added, it's fast and easy to switch between clusters. Since the requirement for acting a cluster is just a cube. Conflict file lens works with any any certified Cooper needs distributions where user might have obtained to keep conflict. Five. For example, Documented price Container Cloud. You see T e. K s G. K. A. K s rancher opens it. Minnick YouTube many, many other flavors off uber Nitties They all work straight out off the box. The creating above lens is that you will get one unified I e across all your clusters. >>No matter what's the flavor on. There is absolutely nothing that you need to install in. Cluster is in itself is great because most off the developers we're talking about in here do not have sufficient right to install anything like this in their clusters. Since we're now talking about access control, let's discuss how the role based access control is taken in account with lens. It's all about uber needs built in role based access control. As you know, clusters may be configured to use any supported identity providers, since lens will authenticate uses the Cooper needs with Cuba conflict file Cooper needs are back is automatically enforced. This is also reflected on the user interface user. Will Onley see those resources they are allowed to access? Lens do not need admin level privileges, service accounts or any other solution that would by bus. The Cooper needs are back. Next. We have a smart terminal less has a built in smart terminal. It comes with bundled common line tools such as cube cattle on help. It's different from your native terminal because the smart terminal will always have cube cattle command available on bond. It will automatically >>switch the version off cube cattle to match the currently selected Cooper Needs Cluster a P I. If FBI compatible version is not found, it will be downloaded automatically in the background. In addition to making sure you are always using the right version off cube kuttel the Smart Terminal will automatically assigned the Cube conflict context to match your currently selected co Bernie. This cluster as a summary. When you use lens with building Smart Terminal, you are always using the right version off cube cattle and context. I feel there is still something more I want to share with you. Visualizations lenses Very diesel on There is a lot of detail in the user experience. One of the great features in Lens is that building in the creation with Prometheus to visualize everything. As you might know, people working on the ups and i d at me inside of things have learned to write complex Primedia Square ease. Most likely, they have created beautiful death sport to look at data. Looking at the cluster's from the bird's eye perspective. If you are a developer, you are interested in your own stuff. Bird side perspective might be nice, but it doesn't help you to debug and trouble. Suit your own application. You don't necessarily have access to or want to learn Prometheus to write your own queries on out of context that sports. That is why lens will provide automatically civilization for all supportive resource types including the aggregated Use it, >>David Little person. Or, to be honest, ops on Idea Means to will get all the data they need, always in the right context. The basic metrics include CPU memory on disk with total capacity actual use. It requests on limits. The unrest metrics include bytes sent success, failure on request and response to race. Both statistics also include network bytes sent and received. Persistent pulling. Unclaimed metrics include disk usage and capacity. Wow, that was a lot on. To be honest, we are just barely scratching the surface off the available features. Let's move on and talk about lens from the community on open source project perspective. We'll start with statistic, not because I like statistics in particular, but because this project has some mind blowing stats to share. Let's remind ourselves that lens was made open source just a half a year ago. Since then, over 600,000 downloads over 50,000 users over >>8000 star gazers on get top. The users come from all around the world. It's one off the fastest training open source projects on git hub and definitely in Cuba needs ecosystem. It's the number one e or u I or whatever you wanna call it for Cuban, it is on. If you are not using it yet, you're probably missing out some something great. What's coming on next? We are working hard every day to make lens better. Our focus as a leader in this open source project is to remain vendor Notre Look Ways for collaboration with other vendors in the cloud Native technology ecosystem on focus on making features that directing most value for our users. Against this background, the near future roadmap includes exciting features like extensive a P I. While the building features off, lens might feel great. It's just the beginning. Lens extensive a p I that is going to be a new feature released as part off Lens 4.0, we'll let you at custom visualizations on functionality to support your preferred development. Work flaws. The Extensions AP I will provide options for extensive creators to but directly into the lens You I we are already working with the number off cloud Native technology ecosystem vendors to get their technology is deeply integrated on therefore more accessible true lens, for example, on extension for a container >>image scanning technology vendor, I might add a warning icon next to a port or a deployment where vulnerable image is detected in a decent. This extension might provide more details about this vulnerability when the port or deployment is clicked. This is just a simple example, but I hope you get the idea on Really, this is just beginning. We want to >>bring entire Coburn Edie's ecosystem together in a listen to extensions. A p I. We will work on features to enhance Cooper needs Developer were close, both locally on remote, enable teamwork and naturally improve the usability on fixed box reported by our users. There are so many great things coming. It's impossible to list everything in here. If you are interested, please take a look at the epics listed on our guitar free ball. Once again, if you're not using lens already, you're probably missing out on something great. Download and get started today. For the most amazing entrant experience, check out the Docker Enterprise Container Cloud as well. I wish you all a great time with Coburn. Edie's I'm looking forward to meet you all in person someday. Take care. Bye bye
SUMMARY :
The clusters provided by the ops and I T at means It's been out on Lee for just six months entry because it will let you explore your clusters on workloads security, something that is built into the system from You should be able to choose the parts you want to use, New people peace is the Monte Cluster management from the developers you will get one unified I e across all your clusters. Cluster is in itself is great because most off the developers addition to making sure you are always using the right version off cube kuttel the Let's move on and talk about lens from the community on functionality to support your preferred development. is just a simple example, but I hope you get the idea on Really, Edie's I'm looking forward to meet you all
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
20% | QUANTITY | 0.99+ |
six months | QUANTITY | 0.99+ |
Cuba | LOCATION | 0.99+ |
David Little | PERSON | 0.99+ |
First | QUANTITY | 0.99+ |
Prometheus | TITLE | 0.99+ |
over 600,000 downloads | QUANTITY | 0.99+ |
Linux | TITLE | 0.98+ |
Miska | PERSON | 0.98+ |
today | DATE | 0.98+ |
Mirandes | ORGANIZATION | 0.98+ |
Edie | PERSON | 0.98+ |
over 50,000 users | QUANTITY | 0.98+ |
both | QUANTITY | 0.98+ |
Cooper | PERSON | 0.97+ |
uber | ORGANIZATION | 0.97+ |
Secondly | QUANTITY | 0.97+ |
Lee | PERSON | 0.97+ |
One | QUANTITY | 0.97+ |
hundreds | QUANTITY | 0.96+ |
Coburn Edie | ORGANIZATION | 0.96+ |
YouTube | ORGANIZATION | 0.96+ |
FBI | ORGANIZATION | 0.96+ |
Five | QUANTITY | 0.95+ |
12 factor | QUANTITY | 0.95+ |
Mac OS Windows | TITLE | 0.95+ |
half a year ago | DATE | 0.94+ |
Day two | QUANTITY | 0.92+ |
Lens 4.0 | OTHER | 0.9+ |
Edie | ORGANIZATION | 0.9+ |
Both statistics | QUANTITY | 0.89+ |
Lens | ORGANIZATION | 0.87+ |
Emmett | TITLE | 0.85+ |
Coburn ET | ORGANIZATION | 0.85+ |
single system | QUANTITY | 0.85+ |
Cooper | ORGANIZATION | 0.84+ |
Coburn | ORGANIZATION | 0.84+ |
hundreds of thousands off | QUANTITY | 0.84+ |
Notre | ORGANIZATION | 0.82+ |
>8000 star | QUANTITY | 0.77+ |
Cuban | OTHER | 0.75+ |
Coburn | PERSON | 0.74+ |
least two important stakeholders | QUANTITY | 0.72+ |
Lee | ORGANIZATION | 0.72+ |
Primedia Square | ORGANIZATION | 0.72+ |
2020 Virtual conference | EVENT | 0.69+ |
ways | QUANTITY | 0.69+ |
Caribbean | LOCATION | 0.69+ |
Coburn Ages | ORGANIZATION | 0.68+ |
Minnick | ORGANIZATION | 0.67+ |
one | QUANTITY | 0.66+ |
Enterprise Container | ORGANIZATION | 0.62+ |
Docker | ORGANIZATION | 0.6+ |
Bernie | PERSON | 0.59+ |
lend | ORGANIZATION | 0.53+ |
Docker Enterprise Container Cloud | TITLE | 0.53+ |
fix | TITLE | 0.52+ |
Cooper | COMMERCIAL_ITEM | 0.5+ |
Onley | PERSON | 0.45+ |
G. | PERSON | 0.44+ |
Cluster | ORGANIZATION | 0.27+ |
ON DEMAND SWARM ON K8S FINAL NEEDS CTA SLIDE
>>welcome to the session. Long live swarm with containers and kubernetes everywhere we have this increasing cloud complexity at the same time that we're facing economic uncertainty and, of course, to navigate this. For most companies, it's a matter of focusing on speed and on shipping and iterating their code faster. Now. For many, Marantz is customers. That means using docker swarm rather than kubernetes to handle container orchestration. We really believe that the best way to increase your speed to production is choice, simplicity and security. So we wanted to bring you a couple of experts to talk about the state of swarm and Docker enterprise and how you can make best use of both of you. So let's get to it. Well, good afternoon or good morning, depending on where you are on and welcome to today's session. Long live swarm. I am Nick Chase. I'm head of content here at Mantis and I would like to introduce you to our two Panelists today eight of Manzini. Why don't you introduce yourself? >>I am a van CNI. I'm a solutions architect here at Moran Tous on work primarily with Docker Enterprise System. I have a long history of working with support team. Um, at what used to be Ah Docker Enterprise, part of Docker Inc. >>Yeah, Okay. Great. And Don Power. >>I, um Yeah, I'm Don Power on the docker. Captain Docker, community leader. Right now I run our Dev Ops team for Citizens Bank out of Nashville, Tennessee, and happy to be here. >>All right, Excellent. So All right, so thank you both for coming. Now, before we say anything else, I want to go ahead and kind of name the elephant in the room. There's been a lot of talk about the >>future. Yeah, that's right. Um, swarm as it stands right now, um, we have, ah, very vested interest in keeping our customers on who want to continue using swarm, functional and keeping swarm a viable alternative or complement to kubernetes. However you see the orchestration war playing out as it were. >>Okay? It's hardly a war at this point, but they do work together, and so that's >>absolutely Yeah, I I definitely consider them more of like, complimentary services, um, using the right tool for the job. Sort of sense. They both have different design goals when they were originally created and set out so I definitely don't see it as a completely one or the other kind of decision and that they could both be used in the same environment and similar clusters to run whatever workload that you have. >>Excellent. And we'll get into the details of all that as we go along. So that's terrific. So I have not really been involved in in the sort of swarm area. So set the stage for us where we kind of start out with all of this. Don I know that you were involved and so guys said, set the stage for us. >>Sure, Um I mean so I've been a heavy user of swarm in my past few roles. Professionally, we've been running containers in production with Swarm for coming up on about four years. Now, Um, in our case, we you know, we looked at what was available at the time, and of course you had. Kubernetes is your biggest contender out there, but like I just mentioned, the one of the things that really led us to swarm is it's design goals were very different than kubernetes. So Kubernetes tries to have an answer for absolutely every scenario where swarm tries to have an answer for, like, the 80% of problems or challenges will say that you might come across 80% of the workloads. Um, I had a better way of saying that, but I think I got my point across >>E Yeah, I think I think you hit the nail on the head. Um, Kubernetes in particular with the way that kubernetes itself is an a P I I believe that kubernetes was, um, you know, written as a toolkit. It wasn't really intended to be used by end users directly. It was really a way to build platforms that run containers. And because it's this really, really extensible ap I you can extend it to manage all sorts of resource is swarm doesn't have that X sensibility aspect, but what it was designed to do, it does very, very well and very easily in a very, very simple sort of way. Um, it's highly opinionated about the way that you should use the product, but it works very effectively. It's very easy to use. It's very low. Um, not low effort, but low. Ah, low barrier to entry. >>Yes. Yes. Absolutely. I was gonna touch on the same thing. It's very easy for someone to come in. Pick up swarm. You know they don't They don't have to know anything about the orchestrator on day one. Most people that are getting into this space are very familiar with Docker. Compose um, and entering from Docker compose into swarm is changing one command that you would run on the command line. >>Yeah, very, very trivial to if you are already used to building docker files using composed, organize your deployment into stacks of related components. It's trivial to turn on swarm mode and then deploy your container set to a cluster. >>Well, excellent. So answer this question for me. Is the swarm of today the same as the swarm of, you know, the original swarm. So, like when swim first started is that the same is what we have now >>it's kind of ah, complicated story with the storm project because it's changed names and forms a few times. Originally in is really somewhere around 2014 in the first version, and it was a component that you really had to configure and set up separately from Docker Ah, the way that it was structured. Ah, you would just have docker installed on a number of servers are machines in your cluster. And then you would organize them into a swarm by bringing your own database and some of the tooling to get those nodes talking to each other and to organize your containers across all of your docker engines. Ah, few years later, the swarm project was retooled and baked into the docker engine. And, um, this is where we sort of get the name change from. So originally it was a feature that we called swarm. Ah. Then the Swarm Kit project was released on Get Hub and baked directly into the engine, where they renamed it as swarm mode. Because now it is a motile option that you just turn on as a button in the docker engine and because it's already there the, um, the tuning knobs that you haven't swarm kit with regard to how what my time outs are and some of these other sort of performance settings there locked there, they're there. It's part of the opinionated set of components that builds up the docker engine is that we bring in the Swarm Kit project with a certain set of defaults and settings. And that is how it operates in today's version of Docker engine. >>Uh, okay for that, that makes sense. That makes sense. So ah, so don, I know you have pretty strong feelings about this topic, but it is swarm still viable in a world that's sort of increasingly dominated by Kubernetes. >>Absolutely. And you were right. I'm very passionate about this topic where I work. We're we're doing almost all of our production work lives on swarm we only have out of Ah, we've got something like 600 different services between three and 4000 containers. At any given point in time. Out of all of those projects, all of those services we've only run into two or three that don't kind of fit into the opinionated model of swarm. So we are running those on KUBERNETES in the same cluster using Moranis is Docker enterprise offering. But, um, no, that's a very, very small percentage of services that we didn't have an answer for in swarm with one. The one case that really gets us just about every time is scaling state full services. But you're gonna have very few staple services in most environments for things like micro service architecture, which is predominantly what we build out. Swarm is perfect. It's simple. It's easy to use you, don't you? Don't end up going for miles of yamma files trying to figure out the one setting that you didn't get exactly right? Um yeah, the other Thea the other big piece of it that way really led us to adopting it so heavily in the beginning is, you know, the overlay network. So your networks don't have to span the whole cluster like they do with kubernetes. So we could we could set up a network isolation between service A and service B, just by use using the built in overlay networks. That was a huge component that, like I said, let us Teoh adopting it so heavily when we first got started. >>Excellent. You look like you're about to say something in a >>Yeah, I think that speaks to the design goals for each piece of software. On the way that I've heard this described before is with regard to the networking piece the ah, the docker networking under the hood, um, feels like it was written by a network engineer. The way that the docker engine overlay networks communicate uses ah, VX lan under the hood, which creates pseudo V lands for your containers. And if two containers aren't on the same Dylan, there's no way they can communicate with each other as opposed to the design of kubernetes networking, which is really left to the C and I implementation but still has the design philosophy of one big, flat sub net where every I p could reach every other i p and you control what is allowed to access, what by policy. So it's more of an application focused Ah design. Whereas in Docker swarm on the overlay networking side, it's really of a network engineering sort of focus. Right? >>Okay, got it. Well, so now how does all this fit in with Docker enterprise now? So I understand there's been some changes on how swarm is handled within Docker Enterprise. Coming with this new release, >>Docker s O swarm Inside Docker Enterprise is represented as both the swarm classic legacy system that we shift way back in 2014 on and then also the swarm mode that is curly used in the docker engine. Um, the Swarm Classic back end gives us legacy support for being able to run unmanaged plane containers onto a cluster. If you were to take Docker ce right now, you would find that you wouldn't be able to just do a very basic docker run against a whole cluster of machines. You can create services using the swarms services, a p I but, um, that that legacy plane container support is something that you have to set up external swarm in order to provide. So right now, the architecture of Docker Enterprise UCP is based on some of that legacy code from about five or six years ago. Okay. Ah, that gives us ability to deploy plane containers for use cases that require it as well as swarm services for those kinds of workloads that might be better served by the built in load balancing and h A and scaling features that swarm provides. >>Okay, so now I know that at one point kubernetes was deployed within Docker Enterprise as you create a swarm cluster and then deploy kubernetes on top of swarm. >>Correct? That is how the current architecture works. >>Okay. All right. And then, um what is what is where we're going with this like, Are we supposed to? Are we going to running Swarm on top of kubernetes? What's >>the the design goals for the future of swarm within branches? Stocker Enterprise are that we will start the employing Ah, like kubernetes cluster features as the base and a swarm kit on top of kubernetes. So it is like you mentioned just a reversal of the roles. I think we're finding that, um, the ability to extend kubernetes a p I to manage resource is is valuable at an infrastructure and platform level in a way that we can't do with swarm. We still want to be able to run swarm workloads. So we're going to keep the swarm kit code the swarm kit orchestration features to run swarm services as a part of the platform to keep the >>got it. Okay, so, uh, if I'm a developer and I want to run swarm, but my company's running kubernetes what? What are my one of my options there? Well, I think >>eight touched on it pretty well already where you know, it depends on your design goals, and you know, one of the other things that's come up a few times is Thea. The level of entry for for swarm is much, much simpler than kubernetes. So I mean, it's it's kind of hard to introduce anything new. So I mean, a company, a company that's got most of their stuff in kubernetes and production is gonna have a hard time maybe looking at a swarm. I mean, this is gonna be, you know, higher, higher up, not the boots on the ground. But, um, you know, the the upper management, that's at some point, you have to pay for all their support, all of it. What we did in our approach. Because there was one team already using kubernetes. We went ahead and stood up a small cluster ah, small swarm cluster and taught the developers how to use it and how to deploy code to it. And they loved it. They thought it was super simple. A time went on, the other teams took notice and saw how fast these guys were getting getting code deployed, getting services up, getting things usable, and they would look over at what the innovation team was doing and say, Hey, I I want to do that to, uh, you know, so there's there's a bunch of different approaches. That's the approach we took and it worked out very well. It looks like you wanted to say something too. >>Yeah, I think that if you if you're if you're having to make this kind of decision, there isn't There isn't a wrong choice. Ah, it's never a swarm of its role and your organization, right? Right. If you're if you're an individual and you're using docker on your workstation on your laptop but your organization wants to standardize on kubernetes there, there are still some two rules that Mike over Ah, pose. And he's manifest if you need to deploy. Coop resource is, um if you are running Docker Enterprise Swarm kit code will still be there. And you can run swarm services as regular swarm workloads on that component. So I I don't want to I don't want people to think that they're going to be like, locked into one or the other orchestration system. Ah, there the way we want to enable developer choice so that however the developer wants to do their work, they can get it done. Um Docker desktop. Ah, ships with that kubernetes distribution bundled in it. So if you're using a Mac or Windows and that's your development, uh, system, you can run docker debt, turn on your mode and run the kubernetes bits. So you have the choices. You have the tools to deploy to either system. >>And that's one of the things that we were super excited about when they introduced Q. Burnett ease into the Docker Enterprise offering. So we were able to run both, so we didn't have to have that. I don't want to call it a battle or argument, but we didn't have to make anybody choose one or the other. We, you know, we gave them both options just by having Docker enterprise so >>excellent. So speaking of having both options, let's just say for developers who need to make a decision while should I go swarm, or should I go kubernetes when it sort of some of the things that they should think about? >>So I think that certain certain elements of, um, certain elements of containers are going to be agnostic right now. So the the the designing a docker file and building a container image, you're going to need to know that skill for either system that you choose to operate on. Ah, the swarm value. Some of the storm advantage comes in that you don't have to know anything beyond that. So you don't have to learn a whole new A p I a whole new domain specific language using Gamel to define your deployment. Um, chances are that if you've been using docker for any length of time, you probably have a whole stack of composed files that are related to things that you've worked on. And, um, again, the barrier to entry to getting those running on swarm is very low. You just turn it on docker stack, deploy, and you're good to go. So I think that if you're trying to make that choice, if you I have a use case that doesn't require you to manage new resource is if you don't need the Extensible researchers part, Ah, swarm is a great great, great viable option. >>Absolutely. Yeah, the the recommendation I've always made to people that are just getting started is start with swarm and then move into kubernetes and going through the the two of them, you're gonna figure out what fits your design principles. What fits your goals. Which one? You know which ones gonna work best for you. And there's no harm in choosing one or the other using both each one of you know, very tailor fit for very various types of use cases. And like I said, kubernetes is great at some things, but for a lot of other stuff, I still want to use swarm and vice versa. So >>on my home lab, for all my personal like services that I run in my, uh, my home network, I used storm, um, for things that I might deploy onto, you know, a bit this environment, a lot of the ones that I'm using right now are mainly tailored for kubernetes eso. I think especially some of the tools that are out there in the open source community as well as in docker Enterprise helped to bridge that gap like there's a translator that can take your compose file, turn it into kubernetes. Yeah, Mel's, um, if if you're trying to decide, like on the business side, should we standardize on former kubernetes? I think like your what? What functionality are you looking at? Out of getting out of your system? If you need things like tight integration into a ah infrastructure vendor such as AWS Azure or VM ware that might have, like plug ins for kubernetes. You're now you're getting into that area where you're managing Resource is of the infrastructure with your orchestration. AP I with kube so things like persistent volumes can talk to your storage device and carve off chunks of storage and assign those two pods if you don't have that need or that use case. Um, you know, KUBERNETES is bringing in a lot of these features that you maybe you're just not taking advantage of. Um, similarly, if you want to take advantage of things like auto scaling to scale horizontally, let's say you have a message queue system and then a number of workers, and you want to start scaling up on your workers. When your CPU hits a certain a metric. That is something that Kubernetes has built right into it. And so, if you want that, I would probably suggest that you look at kubernetes if you don't need that, or if you want to write some of that tooling yourself. Swarm doesn't have an object built into it that will do automatic horizontal scaling based on some kind of metric. So I always consider this decision as a what features are the most I available to you and your business that you need to Yep. >>All right. Excellent. Well, and, ah, fortunately, of course, they're both available on Docker Enterprise. So aren't we lucky? All right, so I am going to wrap this up. I want to thank Don Bauer Docker captain, for coming here and spending some time with us and eight of Manzini. I would like to thank you. I know that the the, uh, circumstances are less than ideal here for your recording today, but we appreciate you joining us. Um and ah, both of you. Thank you very much. And I want to invite all of you. First of all, thank you for joining us. We know your time is valuable and I want to invite you all Teoh to take a look at Docker Enterprise. Ah, follow the link that's on your screen and we'll see you in the next session. Thank you all so much. Thank you. >>Thank you, Nick.
SUMMARY :
So we wanted to bring you a couple of experts to talk about the state of swarm I have a long history of working with support Tennessee, and happy to be here. kind of name the elephant in the room. However you see the orchestration to run whatever workload that you have. Don I know that you were involved Um, in our case, we you know, we looked at what was Um, it's highly opinionated about the way that you should use is changing one command that you would run on the command line. Yeah, very, very trivial to if you are already used to building docker of, you know, the original swarm. in the first version, and it was a component that you really had to configure and set up separately So ah, so don, I know you have pretty strong to figure out the one setting that you didn't get exactly right? You look like you're about to say something in a On the way that I've heard this described before is with regard to the networking piece Well, so now how does all this fit in with Docker you have to set up external swarm in order to provide. was deployed within Docker Enterprise as you create a swarm cluster That is how the current architecture works. is what is where we're going with this like, Are we supposed to? a part of the platform to keep the I think I mean, this is gonna be, you know, higher, So you have the choices. And that's one of the things that we were super excited about when they introduced Q. So speaking of having both options, let's just say Some of the storm advantage comes in that you don't have to know anything beyond the two of them, you're gonna figure out what fits your design principles. available to you and your business that you need to Yep. I know that the the, uh, circumstances are less than
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Nick Chase | PERSON | 0.99+ |
80% | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
2014 | DATE | 0.99+ |
Citizens Bank | ORGANIZATION | 0.99+ |
three | QUANTITY | 0.99+ |
Marantz | ORGANIZATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
both | QUANTITY | 0.99+ |
two pods | QUANTITY | 0.99+ |
Mantis | ORGANIZATION | 0.99+ |
first version | QUANTITY | 0.99+ |
600 different services | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
Ah Docker Enterprise | ORGANIZATION | 0.99+ |
both options | QUANTITY | 0.99+ |
each piece | QUANTITY | 0.99+ |
Docker Inc. | ORGANIZATION | 0.99+ |
few years later | DATE | 0.99+ |
one case | QUANTITY | 0.99+ |
4000 containers | QUANTITY | 0.98+ |
Docker Enterprise System | ORGANIZATION | 0.98+ |
Q. Burnett | PERSON | 0.98+ |
one team | QUANTITY | 0.98+ |
Manzini | ORGANIZATION | 0.98+ |
Don Power | PERSON | 0.98+ |
Docker Enterprise | ORGANIZATION | 0.98+ |
Docker | TITLE | 0.98+ |
First | QUANTITY | 0.98+ |
Docker Enterprise | TITLE | 0.97+ |
one | QUANTITY | 0.97+ |
Moranis | ORGANIZATION | 0.97+ |
about four years | QUANTITY | 0.97+ |
two Panelists | QUANTITY | 0.97+ |
Nick | PERSON | 0.97+ |
Stocker Enterprise | ORGANIZATION | 0.97+ |
six years ago | DATE | 0.96+ |
Dylan | PERSON | 0.95+ |
Moran Tous | ORGANIZATION | 0.95+ |
Don Bauer | PERSON | 0.95+ |
two containers | QUANTITY | 0.95+ |
two rules | QUANTITY | 0.94+ |
Windows | TITLE | 0.93+ |
Nashville, Tennessee | LOCATION | 0.93+ |
one command | QUANTITY | 0.93+ |
first | QUANTITY | 0.92+ |
Kubernetes | ORGANIZATION | 0.92+ |
eight | QUANTITY | 0.91+ |
Docker enterprise | TITLE | 0.89+ |
KUBERNETES | ORGANIZATION | 0.88+ |
Gamel | TITLE | 0.88+ |
day one | QUANTITY | 0.87+ |
one big, flat sub net | QUANTITY | 0.86+ |
Mac | COMMERCIAL_ITEM | 0.83+ |
service A | OTHER | 0.82+ |
Docker enterprise | TITLE | 0.81+ |
miles | QUANTITY | 0.8+ |
each one | QUANTITY | 0.8+ |
Mike ov | PERSON | 0.78+ |
docker | ORGANIZATION | 0.78+ |
service B | OTHER | 0.77+ |
Azure | TITLE | 0.75+ |
ON DEMAND SPEED K8S DEV OPS SECURE SUPPLY CHAIN
>> In this session, we will be reviewing the power and benefits of implementing a secure software supply chain and how we can gain a cloud like experience with the flexibility, speed and security of modern software delivering. Hi, I'm Matt Bentley and I run our technical pre-sales team here at Mirantis. I spent the last six years working with customers on their containerization journey. One thing almost every one of my customers has focused on is how they can leverage the speed and agility benefits of containerizing their applications while continuing to apply the same security controls. One of the most important things to remember is that we are all doing this for one reason and that is for our applications. So now let's take a look at how we can provide flexibility to all layers of the stack from the infrastructure on up to the application layer. When building a secure supply chain for container focused platforms, I generally see two different mindsets in terms of where their responsibilities lie between the developers of the applications and the operations teams who run the middleware platforms. Most organizations are looking to build a secure, yet robust service that fits their organization's goals around how modern applications are built and delivered. First, let's take a look at the developer or application team approach. This approach falls more of the DevOps philosophy, where a developer and application teams are the owners of their applications from the development through their life cycle, all the way to production. I would refer to this more of a self service model of application delivery and promotion when deployed to a container platform. This is fairly common, organizations where full stack responsibilities have been delegated to the application teams. Even in organizations where full stack ownership doesn't exist, I see the self service application deployment model work very well in lab development or non production environments. This allows teams to experiment with newer technologies, which is one of the most effective benefits of utilizing containers. In other organizations, there is a strong separation between responsibilities for developers and IT operations. This is often due to the complex nature of controlled processes related to the compliance and regulatory needs. Developers are responsible for their application development. This can either include dock at the development layer or be more traditional, throw it over the wall approach to application development. There's also quite a common experience around building a center of excellence with this approach where we can take container platforms and be delivered as a service to other consumers inside of the IT organization. This is fairly prescriptive in the manner of which application teams would consume it. Yeah when examining the two approaches, there are pros and cons to each. Process, controls and compliance are often seen as inhibitors to speed. Self-service creation, starting with the infrastructure layer, leads to inconsistency, security and control concerns, which leads to compliance issues. While self-service is great, without visibility into the utilization and optimization of those environments, it continues the cycles of inefficient resource utilization. And a true infrastructure as a code experience, requires DevOps, related coding skills that teams often have in pockets, but maybe aren't ingrained in the company culture. Luckily for us, there is a middle ground for all of this. Docker Enterprise Container Cloud provide the foundation for the cloud like experience on any infrastructure without all of the out of the box security and controls that our professional services team and your operations teams spend their time designing and implementing. This removes much of the additional work and worry around ensuring that your clusters and experiences are consistent, while maintaining the ideal self service model. No matter if it is a full stack ownership or easing the needs of IT operations. We're also bringing the most natural Kubernetes experience today with Lens to allow for multi-cluster visibility that is both developer and operator friendly. Lens provide immediate feedback for the health of your applications, observability for your clusters, fast context switching between environments and allowing you to choose the best in tool for the task at hand, whether it is the graphic user interface or command line interface driven. Combining the cloud like experience with the efficiencies of a secure supply chain that meet your needs brings you the best of both worlds. You get DevOps speed with all the security and controls to meet the regulations your business lives by. We're talking about more frequent deployments, faster time to recover from application issues and better code quality. As you can see from our clusters we have worked with, we're able to tie these processes back to real cost savings, real efficiency and faster adoption. This all adds up to delivering business value to end users in the overall perceived value. Now let's look and see how we're able to actually build a secure supply chain to help deliver these sorts of initiatives. In our example secure supply chain, where utilizing Docker desktop to help with consistency of developer experience, GitHub for our source control, Jenkins for our CACD tooling, the Docker trusted registry for our secure container registry and the Universal Control Plane to provide us with our secure container runtime with Kubernetes and Swarm, providing a consistent experience, no matter where our clusters are deployed. You work with our teams of developers and operators to design a system that provides a fast, consistent and secure experience. For my developers, that works for any application, Brownfield or Greenfield, Monolith or Microservice. Onboarding teams can be simplified with integrations into enterprise authentication services, calls to GitHub repositories, Jenkins access and jobs, Universal Control Plan and Docker trusted registry teams and organizations, Kubernetes namespace with access control, creating Docker trusted registry namespaces with access control, image scanning and promotion policies. So, now let's take a look and see what it looks like from the CICD process, including Jenkins. So let's start with Docker desktop. From the Docker desktop standpoint, we'll actually be utilizing visual studio code and Docker desktop to provide a consistent developer experience. So no matter if we have one developer or a hundred, we're going to be able to walk through a consistent process through Docker container utilization at the development layer. Once we've made our changes to our code, we'll be able to check those into our source code repository. In this case, we'll be using GitHub. Then when Jenkins picks up, it will check out that code from our source code repository, build our Docker containers, test the application that will build the image, and then it will take the image and push it to our Docker trusted registry. From there, we can scan the image and then make sure it doesn't have any vulnerabilities. Then we can sign them. So once we've signed our images, we've deployed our application to dev, we can actually test our application deployed in our real environment. Jenkins will then test the deployed application. And if all tests show that as good, we'll promote our Docker image to production. So now, let's look at the process, beginning from the developer interaction. First of all, let's take a look at our application as it's deployed today. Here, we can see that we have a change that we want to make on our application. So our marketing team says we need to change containerize NGINX to something more Mirantis branded. So let's take a look at visual studio code, which we'll be using for our ID to change our application. So here's our application. We have our code loaded and we're going to be able to use Docker desktop on our local environment with our Docker desktop plugin for visual studio code, to be able to build our application inside of Docker, without needing to run any command line specific tools. Here with our code, we'll be able to interact with Docker maker changes, see it live and be able to quickly see if our changes actually made the impact that we're expecting our application. So let's find our updated tiles for application and let's go ahead and change that to our Mirantis sized NGINX instead of containerized NGINX. So we'll change it in a title and on the front page of the application. So now that we've saved that changed to our application, we can actually take a look at our code here in VS code. And as simple as this, we can right click on the Docker file and build our application. We give it a name for our Docker image and VS code will take care of the automatic building of our application. So now we have a Docker image that has everything we need in our application inside of that image. So, here we can actually just right click on that image tag that we just created and do run. This will interactively run the container for us. And then once our containers running, we can just right click and open it up in a browser. So here we can see the change to our application as it exists live. So, once we can actually verify that our applications working as expected, we can stop our container. And then from here, we can actually make that change live by pushing it to our source code repository. So here, we're going to go ahead and make a commit message to say that we updated to our Mirantis branding. We will commit that change and then we'll push it to our source code repository. Again, in this case, we're using GitHub to be able to use as our source code repository. So here in VS code, we'll have that pushed here to our source code repository. And then, we'll move on to our next environment, which is Jenkins. Jenkins is going to be picking up those changes for our application and it checked it out from our source code repository. So GitHub notifies Jenkins that there's a change. Checks out the code, builds our Docker image using the Docker file. So we're getting a consistent experience between the local development environment on our desktop and then in Jenkins where we're actually building our application, doing our tests, pushing it into our Docker trusted registry, scanning it and signing our image in our Docker trusted registry and then deploying to our development environment. So let's actually take a look at that development environment as it's been deployed. So, here we can see that our title has been updated on our application, so we can verify that it looks good in development. If we jump back here to Jenkins, we'll see that Jenkins go ahead and runs our integration tests for our development environment. Everything worked as expected, so it promoted that image for our production repository in our Docker trusted registry. We're then, we're going to also sign that image. So we're assigning that yes, we've signed off that has made it through our integration tests and it's deployed to production. So here in Jenkins, we can take a look at our deployed production environment where our application is live in production. We've made a change, automated and very secure manner. So now, let's take a look at our Docker trusted registry, where we can see our name space for our application and our simple NGINX repository. From here, we'll be able to see information about our application image that we've pushed into the registry, such as the image signature, when it was pushed by who and then, we'll also be able to see the results of our image. In this case, we can actually see that there are vulnerabilities for our image and we'll actually take a look at that. Docker trusted registry does binary level scanning. So we get detailed information about our individual image layers. From here, these image layers give us details about where the vulnerabilities were located and what those vulnerabilities actually are. So if we click on the vulnerability, we can see specific information about that vulnerability to give us details around the severity and more information about what exactly is vulnerable inside of our container. One of the challenges that you often face around vulnerabilities is how exactly we would remediate that in a secure supply chain. So let's take a look at that. In the example that we were looking at, the vulnerability is actually in the base layer of our image. In order to pull in a new base layer for our image, we need to actually find the source of that and update it. One of the ways that we can help secure that as a part of the supply chain is to actually take a look at where we get our base layers of our images. Docker hub really provides a great source of content to start from, but opening up Docker hub within your organization, opens up all sorts of security concerns around the origins of that content. Not all images are made equal when it comes to the security of those images. The official images from Docker hub are curated by Docker, open source projects and other vendors. One of the most important use cases is around how you get base images into your environment. It is much easier to consume the base operating system layer images than building your own and also trying to maintain them. Instead of just blindly trusting the content from Docker hub, we can take a set of content that we find useful such as those base image layers or content from vendors and pull that into our own Docker trusted registry, using our mirroring feature. Once the images have been mirrored into a staging area of our Docker trusted registry, we can then scan them to ensure that the images meet our security requirements. And then based off of the scan result, promote the image to a public repository where you can actually sign the images and make them available to our internal consumers to meet their needs. This allows us to provide a set of curated content that we know is secure and controlled within our environment. So from here, we can find our updated Docker image in our Docker trusted registry, where we can see that the vulnerabilities have been resolved. From a developer's point of view, that's about as smooth as the process gets. Now, let's take a look at how we can provide that secure content for our developers in our own Docker trusted registry. So in this case, we're taking a look at our Alpine image that we've mirrored into our Docker trusted registry. Here, we're looking at the staging area where the images get temporarily pulled because we have to pull them in order to actually be able to scan them. So here we set up mirroring and we can quickly turn it on by making it active. And then we can see that our image mirroring, we'll pull our content from Docker hub and then make it available in our Docker trusted registry in an automatic fashion. So from here, we can actually take a look at the promotions to be able to see how exactly we promote our images. In this case, we created a promotion policy within Docker trusted registry that makes it so that content gets promoted to a public repository for internal users to consume based off of the vulnerabilities that are found or not found inside of the Docker image. So our actual users, how they would consume this content is by taking a look at the public to them, official images that we've made available. Here again, looking at our Alpine image, we can take a look at the tags that exist and we can see that we have our content that has been made available. So we've pulled in all sorts of content from Docker hub. In this case, we've even pulled in the multi architecture images, which we can scan due to the binary level nature of our scanning solution. Now let's take a look at Lens. Lens provides capabilities to be able to give developers a quick opinionated view that focuses around how they would want to view, manage and inspect applications deployed to a Kubernetes cluster. Lens integrates natively out of the box with Universal Control Plane clam bundles. So you're automatically generated TLS certificates from UCP, just work. Inside our organization, we want to give our developers the ability to see their applications in a very easy to view manner. So in this case, let's actually filter down to the application that we just employed to our development environment. Here, we can see the pod for application. And when we click on that, we get instant detailed feedback about the components and information that this pod is utilizing. We can also see here in Lens that it gives us the ability to quickly switch contexts between different clusters that we have access to. With that, we also have capabilities to be able to quickly deploy other types of components. One of those is helm charts. Helm charts are a great way to package up applications, especially those that may be more complex to make it much simpler to be able to consume and inversion our applications. In this case, let's take a look at the application that we just built and deployed. In this case, our simple NGINX application has been bundled up as a helm chart and is made available through Lens. Here, we can just click on that description of our application to be able to see more information about the helm chart. So we can publish whatever information may be relevant about our application. And through one click, we can install our helm chart. Here, it will show us the actual details of the helm charts. So before we install it, we can actually look at those individual components. So in this case, we can see this created an ingress rule. And then this will tell Kubernetes how did it create this specific components of our application. We'd just have to pick a namespace to deploy it to and in this case, we're actually going to do a quick test here because in this case, we're trying to deploy the application from Docker hub. In our Universal Control Plane, we've turned on Docker content trust policy enforcement. So this is actually going to fail to deploy. Because we're trying to employ our application from Docker hub, the image hasn't been properly signed in our environment. So the Docker content trust policy enforcement prevents us from deploying our Docker image from Docker hub. In this case, we have to go through our approved process through our secure supply chain to be able to ensure that we know where our image came from and that meets our quality standards. So if we comment out the Docker hub repository and comment in our Docker trusted registry repository and click install, it will then install the helm chart with our Docker image being pulled from our DTR, which then it has a proper signature. We can see that our application has been successfully deployed through our home chart releases view. From here, we can see that simple NGINX application and in this case, we'll get details around the actual deployed helm chart. The nice thing is, is that Lens provides us this capability here with helm to be able to see all of the components that make up our application. From this view, it's giving us that single pane of glass into that specific application, so that we know all of the components that is created inside of Kubernetes. There are specific details that can help us access the applications such as that ingress rule that we just talked about, gives us the details of that, but it also gives us the resources such as the service, the deployment and ingress that has been created within Kubernetes to be able to actually have the application exist. So to recap, we've covered how we can offer all the benefits of a cloud like experience and offer flexibility around DevOps and operations control processes through the use of a secure supply chain, allowing our developers to spend more time developing and our operators, more time designing systems that meet our security and compliance concerns.
SUMMARY :
of our application to be
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Matt Bentley | PERSON | 0.99+ |
GitHub | ORGANIZATION | 0.99+ |
First | QUANTITY | 0.99+ |
one reason | QUANTITY | 0.99+ |
Mirantis | ORGANIZATION | 0.99+ |
One | QUANTITY | 0.99+ |
NGINX | TITLE | 0.99+ |
Docker | TITLE | 0.99+ |
two approaches | QUANTITY | 0.99+ |
Monolith | ORGANIZATION | 0.99+ |
one | QUANTITY | 0.98+ |
UCP | ORGANIZATION | 0.98+ |
Kubernetes | TITLE | 0.98+ |
One thing | QUANTITY | 0.98+ |
one developer | QUANTITY | 0.98+ |
Jenkins | TITLE | 0.98+ |
today | DATE | 0.98+ |
Brownfield | ORGANIZATION | 0.97+ |
both worlds | QUANTITY | 0.97+ |
two | QUANTITY | 0.97+ |
both | QUANTITY | 0.96+ |
one click | QUANTITY | 0.96+ |
Greenfield | ORGANIZATION | 0.95+ |
each | QUANTITY | 0.95+ |
single pane | QUANTITY | 0.92+ |
Docker hub | TITLE | 0.91+ |
a hundred | QUANTITY | 0.91+ |
Lens | TITLE | 0.9+ |
Docker | ORGANIZATION | 0.9+ |
Microservice | ORGANIZATION | 0.9+ |
VS | TITLE | 0.88+ |
DevOps | TITLE | 0.87+ |
K8S | COMMERCIAL_ITEM | 0.87+ |
Docker hub | ORGANIZATION | 0.85+ |
ways | QUANTITY | 0.83+ |
Kubernetes | ORGANIZATION | 0.83+ |
last six years | DATE | 0.82+ |
Jenkins | PERSON | 0.72+ |
One of | QUANTITY | 0.7+ |
ON DEMAND SEB CONTAINER JOURNEY DEV TO OPS FINAL
>> So, hi, my name is Daniel Terry, I work as Lead Designer at SEB. So, today we will go through why we are why we are Mirantis' customer, why we choose Docker Enterprise, and mainly what challenges we were facing before we chose to work with Docker, and where we are today, and our keys to success. >> Hi, my name is Johan, I'm a senior developer and a Tech Lead at SEB. I was in the beginning with Docker for like, four years ago. And as Daniel was saying here, we are going to present to you our journey with Docker and the answers. >> Yeah, who are we? We are SEB group. So we are a classic, financial large institutions. So, classic and traditional banking services. In Sweden, we are quite a big bank, one of the largest. And we are on a journey of transforming the bank so it has to be online 24-seven. People can do their banking business every day, whenever they want, nothing should stop them to be online. So this is putting a lot of pressure on us on infrastructure to be able to give them that service. (drum fill) >> So our timeline here. Is look, we started out with how to facilitate the container technology it has to be. 2016. And, in 2018, we had the first Docker running in SEB in a standalone mode. You need that. We didn't have any swarm, or given up this cluster since a while. For 2019, we have our first Docker-prise enterprise cluster at SEB. And today, 2020, we have the latest and greatest version of Docker installed. We are running around approximately two and a fifth at 450 specs. Around a thousand services and around 1500 containers. So, developer challenges. As for me as a developer, previous to Docker was really, really hard to get things in production. Times. It took big things and ordering services and infrastructures was a pain in the... yeah, you know what I mean? So for me, it was all about processes. We use natural processes and meaning that I wasn't able to, to see maintaining my system in production. I was handing that over to our operations teams and operation teams in that time, they didn't know how the application works. They didn't know how to troubleshoot it and see, well, what's going wrong. They were experts on the infrastructure and the platforms, but not on our applications. We were working in silos, meaning that I as a developer, only did developing things. The operations side did their things, and the security side did their things. But we didn't work as a team. I mean, today we have a completely different way of working. We will not see shapes. I mean, we have persons that were really good in maybe MQ technologies, or in some programming language and so on, but we didn't have the knowledge in the team techs to solve things, as we should have. Long lead times. I mean, everything we were trying to do had to follow the processes as we had. I mean that we should fill in some forms, send it away, hopefully someone was getting, getting back to us and saying, yeah guys, we can help you out with these services or this infrastructure, but it takes a really long time to do that. I mean, ordering infrastructure is when you're not an expert on that really hard to do. And often the orders we made or placed were wrong. When we have forms to fill in, it wasn't possible for us to do things automatically. Meaning that we didn't have the code, or the infrastructure as code. Meaning that if we didn't get the right persons into the meetings the first time, we didn't have the possibility to do it the right way, meaning that we had to redo and redo, and hopefully sometimes we got the right. We didn't have consistent set ups between the environments. When we order, as for example, a test environment, we could maybe order it with some minor resources, less CPU, so less memory, less disc or whatever. Or actually less performance on the hardware, but then we moved up to production. We realized that we have different hardware, different discs, different memories, and that could actually cause some serious problems in applications, access-wise. I mean, everyone likes to have exercise, especially if you are the maintainer of the system. That was really, really hard to get. I mean, every system has their own services, their own service, and therefore they need to apply for access to those other services. But today there's a complete difference since we only have one class to produce. Since we don't have infrastructure as a code back then, there were really lots of human errors. I mean, everyone was doing things manually. When you're coming from the Windows perspective, everything is a UI. You tend to prefer that way of working, meaning that if you used to click something in between the environments, the environments will not look the same. Life cycles. I mean, just imagine. When we have the server installed, it's like a pet. You have everything configured all from certificates to port openings, cartels, install patches, you name it. And then imagine that Windows are terminating a version and you need to reinstall that. Everything needs to be redone from the beginning. So there was a really long time taking to, to do the LCM activities, General lack of support of Microservice architecture was really also, a thing that are driving us forward with the containers technology, since we can't scale our applications in the same way as for containers. We, for example, couldn't have two applications or two processes using the same TCP port. For example, if you'd like to scale a web server, you can't do that on the same hardware. You need to have two different servers. And just imagine replacing all the excesses, replacing all the orders again for more hardware, and then manually a setting up there. The low balancer in front is a really huge task to do. And necessarily if you don't have the knowledge how the infrastructure is where you're working, then it's also really hard for you as a developer to do things right. Traditionalist. I mean, the services for us are like pets. They were really, really hard to set up. It'd take maybe a week or so. And if something was wrong with them, we will try to fix them as a pet. I mean, we couldn't just kill them and throw them away. It will actually destroy the application as this, our, like a unit box where all our things are installed. >> So, coming in from the infrastructure part of this, we've also seen challenges. For my team, we're coming from a Windows environment. So doing like a DevOps journey, which we want to do, makes it harder due to our nature in our environments. We are not used to, maybe use API, so we are not used to giving open APIs to our developers to do changes on the servers. Since we are a bank, we don't allow users to log into the servers, which means we have to do things for them all the time. This was very time consuming. And a lot of the challenges we actually still are seeing is the existing infrastructure. You can't just put that container platform on it, and thinking you're sold and everything. One of the biggest issues for us is, has been to getting servers. Windows servers usually takes like 15 minutes, Linux servers can takes up to two week in a bad day. So we really lack like, infrastructure as code. If we want a low balancer, that is also an order form. If we want the firewall opening, that's an order form. Hopefully they will not deny it. So it will go faster. So it's a lot of old processes that we need to go through. So what we wanted to do is that we want to move all of these things to the developers, so they can do it. They can own up their problems, but with our old infrastructure, that wasn't possible. We are a heavily ITIL-based organization, meaning that everything went from a cab. Still does in some way, we have one major service window every month where we take everything down. There is a lot of people involved in everything. So it's quite hard to know what will be done during the maintenance window. We lack supporting tools, or we lacked supporting tools, like log-in, good log-in tools. We have a bunch of CI/CD tools, but the maturity level of the infrastructure team wasn't that good. Again, order form and processes. If we want to, like, procure, do our procurement on a new like, storage system, or a backup system, we talk about here. So to do it is, for us, with containers, it would solve a lot of problems, because we cause we would then move the problems, not maybe move the problems to the developer, but we would make it able for them to own their own problems. So everything that we have talked about up till now boils down to business drivers. So the management's gave- gave us some policies to, or what they, how they want to change the company, so we can be this agile and fast moving bank. So one of the biggest drivers are cloud readiness, where Containers comes in perfectly. So we can build it on premises, and then we can move it to the cloud when we are ready but we can't, but we also need an exit strategy to move it back on premises if we need, due to hard regulations. Maybe you can throw it in the air. >> Absolutely. I definitely can. You're absolutely right. We need to develop things in a certain way. So we can move from infrastructure to infrastructure depending, or regardless of the vendor. Meaning that if we are able to run it on-prem, we should be able to run it in cloud or vice versa. We should also be able to move between clouds, and not be forced into one cloud provider. So that's really important for us at SEB. Short time to market is also a thing here. I mean, we are working with the huge customers. I can't name them, but they're really huge. And they need to have us being moving forward. I mean, able to really fast switch from one technology, maybe to another, we are here for them. And it's really important to us to be really fast for us to get new things out in production. All right, maybe. Nothing else? >> I don't, don't really. From the upside, we are in a huge staff DevOps transition. So, or a forced DevOps transition, which means we need to start looking at new infrastructure solutions, maybe deploy our infrastructure parts inside of containers to be able to use it the same way in the cloud. That's what we do prior, do here on premises, we have private clouds which are built on techno- technology, container technology today. So this fits quite good to have the Docker platform being one part of that one. >> Yeah. And this is solid, we are also working really, really actively on open source platforms and open source drivers. We can see that we have a huge amount of vendors in SEB, really huge ones, but we can also see that we can, facilitate open source platforms, and open source technology as well. So container technology will bring that for us. I mean, instead of having a SaaS platform and SaaS services, we can actually instantiate our own with containers and stuff. >> Also we are, since we are quite heavily regulated, the process of going through to you as like a SaaS service can take up to two years for us to go through, and then maybe the SaaS service, is it, is it what we want to use anymore? So, also we want to develop the things in our own premises and maybe, and scale it to the cloud if we need. And also we want to be an attractive employer, where maybe it's not that, the coolest thing for a young student to work in mainframe, we have a mainframe it's, it's not going anywhere, but it's hard to get people, and we want to be an attractive employer, and everyone is talking Kubernetics and containers or, and clouds. So we need to transition into those technologies. >> Yeah, we need to be open minded and necessarily facilitate the new technologies. So we can actually attract new employees. So it's really important to us to have an open mind. Our experience with Docker Containers. I mean, as I said before, scalability is a really important thing for us today. When we are using a more microservice architecture, we need to be able to Skype. We need to be scaling horizontally instead of vertically. So for that, containers are perfect storage. As we said before, we have a huge problem with environments being differently set up, since it was often manually done. Today, as we have a infrastructure as a code, it's really, really nice to have the same things exactly configured, the same in all environments. And we also have the same tooling, meaning that if I can run it on my machine, it's the same tooling I will be using to run it for test purposes or in production. That's a huge benefit for us as a developer. Time to market. I mean, today, we don't have to order service, we are using the service approach here. So we have a container cluster that are actually just sitting there waiting for our services to be hosted. So no more forms, no more calls, no more meetings before we can set up anything. We also own our problems. I mean, before, as I said, we have the processes, meaning that we ship our applications to any server. And then the operation sites take over. That's not the case now. We are actually using this as we should in DevOps. Meaning the other teams are actually responsible for all their errors as well. Even if it's on the infrastructure part, it's completely different if it's a platform's problem, because then it's the platform's team, and we can use different windows. We can try stuff out, we have an open mind. And that says that I can download and try any container image I would like on my developer machine. It's not maybe, okay to run it in production without having the security people look at it. But normally it's really, really much faster instead of waiting maybe six months, we can maybe wait one week or so. And of course less to none LCM activities. I mean, as I said before, it will take months, maybe, to do an LCM activity on multiple servers. Today, our LCM activities more or less are just switching to a new version of the image from Docker hub. That's all we have to do. So that's actually maintained during the processes we have in CI/CD pipelines. >> And the last one. So our keys to success: you should get demanded from the managers and management that everything should be a container. All the new development has to go through a container before you start ordering servers. Everything shall go through a CI/CD pipeline. We don't actually, here at SEB. Our developers build their own CI/CD pipeline. We just provide a platform for them to use it against, and the CI/CD to systems, but they build everything for themselves. Cause they know how their application works, how it should be deployed, with what tools. We just provide them with a tool set. Build a Cross Team. So you should incorporate all the processes that you need, but you should focus on the developer part, because you are building a platform for the developers, not for operations or security. >> And then maybe >> A lot of... >> you'll be able to take flight >> Yeah. Luck has nothing to do with it? Yes, it has. Of course, luck has something to do with it, even if you're really passionate, even if you're really good at some things. I mean, we got some really nice help from Dr. Inc. We were really... Came in with the technology in the right time for us to be, and we had really engaged people with these projects and that's a really luck for us to have. >> Yeah. And also we... I want to thank our colleagues, because we have another container team who started before us. And they have actually run into a lot of organizational problems, which they have sold, so we could piggyback on that, on those solutions. Also, start small and scale it. This is where Docker swarm comes, fits perfectly. So we have actually, we started with swarm. We are moving into Kubernetics in this platform. We will not force-move anything. The developers just should show us, what their- fits their needs. Thank you! >> Thank you very much.
SUMMARY :
So, today we will go through we are going to present to you our journey So we are a classic, had to follow the processes as we had. So everything that we have maybe to another, we are here for them. we have private clouds can also see that we can, to the cloud if we need. the processes we have in CI/CD pipelines. and the CI/CD to systems, I mean, we got some really So we have actually,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Daniel | PERSON | 0.99+ |
Johan | PERSON | 0.99+ |
15 minutes | QUANTITY | 0.99+ |
2018 | DATE | 0.99+ |
Sweden | LOCATION | 0.99+ |
Daniel Terry | PERSON | 0.99+ |
2019 | DATE | 0.99+ |
SEB | ORGANIZATION | 0.99+ |
six months | QUANTITY | 0.99+ |
2016 | DATE | 0.99+ |
one week | QUANTITY | 0.99+ |
2020 | DATE | 0.99+ |
Docker | ORGANIZATION | 0.99+ |
today | DATE | 0.99+ |
Skype | ORGANIZATION | 0.99+ |
450 specs | QUANTITY | 0.99+ |
two processes | QUANTITY | 0.99+ |
one class | QUANTITY | 0.99+ |
Today | DATE | 0.99+ |
first | QUANTITY | 0.99+ |
two applications | QUANTITY | 0.99+ |
four years ago | DATE | 0.99+ |
One | QUANTITY | 0.98+ |
first time | QUANTITY | 0.98+ |
a week | QUANTITY | 0.98+ |
one | QUANTITY | 0.98+ |
24 | QUANTITY | 0.98+ |
Linux | TITLE | 0.97+ |
two different servers | QUANTITY | 0.97+ |
around 1500 containers | QUANTITY | 0.97+ |
Windows | TITLE | 0.96+ |
Dr. Inc. | ORGANIZATION | 0.94+ |
up to two years | QUANTITY | 0.92+ |
one part | QUANTITY | 0.92+ |
one technology | QUANTITY | 0.89+ |
approximately two | QUANTITY | 0.89+ |
Mirantis' | ORGANIZATION | 0.88+ |
Around a thousand services | QUANTITY | 0.87+ |
Docker Enterprise | ORGANIZATION | 0.85+ |
one cloud provider | QUANTITY | 0.8+ |
fifth | QUANTITY | 0.79+ |
Docker | TITLE | 0.79+ |
up to two week | QUANTITY | 0.73+ |
one major service | QUANTITY | 0.72+ |
around | QUANTITY | 0.7+ |
Kubernetics | ORGANIZATION | 0.67+ |
seven | QUANTITY | 0.65+ |
DevOps | TITLE | 0.6+ |
every | QUANTITY | 0.56+ |
swarm | ORGANIZATION | 0.53+ |
ON DEMAND MIRANTIS OPENSTACK ON K8S FINAL
>> Hi, I'm Adrienne Davis, Customer Success Manager on the CFO-side of the house at Mirantis. With me today is Artem Andreev, Product Manager and expert, who's going to enlighten us today. >> Hello everyone. It's great to hear all of you listening to our discussion today. So my name is Artem Andreev. I'm a Product Manager for Mirantis OpenStack line of products. That includes the current product line that we have in the the next generation product line that we're about to launch quite soon. And actually this is going to be the topic of our presentation today. So the new product that we are very, very, very excited about, and that is going to be launched in a matter of several weeks, is called Mirantis OpenStack on Kubernetes. For those of you who have been in Mirantis quite a while already, Mirantis OpenStack on Kubernetes is essentially a reincarnation of our Miranti Cloud Platform version one, as we call it these days. So, and the theme has reincarnated into something more advanced, more robust, and altogether modern, that provides the same, if not more, value to our customers, but packaged in a different shape. And well, we're very excited about this new launch, and we would like to share this excitement with you Of course. As you might know, recently a few months ago, Mirantis acquired Docker Enterprise together with the advanced Kubernetes technology that Docker Enterprise provides. And we made this technology the piece and parcel of our product suite, and this naturally includes OpenStack Mirantis, OpenStack on Kubernetes as well, since this is a part of our product suite. And well, the Kubernetes technology in question, we call Docker Enterprise Container Cloud these days, I'm going to refer to this name a lot over the course of the presentation. So I would like to split today's discussions to several major parts. So for those of you who do not know what OpenStack is in general, a quick recap might be helpful to understand the value that it provides. I will discuss why someone still needs OpenStack in 2020. We will talk about what a modern OpenStack distribution is supposed to do to the expectation that is there. And of course, we will go into a bit of details of how exactly Mirantis OpenStack on Kubernetes works, how it helps to deploy and manage OpenStack clouds. >> So set the stage for me here. What's the base environment we were trying to get to? >> So what is OpenStack? One can think of OpenStack as a free and open source alternative to VMware, and it's a fair comparison. So OpenStack, just as VMware, operates primarily on Virtual Machines. So it gives you as a user, a clean and crispy interface to launch a virtual VM, to configure the virtual networking to plug this VM into it to configure and provision virtual storage, to attach to your VM, and do a lot of other things that actually a modern application requires to run. So the idea behind OpenStack is that you have a clean and crispy API exposed to you as a user, and alters little details and nuances of the physical infrastructure configuration provision that need to happen just for the virtual application to work are hidden, and spread across multiple components that comprise OpenStack per se. So as compared again, to a VMware, the functionality is pretty much similar, but actually OpenStack can do much more than just Vms, and it does that, at frankly speaking much less price, if we do the comparison. So what OpenStack has to offer. Naturally, the virtualization, networking, storage systems out there, it's just the basic entry level functionality. But of course, what comes with it is the identity and access management features, or practical user interface together with the CLI and command line tools to manage the cloud, orchestration functionality, to deploy your application in the form of templates, ability to manage bare metal machines, and of course, some nice and fancy extras like DNSaaS service, Metering, Secret Management, and Load Balancing. And frankly speaking, OpenStack can actually do even more, depending on the needs that you have. >> We hear so much about containers today. Do applications even need VMs anymore? Can't Kubernetes provide all these services? And even if IaaS is still needed, why would one bother with building their own private platform, if there's a wide choice of public solutions for virtualization, like Amazon web services, Microsoft Azure, and Google cloud platform? >> Well, that's a very fair question. And you're absolutely correct. So the whole trend (audio blurs) as the States. Everybody's talking about containers, everybody's doing containers, but to be realistic, yes, the market still needs VMs. There are certain use cases in the modern world. And actually these use cases are quite new, like 5G, where you require high performance in the networking for example. You might need high performance computing as well. So when this takes quite special hardware and configuration to be provided within your infrastructure, that is much more easily solved with the Vms, and not containers. Of course not to mention that, there are still legacy applications that you need to deal with, and that well, they have just switched from the server-based provision into VM-based provision, and they need to run somewhere. So they're not just ready for containers. And well, if we think about, okay, VMs are still needed, but why don't I just go to a public infrastructure as a service provider and run my workloads there? Now if you can do that, but well, you have to be prepared to pay a lot of money, once you start running your workloads at scale. So public IaaSes, they actually tend to hit your pockets heavily. And of course, if you're working in a highly regulated area, like enterprises cover (audio blurs) et cetera, so you have to comply with a lot of security regulations and data placement regulations. And well, public IaaSes, let's be frank, they're not good at providing you with this transparency. So you need to have full control over your whole stack, starting from the hardware to the very, very top. And this is why private infrastructure as a service is still a theme these days. And I believe that it's going to be a theme for at least five years more, if not more. >> So if private IaaSes are useful and demanded, why does Mirantis just stick to the OpenStack that we already have? Why did we decide to build a new product, rather than keep selling the current one? >> Well, to answer this question, first, we need to see what actually our customers believe more in infrastructure as a service platform should be able to provide. And we've compiled this list into like five criteria. Naturally, private IaaS needs to be reliable and robust, meaning that whatever happens on the underneath the API, that should not be impacting the business generated workloads, this is a must, or impacting them as little as possible, the platform needs to be secure and transparent, going back to the idea of working in the highly regulated areas. And this is again, a table stake to enter the enterprise market. The platform needs to be simple to deploy (audio blurs) 'cause well, you as an operator, you should not be thinking about the internals, but try to focus in on enabling your users with the best possible experience. Updates, updates are very important. So the platform needs to keep up with the latest software patches, bug fixes, and of course, features, and upgrading to a new version must not take weeks or months, and has as little impact on the running workloads as possible. And of course, to be able to run modern application, the platform needs to provide the comparable set of services, just as a public cloud so that you can move your application across your terms in the private or public cloud without having to change it severally, so-called the feature parity, it needs to be there. And if we look at the architecture of OpenStack, and we know OpenStack is powerful, it can do a lot. We've just discussed that, right? But the architecture of OpenStack is known to be complex. And well, tell me, how would you enable the robustness and robustness and reliability in this complex system? It's not easy, right? So, and actually this diagrams shelves, just like probably a third part of the modern update OpenStack cloud. So it's just a little illustration. It's not the whole picture. So imagine how hard it is to make a very solid platform out of this architecture. And well, naturally this also imposes some challenges to provide the transparency and security, 'cause well, the more complex the system is, the harder it is to manage, and well the harder it is to see what's on the inside, and well upgrades, yeah. One of the biggest challenges that we learned from our past previous history, well that many of our customers prefer to stay on the older version of OpenStack, just because, well, they were afraid of upgraded, cause they saw upgrades as time-consuming and risky and divorce. And well, instead of just switching to the latest and greatest software, they preferred reliability by sticking to the old stuff. Well, why? Well, 'cause potentially that meant implied certain impact on their workloads and well an upgrade required thorough planning and execution, just to be as as riskless as possible. And we are solving all of these challenges, of managing a system as complex as OpenStack is with Kubernetes. >> So how does Kubernetes solve these problems? >> Well, we look at OpenStack as a typical microservice architecture application, that is organized into multiple little moving parts, demons that are connected to each other and that talk to each other through the standard API. And altogether, that feels as very good feet to run on top of a Kubernetes cluster, because many of the modern applications, they fall exactly on the same pattern. >> How exactly did you put OpenStack on Kubernetes? >> Well, that's not easy. I'm going to be frank with you. And if you look at the architectural diagram, so this is a stack of Miranda's products represented with a focus of course, on the Mirantis OpenStack, as a central part. So what you see in the middle shelving pink, is Mirantis OpenStack on Kubernetes itself. And of course around that are supporting components that are needed to be there, to run OpenStack on Kubernetes successfully. So on the very bottom, there is hardware, networking, storage, computing, hardware that somebody needs to configure provision and manage, to be able to deploy the operating system on top of it. And this is just another layer of complexity that abstracts the Mirantis OpenStack on Kubernetes just from the under lake. So once we have operating system there, there needs to be a Kubernetes cluster, deployed and managed. And as I mentioned previously, we are using the capabilities that this Kuberenetes cluster provides to run OpenStack itself, the control plane that way, because everything in Mirantis OpenStack on Kuberentes is a container, or whatever you can think of. Of course naturally, it doesn't sound like an easy task to manage this multi-layered pie. And this is where Docker Enterprise Container Cloud comes into play, 'cause this is our single pane of glass into day one and day two operations for the hardware itself, for the operating system, and for Docker Enterprise Kubernetes. So it solves the need to have this underlay ready and prepared. And once the underlay is there, you go ahead, and deploy Mirantis OpenStack on Kubernetes, just as another Kubernetes application, application following the same practices and tools as you use with any other applications. So naturally of course, once you have OpenStack up and running, you can use it to create your own... To give your users ability to create their own private little Kubernetes clusters inside OpenStack projects. And this is one of the measure just cases for OpenStack these days, again, being an underlay for containers. So if you look at the operator experience, how does it look like for a human operator who is responsible for deployment the management of the cloud to deal with Mirantis OpenStack on Kubernetes? So first, you deploy Docker Enterprise Container Cloud, and you use the built-in capabilities that it provides to provision your physical infrastructure, that you discover the hardware nodes, you deploy operating system there, you do configuration of the network interfaces in storage devices there, and then you deploy Kubernetes cluster on top of that. This Kubernetes cluster is going to be dedicated to Mirantis OpenStack on Kuberenetes itself. So it's a special (indistinct) general purpose thing, that well is dedicated to OpenStack. And that means that inside of this cluster, there are a bunch of life cycle management modules, running as Kubernetes operators. So OpenStack itself has its own LCM module or operator. There is a dedicated operator for Ceph, cause Ceph is our major storage solution these days, that we integrate with. Naturally, there is a dedicated lifecycle management module for Stack Light. Stack Light is our operator, logging monitoring alerting solution for OpenStack on Kubernetes, that we bundle toegether with the whole product suite. So Kubernetes operators, directly through, it keeps the TL command or through the practical records that are provided by Docker Enterprise Container Cloud, as a part of it, to deploy the OpenStack staff and Stack Light clusters one by one, and connect them together. So instead of dealing with hundreds of YAML files, while it's five definitions, five specifications, that you're supposed to provide these days and that's safe. And although data management is performed through these APIs, just as the deployment as easily. >> All of this assumes that OpenStack has containers. Now, Mirantis was containerizing back long before Kubernetes even came along. Why did we think this would be important? >> That is true. Well, we've been containerizing OpenStack for quite a while already, it's not a new thing at all. However, is the way that we deploy OpenStack as a Kubernetes application that matters, 'cause Kubernetes solves a whole bunch of challenges that we have used to deal with, with MCP1, when deploying OpenStack on top of bare operating systems as packages. So, naturally Kubernetes provides us with... Allows us to achieve reliability through the self (audio blurs) auto-scaling mechanisms. So you define a bunch of policies that describe the behavior of OpenStack control plane. And Kubernetes follows these policies when things happen, and without actually any need for human interaction. So isolation of the dependencies or OpenStack services within Docker images is a good thing, 'cause previously we had to deal with packages and conflicts in between the versions of different libraries. So now we just ship everything together as a Docker image, and I think that early in updates is an advanced feature that Kubernetes provides natively. So updating OpenStack has never been as easy as with Kubernetes. Kubernetes also provides some fancy building blocks for network and like hold balancing, and of course, collegial tunnels, and service meshes. They're also quite helpful when dealing with such a complex application like OpenStack when things need to talk to each other and without any problem in the configuration. So Helm Reconciling is a place that also has a great deal of role. So it actually is our soul for Kubernetes. We're using Helm Bubbles, which are for opens, provide for OpenStack into upstream, as our low level layer of logic to deploy OpenStack app services and connect them to each other. And they'll naturally automatic scale-up of control plane. So adding in, YouNote is easy, you just add a new Kubernetes work up with a bunch of labels there and well, it handles the distribution of the necessary service automatically. Naturally, there are certain drawbacks. So there's fancy features come at a cost. Human operators, they need to understand Kubernetes and how it works. But this is also a good thing because everything is moving towards Kubernetes these days, so you would have to learn at some point anyway. So you can use this as a chance to bring yourself to the next level of knowledge. OpenStack is not 100% Cloud Native Application by itself. Unfortunately, there are certain components that are stateful like databases, or NOAA compute services, or open-the-switch demons, and that have to be dealt with very carefully when doing operates, updates, and all the whole deployment. So there's extra life cycle management logic build team that handles these components carefully for you. So, a bit of a complexity we had to have. And naturally, Kubernetes requires resources, and keeping the resources itself to run. So you need to have this resources available and dedicated to Kubernetes control plane, to be able to control your application, that is all OpenStack and stuff. So a bit of investment is required. >> Can anybody just containerize OpenStack services and get these benefits? >> Well, yes, the idea is not new, there's a bunch of OpStream open, sorry, community projects doing pretty much the same thing. So we are not inventing a rocket here, let's be fair. However, it's only the way that Kubernetes cooks OpenStack, gives you the robustness and reliability that enterprise and like big customers actually need. And we're doing a great deal of a job, ultimating all the possible day to work polls and all these caveats complexities of the OpenStack management inside our products. Okay, at this point, I believe we shall wrap this discussion a bit up. So let me conclude for you. So OpenStack is an opensource infrastructure as a service platform, that still has its niche in 2020th, and it's going to have it's niche for at least five years. OpenStack is a powerful but very complex tool. And the complexities of OpenStack and OpenStack life cycle management, are successfully solved by Mirantis, through the capabilities of Kubernetes distribution, that provides us with the old necessary primitives to run OpenStack, just as another containerized application these days.
SUMMARY :
on the CFO-side of the house at Mirantis. and that is going to be launched So set the stage for me here. So as compared again, to a VMware, And even if IaaS is still needed, and they need to run somewhere. So the platform needs to keep up and that talk to each other of the cloud to deal with All of this assumes that and keeping the resources itself to run. and it's going to have it's
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Adrienne Davis | PERSON | 0.99+ |
Artem Andreev | PERSON | 0.99+ |
2020 | DATE | 0.99+ |
five specifications | QUANTITY | 0.99+ |
five definitions | QUANTITY | 0.99+ |
Mirantis | ORGANIZATION | 0.99+ |
100% | QUANTITY | 0.99+ |
OpenStack | TITLE | 0.99+ |
hundreds | QUANTITY | 0.99+ |
Ceph | ORGANIZATION | 0.99+ |
Microsoft | ORGANIZATION | 0.98+ |
today | DATE | 0.98+ |
One | QUANTITY | 0.98+ |
five criteria | QUANTITY | 0.98+ |
first | QUANTITY | 0.98+ |
Kubernetes | TITLE | 0.97+ |
2020th | DATE | 0.96+ |
one | QUANTITY | 0.95+ |
ORGANIZATION | 0.93+ | |
MCP1 | TITLE | 0.92+ |
two | QUANTITY | 0.92+ |
Mirantis OpenStack | TITLE | 0.91+ |
Mirantis OpenStack | TITLE | 0.91+ |
YouNote | TITLE | 0.9+ |
Docker Enterprise | ORGANIZATION | 0.9+ |
Helm Bubbles | TITLE | 0.9+ |
Kubernetes | ORGANIZATION | 0.9+ |
least five years | QUANTITY | 0.89+ |
single | QUANTITY | 0.89+ |
Mirantis OpenStack on Kubernetes | TITLE | 0.88+ |
few months ago | DATE | 0.86+ |
OpenStack on Kubernetes | TITLE | 0.86+ |
Docker Enterprise | TITLE | 0.85+ |
K8S | TITLE | 0.84+ |
ON DEMAND BUILDING MULTI CLUSTER CONTAINER PLATFORM SPG FINAL 2
>> Hello, everyone. I'm Khalil Ahmad, Senior Director, Architecture at S&P Global. I have been working with S&P Global for six years now. Previously, I worked for Citigroup and Prudential. Overall, I have been part of IT industry for 30 years, and most of my professional career has been within financial sector in New York City metro area. I live in New Jersey with my wife and son, Daniel Khalil. I have a Master degree in software engineering from the University of Scranton, and Master in mathematics University of Punjab, Lahore. And currently I am pursuing TRIUM global Executive MBA. A joint program from the NYU Stern, LSE and HEC Paris. So today, I'm going to talk about building multi-cluster scalable container platform, supporting on-prem hybrid and multicloud use cases, how we leverage that with an S&P Global and what was our best story. As far as the agenda is concerned, I will go over, quickly the problem statement. Then I will mention the work of our core requirements, how we get solutioning, how Docker Enterprise helped us. And at the end, I will go over the pilot deployment for a proof of concept which we leverage. So, as far as the problem statement is concerned. Containers, as you all know, in the enterprise are becoming mainstream but expertise remains limited and challenges are mounting as containers enter production. Some companies are building skills internally and someone looking for partners that can help catalyze success, and choosing more integrated solutions that accelerate deployments and simplify the container environment. To overcome the challenges, we at S&P Global started our journey a few years back, taking advantage of both options. So, first of all, we met with all the stakeholder, application team, Product Manager and we define our core requirements. What we want out of this container platform, which supports multicloud and hybrid supporting on-prem as well. So, as you see my core requirements, we decided that we need first of all a roadmap or container strategy, providing guidelines on standards and specification. Secondly, with an S&P Global, we decided to introduce Platform as a Service approach, where we bring the container platform and provide that as a service internally to our all application team and all the Product Managers. Hosting multiple application on-prem as well as in multicloud. Third requirement was that we need Linux and Windows container support. In addition to that, we would also require hosted secure image registry with role based access control and image security scanning. In addition to that, we also started DevOps journey, so we want to have a full support of CI/CD pipeline. Whatever the solution we recommend from the architecture group, it should be easily integrated to the developer workstation. And developer workstation could be Windows, Mac or Linux. Orchestration, performance and control were few other parameter which we'll want to keep in mind. And the most important, dynamic scaling of container clusters. That was something we were also want to achieve, when we introduce this Platform as a Service. So, as far as the standard specification are concerned, we turn to the Open Container Initiative, the OCI. OCI was established in June 2015 by Docker and other leaders in the technology industry. And OCI operates under Linux Foundation, and currently contains two specification, runtime specification and image specification. So, at that time, it was a no brainer, other than to just stick with OCI. So, we are following the industry standard and specifications. Now the next step was, okay, the container platform. But what would be our runtime engine? What would be orchestration? And how we support, in our on-prem as well as in the multicloud infrastructure? So, when it comes to runtime engine, we decided to go with the Docker. Which is by default, runtime engine and Kubernetes. And if I may mention, DataDog in one of their public report, they say Docker is probably the most talked about infrastructure technology for the past few years. So, sticking to Docker runtime engine was another win-win game and we saw in future not bringing any challenge or issues. When it comes to orchestration. We prefer Kubernetes but that time there was a challenge, Kubernetes did not support Windows container. So, we wanted something which worked with a Linux container, and also has the ability or to orchestrate Windows containers. So, even though long term we want to stick to Kubernetes, but we also wanted to have a Docker swarm. When it comes to on-prem and multicloud, technically you could only support as of now, technology may change in future, but as of now, you can only support if you bring your own orchestration too. So, in our case, if we have control over orchestration control and not locked in with one cloud provider, that was the ideal situation. So, with all that, research, R&D and finding, we found Docker Enterprise. Which is securely built, share and run modern applications anywhere. So, when we come across Docker Enterprise, we were pleased to see that it meets our most of the core requirements. Whether it is coming on the developer machine, to integrating their workstation, building the application. Whether it comes to sharing those application, in a secure way and collaborating with our pipeline. And the lastly, when it comes to the running. If we run in hybrid or multicloud or edge, in Kubernetes, Docker Enterprise have the support all the way. So, three area one I just call up all the Docker Enterprise, choice, flexibility and security. I'm sure there's a lot more features in Docker Enterprise as a suite. But, when we looked at these three words very quickly, simplified hybrid orchestration. Define application centric policies and boundaries. Once you define, you're all set. Then you just maintain those policies. Manage diverse application across mixed infrastructure, with secure segmentation. Then it comes to secure software supply chain. Provenance across the entire lifecycle of apps and infrastructure through enforceable policy. Consistently manage all apps and infrastructure. And lastly, when it comes to infrastructure independence. It was easily forever lift and shift, because same time, our cloud journey was in the flight. We were moving from on-prem to the cloud. So, support for lift and shift application was one of our wishlist. And Docker Enterprise did not disappoint us. It also supported both traditional and micro services apps on any infrastructure. So, here we are, Docker Enterprise. Why Docker Enterprise? Some of the items in previous slides I mentioned. But in addition to those industry-leading platform, simplifying the IT operations, for running modern application at scale, anywhere. Docker Enterprise also has developer tools. So, the integration, as I mentioned earlier was smooth. In addition to all these tools, the main two components, the Universal Control Plane and the Docker Trusted Registry, solve lot of our problems. When it comes to the orchestration, we have our own Universal Control Plane. Which under the hood, manages Kubernetes and Docker swarm both clusters. So, guess what? We have a Windows support, through Docker swarm and we have a Linux support through Kubernetes. Now that paradigm has changed, as of today, Kubernetes support Windows container. So, guess what? We are well after the UCP, because we have our own orchestration tool, and we start managing Kubernetes cluster in Linux and introduce now, Windows as well. Then comes to the Docker Trusted Registry. Integrated Security and role based access control, made a very smooth transition from our RT storage to DTR. In addition to that, binary level scanning was another good feature from the security point of view. So that, these all options and our R&D landed the Docker Enterprise is the way to go. And if we go over the Docker Enterprise, we can spin up multiple clusters on-prem and in the cloud. And we have a one centralized location to manage those clusters. >> Khalil: So, with all that, now let's talk about how what was our pilot deployment, for proof of concept. In this diagram, you can see we, on the left side is our on-prem Data Center, on the right side is AWS, US East Coast. We picked up one region three zones. And on-prem, we picked up our Data Center, one of the Data Center in the United States of America, and we started the POC. So, our Universal Control Plane had a five nodes cluster. Docker Trusted Registry, also has a five node cluster. And the both, but in our on-prem Data Center. When it comes to the worker nodes, we have started with 18 node cluster, on the Linux side and the four node cluster on the Windows side. Because the major footprint which we have was on the Linux side, and the Windows use cases were pretty small. Also, this is just a proof of concept. And in AWS, we mimic the same web worker nodes, virtual to what we have on-prem. We have a 13 nodes cluster on Linux. And we started with four node cluster of Windows container. And having the direct connect from our Data Center to AWS, which was previously existing, so we did not have any connectivity or latency issue. Now, if you see in this diagram, you have a centralized, Universal Control Plane and your trusted registry. And we were able to spin up a cluster, on-prem as well as in the cloud. And we made this happen, end to end in record time. So later, when we deploy this in production, we also added another cloud provider. So, what you see the box on the right side, we just duplicate test that box in another cloud platform. So, now other orchestration tool, managing on-prem and multicloud clusters. Now, in your use case, you may find this little, you know, more in favor of on-prem. But that fit in our use case. Later, we did have expanded the cluster of Universal Control Plane and DTR in the cloud as well. And the clusters have gone and hundreds and thousands of worker nodes span over two cloud providers, third being discussed. And this solution has been working so far, very good. We did not see any downtime, not a single instance. And we were able to provide multicloud platform, container Platform as a Service for our S&P Global. Thank you for your time. If any questions, I have put my LinkedIn and Twitter account holder, you're welcome to ask any question
SUMMARY :
and in the cloud. and the Windows use
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Daniel Khalil | PERSON | 0.99+ |
Citigroup | ORGANIZATION | 0.99+ |
S&P Global | ORGANIZATION | 0.99+ |
June 2015 | DATE | 0.99+ |
S&P Global | ORGANIZATION | 0.99+ |
Khalil Ahmad | PERSON | 0.99+ |
LSE | ORGANIZATION | 0.99+ |
six years | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
30 years | QUANTITY | 0.99+ |
New Jersey | LOCATION | 0.99+ |
Prudential | ORGANIZATION | 0.99+ |
United States of America | LOCATION | 0.99+ |
New York City | LOCATION | 0.99+ |
13 nodes | QUANTITY | 0.99+ |
University of Scranton | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.99+ | |
OCI | ORGANIZATION | 0.99+ |
University of Punjab | ORGANIZATION | 0.99+ |
today | DATE | 0.99+ |
Linux | TITLE | 0.99+ |
three words | QUANTITY | 0.99+ |
third | QUANTITY | 0.99+ |
Windows | TITLE | 0.99+ |
Linux Foundation | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.98+ | |
Khalil | PERSON | 0.98+ |
three zones | QUANTITY | 0.98+ |
both | QUANTITY | 0.98+ |
HEC Paris | ORGANIZATION | 0.98+ |
one | QUANTITY | 0.98+ |
Docker | TITLE | 0.98+ |
NYU Stern | ORGANIZATION | 0.98+ |
five nodes | QUANTITY | 0.97+ |
two components | QUANTITY | 0.97+ |
both options | QUANTITY | 0.97+ |
Docker Enterprise | TITLE | 0.97+ |
Secondly | QUANTITY | 0.96+ |
single instance | QUANTITY | 0.96+ |
first | QUANTITY | 0.95+ |
Kubernetes | TITLE | 0.94+ |
two cloud providers | QUANTITY | 0.94+ |
DataDog | ORGANIZATION | 0.93+ |
Docker | ORGANIZATION | 0.93+ |
two | QUANTITY | 0.92+ |
Third requirement | QUANTITY | 0.92+ |
four node | QUANTITY | 0.91+ |
both clusters | QUANTITY | 0.91+ |
TRIUM | ORGANIZATION | 0.91+ |
five node cluster | QUANTITY | 0.88+ |
Docker Enterprise | ORGANIZATION | 0.87+ |
US East Coast | LOCATION | 0.85+ |
one cloud provider | QUANTITY | 0.83+ |
Lahore | LOCATION | 0.82+ |
Open Container Initiative | ORGANIZATION | 0.81+ |
ON DEMAND API GATEWAYS INGRESS SERVICE MESH
>> Thank you, everyone for joining. I'm here today to talk about ingress controllers, API gateways, and service mesh on Kubernetes, three very hot topics that are also frequently confusing. So I'm Richard Li, founder/CEO of Ambassador Labs, formerly known as Datawire. We sponsor a number of popular open source projects that are part of the Cloud Native Computing Foundation, including Telepresence and Ambassador, which is a Kubernetes native API gateway. And most of what I'm going to talk about today is related to our work around Ambassador. So I want to start by talking about application architecture and workflow on Kubernetes and how applications that are being built on Kubernetes really differ from how they used to be built. So when you're building applications on Kubernetes, the traditional architecture is the very famous monolith. And the monolith is a central piece of software. It's one giant thing that you build deploy, run. And the value of a monolith is it's really simple. And if you think about the monolithic development process, more importantly is that architecture is really reflected in that workflow. So with a monolith, you have a very centralized development process. You tend not to release too frequently because you have all these different development teams that are working on different features, and then you decide in advance when you're going to release that particular piece of software and everyone works towards that release train. And you have specialized teams. You have a development team, which has all your developers. You have a QA team, you have a release team, you have an operations team. So that's your typical development organization and workflow with a monolithic application. As organizations shift to microservices, they adopt a very different development paradigm. It's a decentralized development paradigm where you have lots of different independent teams that are simultaneously working on different parts of this application, and those application components are really shipped as independent services. And so you really have a continuous release cycle because instead of synchronizing all your teams around one particular vehicle, you have so many different release vehicles that each team is able to ship as soon as they're ready. And so we call this full cycle development because that team is really responsible not just for the coding of that microservice, but also the testing and the release and operations of that service. So this is a huge change, particularly with workflow, and there's a lot of implications for this. So I have a diagram here that just tries to visualize a little bit more the difference in organization. With the monolith, you have everyone who works on this monolith. With microservices, you have the yellow folks work on the yellow microservice and the purple folks work on the purple microservice and maybe just one person work on the orange microservice and so forth. So there's a lot more diversity around your teams and your microservices, and it lets you really adjust the granularity of your development to your specific business needs. So how do users actually access your microservices? Well, with a monolith, it's pretty straightforward. You have one big thing, so you just tell the internet, well, I have this one big thing on the internet. Make sure you send all your traffic to the big thing. But when you have microservices and you have a bunch of different microservices, how do users actually access these microservices? So the solution is an API gateway. So the API gateway consolidates all access to your microservices. So requests come from the internet. They go to your API gateway. The API gateway looks at these requests, and based on the nature of these requests, it routes them to the appropriate microservice. And because the API gateway is centralizing access to all of the microservices, it also really helps you simplify authentication, observability, routing, all these different cross-cutting concerns, because instead of implementing authentication in each of your microservices, which would be a maintenance nightmare and a security nightmare, you've put all of your authentication in your API gateway. So if you look at this world of microservices, API gateways are a really important part of your infrastructure which are really necessary, and pre-microservices, pre-Kubernetes, an API gateway, while valuable, was much more optional. So that's one of the really big things around recognizing with the microservices architecture, you really need to start thinking much more about an API gateway. The other consideration with an API gateway is around your management workflow, because as I mentioned, each team is actually responsible for their own microservice, which also means each team needs to be able to independently manage the gateway. So Team A working on that microservice needs to be able to tell the API gateway, this is how I want you to route requests to my microservice, and the purple team needs to be able to say something different for how purple requests get routed to the purple microservice. So that's also a really important consideration as you think about API gateways and how it fits in your architecture, because it's not just about your architecture, it's also about your workflow. So let me talk about API gateways on Kubernetes. I'm going to start by talking about ingress. So ingress is the process of getting traffic from the internet to services inside the cluster. Kubernetes, from an architectural perspective, it actually has a requirement that all the different pods in a Kubernetes cluster needs to communicate with each other. And as a consequence, what Kubernetes does is it creates its own private network space for all these pods, and each pod gets its own IP address. So this makes things very, very simple for interpod communication. Kubernetes, on the other hand, does not say very much around how traffic should actually get into the cluster. So there's a lot of detail around how traffic actually, once it's in the cluster, how you route it around the cluster, and it's very opinionated about how this works, but getting traffic into the cluster, there's a lot of different options and there's multiple strategies. There's Pod IP, there's Ingress, there's LoadBalancer resources, there's NodePort. I'm not going to go into exhaustive detail on all these different options, and I'm going to just talk about the most common approach that most organizations take today. So the most common strategy for routing is coupling an external load balancer with an ingress controller. And so an external load balancer can be a hardware load balancer. It can be a virtual machine. It can be a cloud load balancer. But the key requirement for an external load balancer is to be able to attach a stable IP address so that you can actually map a domain name and DNS to that particular external load balancer, and that external load balancer usually, but not always, will then route traffic and pass that traffic straight through to your ingress controller. And then your ingress controller takes that traffic and then routes it internally inside Kubernetes to the various pods that are running your microservices. There are other approaches, but this is the most common approach. And the reason for this is that the alternative approaches really require each of your microservices to be exposed outside of the cluster, which causes a lot of challenges around management and deployment and maintenance that you generally want to avoid. So I've been talking about an ingress controller. What exactly is an ingress controller? So an ingress controller is an application that can process rules according to the Kubernetes ingress specification. Strangely, Kubernetes is not actually shipped with a built-in ingress controller. I say strangely because you think, well, getting traffic into a cluster is probably a pretty common requirement, and it is. It turns out that this is complex enough that there's no one size fits all ingress controller. And so there is a set of ingress rules that are part of the Kubernetes ingress specification that specify how traffic gets routed into the cluster, and then you need a proxy that can actually route this traffic to these different pods. And so an ingress controller really translates between the Kubernetes configuration and the proxy configuration, and common proxies for ingress controllers include HAProxy, Envoy Proxy, or NGINX. So let me talk a little bit more about these common proxies. So all these proxies, and there are many other proxies. I'm just highlighting what I consider to be probably the three most well-established proxies, HAProxy, NGINX, and Envoy Proxy. So HAProxy is managed by HAProxy Technologies. Started in 2001. The HAProxy organization actually creates an ingress controller. And before they created an ingress controller, there was an open source project called Voyager which built an ingress controller on HAProxy. NGINX, managed by NGINX, Inc., subsequently acquired by F5. Also open source. Started a little bit later, the proxy, in 2004. And there's the Nginx-ingress, which is a community project. That's the most popular. As well as the Nginx, Inc. kubernetes-ingress project, which is maintained by the company. This is a common source of confusion because sometimes people will think that they're using the NGINX ingress controller, and it's not clear if they're using this commercially supported version or this open source version. And they actually, although they have very similar names, they actually have different functionality. Finally, Envoy Proxy, the newest entrant to the proxy market, originally developed by engineers at Lyft, the ride sharing company. They subsequently donated it to the Cloud Native Computing Foundation. Envoy has become probably the most popular cloud native proxy. It's used by Ambassador, the API gateway. It's used in the Istio service mesh. It's used in the VMware Contour. It's been used by Amazon in App Mesh. It's probably the most common proxy in the cloud native world. So as I mentioned, there's a lot of different options for ingress controllers. The most common is the NGINX ingress controller, not the one maintained by NGINX, Inc., but the one that's part of the Kubernetes project. Ambassador is the most popular Envoy-based option. Another common option is the Istio Gateway, which is directly integrated with the Istio mesh, and that's actually part of Docker Enterprise. So with all these choices around ingress controller, how do you actually decide? Well, the reality is the ingress specification's very limited. And the reason for this is that getting traffic into a cluster, there's a lot of nuance into how you want to do that, and it turns out it's very challenging to create a generic one size fits all specification because of the vast diversity of implementations and choices that are available to end users. And so you don't see ingress specifying anything around resilience. So if you want to specify a timeout or rate-limiting, it's not possible. Ingress is really limited to support for HTTP. So if you're using gRPC or web sockets, you can't use the ingress specification. Different ways of routing, authentication. The list goes on and on. And so what happens is that different ingress controllers extend the core ingress specification to support these use cases in different ways. So NGINX ingress, they actually use a combination of config maps and the ingress resources plus custom annotations that extend the ingress to really let you configure a lot of the additional extensions that is exposed in the NGINX ingress. With Ambassador, we actually use custom resource definitions, different CRDs that extend Kubernetes itself to configure Ambassador. And one of the benefits of the CRD approach is that we can create a standard schema that's actually validated by Kubernetes. So when you do a kub control apply of an Ambassador CRD, kub control can immediately validate and tell you if you're actually applying a valid schema and format for your Ambassador configuration. And as I previously mentioned, Ambassador's built on Envoy Proxy, Istio Gateway also uses CRDs. They can be used in extension of the service mesh CRDs as opposed to dedicated gateway CRDs. And again, Istio Gateway is built on Envoy Proxy. So I've been talking a lot about ingress controllers, but the title of my talk was really about API gateways and ingress controllers and service mesh. So what's the difference between an ingress controller and an API gateway? So to recap, an ingress controller processes Kubernetes ingress routing rules. An API gateway is a central point for managing all your traffic to Kubernetes services. It typically has additional functionality such as authentication, observability, a developer portal, and so forth. So what you find is that not all API gateways are ingress controllers because some API gateways don't support Kubernetes at all. So you can't, they can't be ingress controllers. And not all ingress controllers support the functionality such as authentication, observability, developer portal, that you would typically associate with an API gateway. So generally speaking, API gateways that run on Kubernetes should be considered a superset of an ingress controller. But if the API gateway doesn't run on Kubernetes, then it's an API gateway and not an ingress controller. So what's the difference between a service mesh and an API gateway? So an API gateway is really focused on traffic into and out of a cluster. So the colloquial term for this is North/South traffic. A service mesh is focused on traffic between services in a cluster, East/West traffic. All service meshes need an API gateway. So Istio includes a basic ingress or API gateway called the Istio Gateway, because a service mesh needs traffic from the internet to be routed into the mesh before it can actually do anything. Envoy Proxy, as I mentioned, is the most common proxy for both mesh and gateways. Docker Enterprise provides an Envoy-based solution out of the box, Istio Gateway. The reason Docker does this is because, as I mentioned, Kubernetes doesn't come package with an ingress. It makes sense for Docker Enterprise to provide something that's easy to get going, no extra steps required, because with Docker enterprise, you can deploy it and get going, get it exposed on the internet without any additional software. Docker Enterprise can also be easily upgraded to Ambassador because they're both built on Envoy. It ensures consistent routing semantics. And also with Ambassador, you get greater security for single sign-on. There's a lot of security by default that's configured directly into Ambassador. Better control over TLS, things like that. And then finally, there's commercial support that's actually available for Ambassador. Istio is an open source project that has a very broad community, but no commercial support options. So to recap, ingress controllers and API gateways are critical pieces of your cloud native stack. So make sure that you choose something that works well for you. And I think a lot of times organizations don't think critically enough about the API gateway until they're much further down the Kubernetes journey. Considerations around how to choose that API gateway include functionality such as how does it do with traffic management and observability? Does it support the protocols that you need? Also nonfunctional requirements such as does it integrate with your workflow? Do you offer commercial support? Can you get commercial support for this? An API gateway is focused on North/South traffic, so traffic into and out of your Kubernetes cluster. A service mesh is focused on East/West traffic, so traffic between different services inside the same cluster. Docker Enterprise includes Istio Gateway out of the box. Easy to use, but can also be extended with Ambassador for enhanced functionality and security. So thank you for your time. Hope this was helpful in understanding the difference between API gateways, ingress controllers, and service meshes, and how you should be thinking about that on your Kubernetes deployment.
SUMMARY :
So ingress is the process
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
2004 | DATE | 0.99+ |
Richard Li | PERSON | 0.99+ |
2001 | DATE | 0.99+ |
Ambassador Labs | ORGANIZATION | 0.99+ |
Cloud Native Computing Foundation | ORGANIZATION | 0.99+ |
each team | QUANTITY | 0.99+ |
Cloud Native Computing Foundation | ORGANIZATION | 0.99+ |
each team | QUANTITY | 0.99+ |
Datawire | ORGANIZATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
each pod | QUANTITY | 0.99+ |
Lyft | ORGANIZATION | 0.99+ |
Nginx, Inc. | ORGANIZATION | 0.99+ |
today | DATE | 0.98+ |
each | QUANTITY | 0.98+ |
Kubernetes | TITLE | 0.98+ |
one person | QUANTITY | 0.98+ |
HAProxy Technologies | ORGANIZATION | 0.98+ |
HAProxy | TITLE | 0.97+ |
Docker Enterprise | TITLE | 0.96+ |
Ambassador | ORGANIZATION | 0.96+ |
both | QUANTITY | 0.96+ |
NGINX | TITLE | 0.96+ |
NGINX, Inc. | ORGANIZATION | 0.96+ |
Docker Enterprise | TITLE | 0.96+ |
Envoy Proxy | TITLE | 0.96+ |
one | QUANTITY | 0.95+ |
one big thing | QUANTITY | 0.95+ |
NGINX ingress | TITLE | 0.95+ |
Docker enterprise | TITLE | 0.94+ |
one particular vehicle | QUANTITY | 0.93+ |
ingress | ORGANIZATION | 0.91+ |
Telepresence | ORGANIZATION | 0.87+ |
F5 | ORGANIZATION | 0.87+ |
Envoy | TITLE | 0.86+ |
Nginx-ingress | TITLE | 0.85+ |
three very hot topics | QUANTITY | 0.82+ |
both mesh | QUANTITY | 0.82+ |
three most well-established proxies | QUANTITY | 0.76+ |
single sign | QUANTITY | 0.75+ |
Istio Gateway | OTHER | 0.75+ |
one giant thing | QUANTITY | 0.73+ |
VMware Contour | TITLE | 0.71+ |
Ingress | ORGANIZATION | 0.7+ |
Docker Enterprise | ORGANIZATION | 0.69+ |
Ambassador | TITLE | 0.67+ |
Voyager | TITLE | 0.67+ |
Envoy | ORGANIZATION | 0.65+ |
Istio Gateway | TITLE | 0.65+ |
Istio | ORGANIZATION | 0.62+ |
Dave Van Everen, Mirantis | Mirantis Launchpad 2020 Preview
>>from the Cube Studios in Palo Alto in Boston, connecting with thought leaders all around the world. This is a cube conversation. >>Hey, welcome back. You're ready, Jeffrey here with the Cuban Apollo Alto studios today, and we're excited. You know, we're slowly coming out of the, uh, out of the summer season. We're getting ready to jump back into the fall. Season, of course, is still covet. Everything is still digital. But you know, what we're seeing is a digital events allow a lot of things that you couldn't do in the physical space. Mainly get a lot more people to attend that don't have to get in airplanes and file over the country. So to preview this brand new inaugural event that's coming up in about a month, we have We have a new guest. He's Dave and Everen. He is the senior vice president of marketing. Former ran tous. Dave. Great to see you. >>Happy to be here today. Thank you. >>Yeah. So tell us about this inaugural event. You know, we did an event with Miranda's years ago. I had to look it up like 2014. 15. Open stack was hot and you guys sponsored a community event in the Bay Area because the open stack events used to move all over the country each and every year. But you guys said, and the top one here in the Bay Area. But now you're launching something brand new based on some new activity that you guys have been up to over the last several months. So let us give us give us the word. >>Yeah, absolutely. So we definitely have been organizing community events in a variety of open source communities over the years. And, you know, we saw really, really good success with with the Cube And are those events in opens tax Silicon Valley days? And, you know, with the way things have gone this year, we've really seen that virtual events could be very successful and provide a new, maybe slightly different form of engagement but still very high level of engagement for our guests and eso. We're excited to put this together and invite the entire cloud native industry to join us and learn about some of the things that Mantis has been working on in recent months. A zwelling as some of the interesting things that are going on in the Cloud native and kubernetes community >>Great. So it's the inaugural event is called Moran Sous launchpad 2020. The Wares and the Winds in September 16th. So we're about a month away and it's all online is their registration. Costars is free for the community. >>It's absolutely free. Eso everyone is welcome to attend You. Just visit Miranda's dot com and you'll see the info for registering for the event and we'd love it. We love to see you there. It's gonna be a fantastic event. We have multiple tracks catering to developers, operators, general industry. Um, you know, participants in the community and eso we'd be happy to see you on join us on and learn about some of the some of the things we're working on. >>That's awesome. So let's back up a step for people that have been paying as close attention as they might have. Right? So you guys purchase, um, assets from Docker at the end of last year, really taken over there, they're they're kind of enterprise solutions, and you've been doing some work with that. Now, what's interesting is we we cover docker con, um, A couple of months ago, a couple three months ago. Time time moves fast. They had a tremendously successful digital event. 70,000 registrants, people coming from all over the world. I think they're physical. Event used to be like four or 5000 people at the peak, maybe 6000 Really tremendous success. But a lot of that success was driven, really by the by the strength of the community. The docker community is so passionate. And what struck me about that event is this is not the first time these people get together. You know, this is not ah, once a year, kind of sharing of information and sharing ideas, but kind of the passion and and the friendships and the sharing of information is so, so good. You know, it's a super or, um, rich development community. You guys have really now taken advantage of that. But you're doing your Miranda's thing. You're bringing your own technology to it and really taking it to more of an enterprise solution. So I wonder if you can kind of walk people through the process of, you know, you have the acquisition late last year. You guys been hard at work. What are we gonna see on September 16. >>Sure, absolutely. And, you know, just thio Give credit Thio Docker for putting on an amazing event with Dr Khan this year. Uh, you know, you mentioned 70,000 registrants. That's an astounding number. And you know, it really is a testament thio. You know, the community that they've built over the years and continue to serve eso We're really, really happy for Docker as they kind of move into, you know, the next the next path in their journey and, you know, focus more on the developer oriented, um, solution and go to market. So, uh, they did a fantastic job with the event. And, you know, I think that they continue toe connect with their community throughout the year on That's part of what drives What drove so many attendees to the event assed faras our our history and progress with with Dr Enterprise eso. As you mentioned mid November last year, we did acquire Doctor Enterprise assets from Docker Inc and, um, right away we noticed tremendous synergy in our product road maps and even in the in the team's eso that came together really, really quickly and we started executing on a Siris of releases. Um that are starting Thio, you know, be introduced into the market. Um, you know, one was introduced in late May and that was the first major release of Dr Enterprise produced exclusively by more antis. And we're going to announce at the launch pad 2020 event. Our next major release of the Doctor Enterprise Technology, which will for the first time include kubernetes related in life cycle management related technology from Mirant is eso. It's a huge milestone for our company. Huge benefit Thio our customers on and the broader user community around Dr Enterprise. We're super excited. Thio provide a lot of a lot of compelling and detailed content around the new technology that will be announcing at the event. >>So I'm looking at the at the website with with the agenda and there's a little teaser here right in the middle of the spaceship Docker Enterprise Container Cloud. So, um, and I glanced into you got a great little layout, five tracks, keynote track D container track operations and I t developer track and keep track. But I did. I went ahead and clicked on the keynote track and I see the big reveal so I love the opening keynote at at 8 a.m. On the 76 on the September 16th is right. Um, I, Enel CEO who have had on many, many times, has the big reveal Docker Enterprise Container Cloud. So without stealing any thunder, uh, can you give us any any little inside inside baseball on on what people should expect or what they can get excited about for that big announcement? >>Sure, absolutely so I definitely don't want to steal any thunder from Adrian, our CEO. But you know, we did include a few Easter eggs, so to speak, in the website on Dr Enterprise. Container Cloud is absolutely the biggest story out of the bunch eso that's visible on the on the rocket ship as you noticed, and in the agenda it will be revealed during Adrian's keynote, and every every word in the product name is important, right? So Dr Enterprise, based on Dr Enterprise Platform Container Cloud and there's the new word in there really is Cloud eso. I think, um, people are going to be surprised at the groundbreaking territory that were forging with with this release along the lines of a cloud experience and what we are going to provide to not only I t operations and the Op Graders and Dev ops for cloud environment, but also for the developers and the experience that we could bring to developers As they become more dependent on kubernetes and get more hands on with kubernetes. We think that we're going thio provide ah lot of ways for them to be more empowered with kubernetes while at the same time lowering the bar, the bar or the barrier of entry for kubernetes. As many enterprises have have told us that you know kubernetes can be difficult for the broader developer community inside the organization Thio interact with right? So this is, uh, you know, a strategic underpinning of our our product strategy. And this is really the first step in a non going launch of technologies that we're going to make bigger netease easier for developing. >>I was gonna say the other Easter egg that's all over the agenda, as I'm just kind of looking through the agenda. It's kubernetes on 80 infrastructure multi cloud kubernetes Miranda's open stack on kubernetes. So Goober Netease plays a huge part and you know, we talk a lot about kubernetes at all the events that we cover. But as you said, kind of the new theme that we're hearing a little bit more Morris is the difficulty and actually managing it so looking, kind of beyond the actual technology to the operations and the execution in production. And it sounds like you guys might have a few things up your sleeve to help people be more successful in in and actually kubernetes in production. >>Yeah, absolutely. So, uh, kubernetes is the focus of most of the companies in our space. Obviously, we think that we have some ideas for how we can, you know, really begin thio enable enable it to fulfill its promise as the operating system for the cloud eso. If we think about the ecosystem that's formed around kubernetes, uh, you know, it's it's now really being held back on Lee by adoption user adoption. And so that's where our focus in our product strategy really lives is around. How can we accelerate the move to kubernetes and accelerate the move to cloud native applications on? But in order to provide that acceleration catalyst, you need to be able to address the needs of not only the operators and make their lives easier while still giving them the tools they need for things like policy enforcement and operational insights. At the same time, Foster, you know, a grassroots, um, upswell of developer adoption within their company on bond Really help the I t. Operations team serve their customers the developers more effectively. >>Well, Dave, it sounds like a great event. We we had a great time covering those open stack events with you guys. We've covered the doctor events for years and years and years. Eso super engaged community and and thanks for, you know, inviting us back Thio to cover this inaugural event as well. So it should be terrific. Everyone just go to Miranda's dot com. The big pop up Will will jump up. You just click on the button and you can see the full agenda on get ready for about a month from now. When when the big reveal, September 16th will happen. Well, Dave, thanks for sharing this quick update with us. And I'm sure we're talking a lot more between now in, uh, in the 16 because I know there's a cube track in there, so we look forward to interview in our are our guests is part of the part of the program. >>Absolutely. Eso welcome everyone. Join us at the event and, uh, you know, stay tuned for the big reveal. >>Everybody loves a big reveal. All right, well, thanks a lot, Dave. So he's Dave. I'm Jeff. You're watching the Cube. Thanks for watching. We'll see you next time.
SUMMARY :
from the Cube Studios in Palo Alto in Boston, connecting with thought leaders all around the world. But you know, what we're seeing is a digital Happy to be here today. But you guys said, and the top one here in the Bay Area. invite the entire cloud native industry to join us and The Wares and the Winds in September 16th. participants in the community and eso we'd be happy to see you on So you guys purchase, um, assets from Docker at the end of last year, you know, focus more on the developer oriented, um, solution and So I'm looking at the at the website with with the agenda and there's a little teaser here right in the on the on the rocket ship as you noticed, and in the agenda it will be revealed So Goober Netease plays a huge part and you know, we talk a lot about kubernetes at all the events that we cover. some ideas for how we can, you know, really begin thio enable You just click on the button and you can see the full agenda on uh, you know, stay tuned for the big reveal. We'll see you next time.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Adrian | PERSON | 0.99+ |
September 16 | DATE | 0.99+ |
Dave | PERSON | 0.99+ |
Jeffrey | PERSON | 0.99+ |
Dave Van Everen | PERSON | 0.99+ |
Jeff | PERSON | 0.99+ |
Everen | PERSON | 0.99+ |
Docker | ORGANIZATION | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
September 16th | DATE | 0.99+ |
Docker Inc | ORGANIZATION | 0.99+ |
Bay Area | LOCATION | 0.99+ |
late May | DATE | 0.99+ |
Enel | PERSON | 0.99+ |
mid November last year | DATE | 0.99+ |
5000 people | QUANTITY | 0.99+ |
four | QUANTITY | 0.99+ |
70,000 registrants | QUANTITY | 0.99+ |
Dr Enterprise | ORGANIZATION | 0.99+ |
Boston | LOCATION | 0.99+ |
today | DATE | 0.99+ |
Mirantis | ORGANIZATION | 0.99+ |
8 a.m. | DATE | 0.99+ |
first time | QUANTITY | 0.98+ |
Docker Enterprise Container Cloud | TITLE | 0.98+ |
Doctor Enterprise | ORGANIZATION | 0.98+ |
Foster | PERSON | 0.98+ |
Dr Enterprise | TITLE | 0.98+ |
2014. 15 | DATE | 0.98+ |
first step | QUANTITY | 0.98+ |
80 | QUANTITY | 0.98+ |
Docker Enterprise Container Cloud | TITLE | 0.98+ |
6000 | QUANTITY | 0.97+ |
this year | DATE | 0.97+ |
Cube Studios | ORGANIZATION | 0.96+ |
late last year | DATE | 0.96+ |
Container Cloud | TITLE | 0.96+ |
five tracks | QUANTITY | 0.96+ |
Easter | EVENT | 0.96+ |
Morris | PERSON | 0.96+ |
The Wares and the Winds | EVENT | 0.95+ |
Miranda | PERSON | 0.94+ |
Dr | PERSON | 0.94+ |
Silicon Valley | LOCATION | 0.94+ |
first time | QUANTITY | 0.94+ |
once a year | QUANTITY | 0.93+ |
each | QUANTITY | 0.9+ |
end | DATE | 0.9+ |
couple of months ago | DATE | 0.88+ |
Launchpad | COMMERCIAL_ITEM | 0.88+ |
Apollo Alto studios | ORGANIZATION | 0.87+ |
Cloud | TITLE | 0.87+ |
about a month | QUANTITY | 0.86+ |
Thio | PERSON | 0.85+ |
Will | PERSON | 0.85+ |
Mantis | PERSON | 0.84+ |
Mirant | ORGANIZATION | 0.84+ |
Thio Docker | PERSON | 0.83+ |
Doctor Enterprise | TITLE | 0.82+ |
a month | QUANTITY | 0.82+ |
Khan | PERSON | 0.81+ |
first major release | QUANTITY | 0.81+ |
last year | DATE | 0.8+ |
2020 | DATE | 0.8+ |
couple three months ago | DATE | 0.79+ |
years | QUANTITY | 0.79+ |
years | DATE | 0.75+ |
Moran Sous launchpad 2020 | EVENT | 0.72+ |
Thio | ORGANIZATION | 0.72+ |
Lee | PERSON | 0.71+ |
Miranda | ORGANIZATION | 0.7+ |
one | QUANTITY | 0.69+ |
Container Cloud | TITLE | 0.67+ |
months | DATE | 0.66+ |
Dr Enterprise Platform | TITLE | 0.65+ |
76 | DATE | 0.64+ |
Chris Brown, Nutanix | DockerCon 2018
>> Live from San Francisco, it's theCUBE! Covering DockerCon 18, brought to you by Docker and it's ecosystem partners. >> Welcome back to theCUBE, I'm Lisa Martin with John Troyer we are live from DockerCon 2018 on a sunny day here in San Francisco at Moscone Center. Excited to welcome to theCUBE Chris Brown the Technical Marketing Manager at Nutanix, Chris welcome to theCUBE! >> Thank you so much for having me. >> So you've been with Nutanix for a couple years, so we'll talk about Nutanix and containers, you have a session control and automate your container journey with Nutanix. Talk to us about what you're gonna be talking about in the session, what's Nutanix's role in helping the customers get over this trepidation of containers? >> Yeah, definitely, and it's, it's a 20 minute session, so we've got a lot of information to cover there 'cause wanna go over a little bit about, you know, who Nutanix is from the beginning to end but, the main part I'm gonna be focusing on in that session is talking about how we, with our com product, can automate VMs and containers together and how we're moving towards being able to, you know, define you application in a blueprint and understand what you're trying to do with your application. You know, one of the things I always say is that nobody runs Sequel because they love running Sequel, they run Sequel to do something, and our goal with the com is to capture that something, what it depends on, what it relies on. Once we understand what this particular component is supposed to do in your application, we can change that, we can move that to another cloud, or we can move it to containers without losing that definition, and without losing its dependence on the other pieces of the infrastructure and exchange information back and forth. So we're talking a little bit about what we're doing today with com and where we're going with it to add Kubernetes support. >> Chris, we're sitting here in the ecosystem expo at DockerCon and your booth is busy, there's a lot of good activity. Are people coming up to you and asking, do they know Nutanix, do they understand who you are, do they just say oh you guys sell boxes? You know you're both a, you're a systems provider, you're a private cloud provider, and a hybrid-cloud provider, do people understand that, the crowd here, and what kinda conversations are you having? >> It's actually really interesting 'cause we're seeing a broad range of people, some customers are comin' up, or some people are coming up that they don't reali--they don't know that other pieces, places their company use Nutanix, but they wanted to learn more about us, so they've got some sort of initiative that you know, a lot of times it is around containers, around understanding, you know, they're starting to figure out, you know, how do we deploy this, how do we connect? You know, we've got something we wanna deploy here and there how do we do that in a scalable way? But we also have some that have no idea who we are and just comin' up like so you've got a booth and some awesome giveaways, (laughing) what do I have to do to get that, and what do you do? And you know, I really kinda summarize it as two main main groups of people that I've seen is, one of 'em is, the people who've been doing containers for forever, they know it, they've been doing it, they're very familiar with the command line, they're ret-- any gooey is too much gooey for them. And then we've got the people who are just getting started, they've kinda been told hey, containers are coming, we need to figure out how to do this, or we've got, we need to start figuring out our containers strategy. And so they're here to learn and figure out how to begin that. And so it's really interesting because those, the ones that are just getting started or just learning, we obviously help out a ton because the people who came before had to go through all the fire, all the configuration, all of the challenges, and figure out there own solutions where as we can, now we kinda come in, there's a little bit more opinionated example of how to do these things. >> So DockerCon, this year is the fifth DockerCon, they've got between five thousand and six thousand people, I was talking with John earlier and Steve Singh as well, that how I really impressed when I was leaving the general session, it was standing room only a sea of heads so they've got, obviously developers here right, sweet spot, IT folks, enterprise architects, and execs, you talked about Nutanix getting those the two polar opposite ends of the spectrum, the container lovers, the ones who are the experts, and the ones going I know I have to do this. I'm curious, what target audience are you talking to that goes hey I'm tasked with doing this, are those developers, are those IT folks, are you talking with execs as well, give us that mix. >> For the most part they are IT folks, you're artusional operators who are trying to figure out this new shift in technology and we have to talk to some developers, and it's actually been interesting to have speak with developers because you know, in general that's not, that hasn't been Nutanix's traditional audience, we've sold this product called infrastructure to develop. But developers, the few developers I've talked to have gotten really receptive and really excited about what we can do and how we can help them do their job faster by getting their IT people on board but for the most part it'd be traditional IT operators who're looking at this new technology and you know, givin' it kind of a little squinty eye, trying to figure out where it's going, because at the end of the day, with any shift in IT, there's never a time where something is completely sunset, I mean people are still using mainframes today, people will be using mainframes forever, people are just starting their virtualization journey today they're just going from bare metal to VMs, so, and then even with that shift, there's always something that gets left behind, so, they're trying to figure out how can we get used to this new container shift because at the end of the day not everything is gonna be containerized because there's just simply some things that won't be able to or they'll scope out the project and then it'll end up falling by the wayside or budget will go somewhere else. So they're trying to figure out how they can understand the container world from the world that they come from, the VM-centric world, and then, you know, it's really interesting to talk to them and show them how we're able to bring those two together and give you, not only bring the container journey up another step, but also carry your VMs along the way as well. >> Chris, Nutanix is at a, the center of several different transitions, right, both old school hardware to kind of hyper converge, but not now also kind of private hybrid-cloud to more kind of multi-cloud, hybrid-cloud. When we're not at DockerCon, so when you're out in the field, how real is multi-cloud, how real is containers in a normal enterprise? >> Definitely, so, multi-cloud is a very hot topic for sure, everyone, there's no company, no IT department that doesn't have some sort of cloud strategy or analyzing it or looking at it. The main way that we get there, or one of the core tools we have is com once again, so, and I'm obviously biased because that's my wheelhouse, right, in marketing, so I talk about that day in day out, but, with com you can add, we support today AHV and EXSi both on and off Nutanix, as well as AWS, AWS gov cloud and GCP, and Azure's coming in down the line that's where Kubernetes will come in as well, so we see a lot of people looking a this and saying hey you know, we do wanna be able to move into AWS, we do wanna be able to move into GCP and use those clouds or unify them together, and some com lets us do that. There's a couple other of prongs to that as well, one of them is Beam, Nutanix Beam, which is a product we announced at DotNext last month, which is around multi-cloud cost optimization, Beam came from an acquisition that of bought metric--the company was called milinjar, I'm probably saying that horribly wrong, but made a product called bought metric which we've rebranded and are integrating into the platform as Nutanix Beam. So what that allows you to do is, you can, it's provided as a SaaS service, so you can go use it today, there's a trial available, all that, you give it AWS credentials and it reaches out and takes a look at your billing account and says hey, we noticed that these VMs are running 50% of the time at no capacity, or they're not being used at all, you can probably cut that down shrink these and save it or hey we noticed that in general you're using this level, this baseline level, you should buy these in reserved instances to save this much per month. And it presents all that up in a really easy to use interface, and then, depending on how you wanna use it, you can even have it automatically go and resize your VMs for you, so it can say, hey you've got a T2 medium or an M2 medium running, it really would make a lot more sense as a you know M2 small. You can, it'll give you the API call, you can go make it on your own, or you can have, if you give crede-- authorization of course, it can go ahead and run that for you and just downsize those and start saving you that money, so that's another fork of that, the multi-cloud strategy. And the last one is one of the other announcements we made around last month which was around--excuse me extract for VMs, so extract is a portfolio of products, we've got extract for DBs where we can scan your sequel databases and move into ESXi or AHV, both from bare metal, or wherever the sequel databases running, extract for VMs allows us to scan the ESXi VMs, and move them over to AHV. And then, we're taking extract for VMs to the next step and being able to scan your AWS VMs and pull them on, back on-prem, if that's what you're looking for as well, so that's right now in beta and they're working on fine tuning that. Because at the end of the day, it's not just enough to view and manage, we really need to get to someplace where we can move workloads between, and put the workload in the right place. Because really with IT, it's always a balance of tools, there's never one golden bullet that solves every problem, every time a new project comes out you're trying to choose the right tool based on the expertise of the team, based on what tools are already in use, based on policy. So, we wanna be able to make sure that we have the tool sets across, that you can choose and change those choices later on, and always use the right thing for the particular application you're running. >> Choice was a big theme this morning during the general session where Docker was talking about choice agility and security. I'm curious with some of the things that were announced, you know they're talking about the only multi-cloud, multi-OS, multi-Linux, they also were talking about, they announced this federated, containerized application management saying hey, containers have always been portable but management hasn't been. I'm curious what your perspectives are on some of the of the evolution that Docker is announcing today, and how will that help Nutanix customers be able to successfully navigate this container journey? >> Definitely. And--(clears throat) you know federation's critical, being able to, container management in general is always a challenge, one of the things that I've heard time and time again is that getting are back to work for Kubernetes has always been very difficult. (laughs) And so, getting that in there, getting, that is such a basic feature that people expect, you're getting the ability to properly federate roles or federate out authentication is huge. There's a reason that SAML took the world by storm, it's that nobody wants to manage passwords, you wanna rely on some external source of truth, being able to pull that in, being able to use some cloud service and have it federated against having Docker federated against other pieces is very important there. I might've gone way off there, but whatever. (laughing) >> No, no, absolutely. >> And then, the other piece of it is that we, with a multi-cloud, with the idea of it doesn't matter whether you're running on-prem or in the cloud or, that is what people need, that's one of the true promises of containers has always been is the portability, so seeing the delivery of that is huge, and being able to provision it on-prem, on Nutanix obviously because that's who I'm here from. (laughing) but, and being able to provision to the cloud and bring those together, that's huge. >> Chris you talked about Kubernetes couple times now, obviously a big topic here, seems to be kind of emerging de facto application deployment configuration for multi-cloud. What's Nutanix doing with Kubernetes? >> Yeah, so I've definitely, Kubernetes is, it's really in many ways winning that particular battle, I mean don't get me wrong Swarm is great, and the other pieces are great, but, Kubernetes is becoming the de facto standard. One of the things we're working on is bringing containers as a service through Kubernetes, natively on Nutanix, to give you an easy way to manage, through Prism manage containers just the way you manage VMs, manage Kubernetes clusters, and you know it's, it's really important that that's, that is just one solution, because we, there's as many different Kubernetes orchestration engines as you can name, every, any name you bring in, so that's my-- >> It's like Linux, back in the day, they're a lot of different distributions or there're a lot of different ways to consume Kubernetes. >> Exactly. And so, we wanna be able to bring a opinionated way of consuming Kubernetes to the platform natively, just as a, so it's a couple of clicks away, it's very easy to do. But that's not the only way that we're doing it, we're also we do have a partnership with Docker where we're doing things like deploying Docker EE through com, or Docker, it's of course all sorts of legalese but, they're working on that so it's natively in everyone's Prism central you can just one click deploy Docker EE, we have a demo running at our booth deploying rancher using com as well, because we wanna be able to provide whatever set of infrastructure makes the most sense for the customer based on, this is what they've used in the past, this is what they're familiar with, or this is what they want. But we also want to offer an opinionated way to deliver containers as a service so that those of you that don't know, or just trying to get started, or that that's what they're looking for, this, when you've got a thousand choices to make everyone's gonna make slightly different ones. So we can't ever offer one, no one can offer the true, this is the only way to do Kubernetes, we need to offer flexibility across as well. >> One of the words we here all the time at trade shows is flexibility. So, love customer stories, as a customer marketing person, I think there's no greater brand validation you can get than the voice of the customer, and I was looking on the Docker website recently and they were saying: customers that migrate to Docker Enterprise Edition, are actually reducing costs by 50%, so, you're a marketing guy, what're some of your favorite examples of customers where Nutanix is really helping them to just kill it on their container journey? >> Yeah, so, there's a, wish I'd thought of this sooner, I shoulda. (laughing) No, but we have a, one of our customers actually, I, this always brings a smile to my face 'cause they they came and saw us last year at the booth, they're one of our existing long time customers, and they're looking to adopt Docker. They came up and we gave 'em a demo, showed them how all the pieces were doing all of the, and he's just looking at it and he's like man, I need this in my life right now, and it was mostly a demo around Docker EE, using the unified control plane, and showing off, using Nutanix drivers showing how we can back up the data and protect individual components of the containers in a very granular fashion. He's like man I need this in my life, this is incredible, and he went and grabbed his friend ran him over, and was like dude we're already using Nutanix look what they can do! And the perfect example of the two kinds of customers, this guy goes like hold on a second, jumps on the command line, like oh yeah I do this all the time from there. (laughing) >> But, that was the, that light up, the light in the eyes of the customer where they were like, this, I need to be able to see this, to be able to use this, and be able to integrate this, that's, I will not forget that anytime soon. That's really why I think we're going down a very good path there, because the ability to, when you have these tinkerers, the people who are really good at code, I mean I spend a lot of time on the command line myself even though I'm in marketing, so, I don't know what I'm doing there, Powerpoints maybe? (laughing) Just because I can understand it from the command line or an expert can understand it, doesn't mean you can share that. I've been tryin' to hand off some of the gear that I manage off to another person, and was like oh you just type out all these commands, and they're like I have no idea what's going on here. (laughing) And so, seeing the customers be able to, to understand what they're more in depth coworkers have done in a gooey fashion, that's just really, that makes a lot of sense to me and it's, I like that a lot. >> It's great. >> Are you seeing any, and the last question is, as we wrap up, some of the, one of the stats actually that was mentioned in the Docker press release this morning about the new announcements was, 85% of enterprise organizations have multi-cloud, and then we were talking with Scott Johnston, their Chief Product Officer, that said, upwards of 90% of IT budgets are spent on keeping the lights on for existing applications, so, there's a lot of need there for enterprises to go this road. I'm wondering, are you seeing at Nutanix, any particular industries that are really leading edge here saying hey we have a lot of money that we're not able to use for innovation, are you seeing that in any specific industries, or is it kinda horizontal? >> I, to be honest, I've seen it kind of horizontally, I mean I've had, I've spoken to many different customers, mostly around com because, but, and they come from all different walks of life. I've seen, I've talked to customers from sled, who've been really excited about their ability to start better doing hadoop, because they do thousands of hadoop clusters a year for their researchers. I've talked to, you know in the cloud or on-prem, or across. I've talked to people in governments, I've talked to people in hospitals and, you know, all sorts of-- >> I can imagine oil and gas, some of those industries that have a ton of data. >> Yeah and it's actually, the oil and gas is really fascinating because a lot of times they, for in a rig, they wanna be able to use compute, but they can't exactly get to a cloud, so how do you, how do you innovate there and on the edge, without, how do you make a change in the core without making it on the edge, and how do you bring those together? So it's, there's really a lot of really fascinating things happening around that, but, I haven't noticed any one industry in particular it's, it's across, it's that everyone is, but then again, by the time they get to me, it's probably self selected. (laughing) But it's across horizontally, is that everyone is looking at how can we use this vast storage, I just found out this is already being used in my environment because it's super easy, how do I, how do I keep a job? (chuckles) Or how do I adopt this and free up my investments in keeping the lights on into innovation, how do I save time, how do I-- Because one of the things that I've noticed with all of this cloud adoption or container adoption all of that is that many times a customer will start making this push, not always from a low level, maybe from a high level, but, they start making this push because they hear it's faster and better and that it'll just solve all their problems if they just start using this. And, because they rush into they don't often they don't solve the fundamental problems that gave 'em the issue to begin with, and so they're just hoping that this new technology fixes it. So, now there's, I am seeing some customers shift back and say hey, I do wanna adopt that, but I need to do it in a smart way, 'cause we just ran to it and that caused us problems. >> Well it sounds like with all the momentum, John, that we've heard in the keynote, the general session this morning, and with some of the guests, you know, I think even Steve Singh was saying only about half of the audience is actually using containers so it's sounds like, with what you're talking about, with what we've heard consistently today, it's sort of the tip of the iceberg, so lots of opportunity. Chris thank you so much for stopping by theCUBE and sharing with us all the exciting things that are going on at Nutanix with containers and more. >> Thank you so much for having me, it was a lot of fun. >> And we wanna thank you for watching theCUBE, Lisa Martin with John Troyer, from DockerCon 2018 stick around we will be right back with our next guest. (bubbly music)
SUMMARY :
brought to you by Docker the Technical Marketing about in the session, move that to another cloud, they understand who you are, they're starting to figure out, you know, and the ones going I and it's actually been interesting to have the center of several and Azure's coming in down the line of the evolution that one of the things that I've heard and being able to provision it on-prem, seems to be kind of emerging de facto just the way you manage VMs, back in the day, they're a or that that's what customers that migrate to and they're looking to adopt Docker. and was like oh you just and the last question is, as we wrap up, and they come from all that have a ton of data. that gave 'em the issue to begin with, and with some of the guests, you know, Thank you so much for we will be right back with our next guest.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Lisa Martin | PERSON | 0.99+ |
50% | QUANTITY | 0.99+ |
Steve Singh | PERSON | 0.99+ |
John | PERSON | 0.99+ |
Nutanix | ORGANIZATION | 0.99+ |
John Troyer | PERSON | 0.99+ |
Chris Brown | PERSON | 0.99+ |
Chris | PERSON | 0.99+ |
20 minute | QUANTITY | 0.99+ |
San Francisco | LOCATION | 0.99+ |
Scott Johnston | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
90% | QUANTITY | 0.99+ |
last year | DATE | 0.99+ |
85% | QUANTITY | 0.99+ |
ESXi | TITLE | 0.99+ |
two kinds | QUANTITY | 0.99+ |
last month | DATE | 0.99+ |
One | QUANTITY | 0.99+ |
one solution | QUANTITY | 0.99+ |
DockerCon 2018 | EVENT | 0.98+ |
two main main groups | QUANTITY | 0.98+ |
five thousand | QUANTITY | 0.98+ |
Moscone Center | LOCATION | 0.98+ |
two | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
thousands | QUANTITY | 0.98+ |
one | QUANTITY | 0.98+ |
Docker | ORGANIZATION | 0.98+ |
DockerCon | EVENT | 0.98+ |
both | QUANTITY | 0.97+ |
Docker EE | TITLE | 0.97+ |
six thousand people | QUANTITY | 0.96+ |
theCUBE | ORGANIZATION | 0.96+ |
Linux | TITLE | 0.96+ |
Kubernetes | TITLE | 0.96+ |
fifth | QUANTITY | 0.95+ |
DockerCon | ORGANIZATION | 0.95+ |
this year | DATE | 0.93+ |
T2 medium | COMMERCIAL_ITEM | 0.93+ |
DotNext | ORGANIZATION | 0.92+ |
two polar | QUANTITY | 0.91+ |
Scott Johnston, Docker | DockerCon 2018
>> Live from San Francisco, it's theCUBE, covering DockerCon '18, brought to you by Docker and it's ecosystem partners. >> Welcome back to theCUBE, we are live at DockerCon 2018 in San Francisco on a spectacular day. I am Lisa Martin with my with my co-host for the day, John Troyer, and we're very pleased to welcome back to theCUBE a distinguished CUBE alumni and Docker veteran, Steve Johnston, Chief Product Officer at Docker. Welcome back. >> Thank you, thank you very much. That's Scott Johnston but that's okay. >> What did I say? Steve? >> Steve. That's okay. >> Oh, I gave you a new name. >> You know, I get that all the time. >> I'm sorry, Scott. >> That's alright. >> This event, between five and six thousand people. >> Yes. >> You were saying in your general session in keynote this morning, that this is the fifth DockerCon. You started a few years ago with just 300 people and when I was walking out of the keynote this morning, I took a photograph, incredible. People as far as the eye can see. It was literally standing room only. >> It's crazy, right? And you think about four years ago, June 2014 when we did our very first DockerCon, here in San Francisco, 300 people, right? And we've gone from 300 to over 5,000 in that time, grown the community, grown the products, grown the partnerships and it's just, it's very humbling, honestly, to be part of something that's literally industry changing. >> You gave some great numbers during your keynote. You talked about 500 customers using Docker Enterprise Edition. >> Yes. >> Some big names. >> Yes. >> MET Life, Visa, PayPal, McKesson, who was on stage and that was a really interesting. McKesson is what, 183 years old? >> Healthcare company, yeah. >> Talking about data, life and death type of data. >> Right. >> Their transformation working with Docker and containers was really pretty impressive. >> It's exciting that companies get their hands on the technology and they start maybe on a small project or a small team but very quickly they see the potential impact of the solution and very quickly, it's almost infectious inside the organization and more and more teams want to jump on, understand how they can use it to help with their applications, their business to get impact in their operations and it just spreads, spreads like wildflower. That was really the story that McKesson was sharing, just how quickly they were seeing the adoption throughout their org. >> I thought that was really interesting and they did point it out on stage, how that developer adoption did help them go to the next level. >> Yes. >> And kind of transform their whole pipeline. >> Yes. >> Now Scott, you've been here the whole line of time and that through line has been, for Docker, that developer experience. >> That's exactly right. >> Now, as Product Lead here, you've got the Docker Desktop side and the Docker EE side and it's clear, there were some great announcements about desktop here, previews today but how do you balance the enterprise side with the developer centric desktop side and that developer experience idea? >> No, it's a great question, John. I'd reshape it almost to say, it's a continuous platform from developer experience to the operation side and you have to stand back and kind of see it as one and less about trading off one versus the other and how do you create an experience that carries all the way through. So a lot of Gareth's demonstration and the Lily Mason play, was showing how you can create apps in Docker very easily as a developer but those same artifacts that they put their apps in to carry all the way through into production, all the way through into operations. So it's about providing a consistent user experience, consistent set of artifacts that can be used by all the different personas that are building software so that they can be successful moving these Docker applications through the entire application development life cycle. Does that make sense? >> It does, thank you. I'd love to get your perspective, when you're talking with enterprises who might have some trepidation about the container journey, they probably know they have to do it to stay agile and competitive. I think in the press release, I believe it was you, that was quoted saying, "An estimated 85% of enterprise organizations are in a multi Cloud world." >> That's right. >> In a multi Cloud strategy. >> That's right. >> So when you're talking with customers, what's that executive conversation like? C level to C level, what are some of the main concerns that you hear and how influential are the developers in that C suite saying, "Hey guys, we've got to go this direction"? >> No, that's right. That's a great question, Lisa and what we hear again, and again, and again, is a realization going on in the C suite, that having software capabilities is strategic to their business, right? That was not always the case, as much as a decade ago, as recently as a decade ago, inside kind of big manufacturing businesses or big verticals that weren't kind of tech first, IT was a back office, right? It was not front and center but now they're seeing the disruption that software can have in other verticals and they're saying, "Wait a minute, we need to make software capabilities a core capability in our business." And who starts that whole cycle? It's the developers, right? If the developers can integrate with the lines of business, understand their objectives, understand how software can help them achieve those objectives, that's where it kicks off the whole process of, "Okay, we're going to build competitive applications. We then need an operations team to manage and deploy those applications to help us deploy them in a competitive way by taking them to the Cloud." So developers are absolutely pivotal in that conversation and core to helping these very large, Fortune 500, hundred year old companies, transform into new, agile, software driven businesses. >> Modernizing enterprise apps has been a theme >> Yes. >> also at Docker for a few years now. >> Yup. >> Up on stage Microsoft demonstrating the results of a multiyear partnership >> That's right. >> between Microsoft and Docker both with Docker integrating well with Windows server as well as, you talked about, Kubernetes now. >> That's right. >> Can you talk a little bit about what the implications of this are? The demo on stage, of course, was a very old enterprise app written in dot net, with just a few clicks, up and running in the Cloud on Kubernetes no less. >> That's right. >> Managed by Docker, that's actually very cool. You want to talk a little bit about, again, your conversations? >> Absolutely. >> Is this all about Cloud native or how much of your conversations are also supporting enterprise apps? >> Tying back to Lisa's question, so how do we help these organizations get started on their transformation? So they realize they need to transform, where do you start? Well guess what? 90% of their IT budget right now is going into these legacy applications and these legacy infrastructures, so if you start there and it can help modernize what they already have and bring it to modern platforms like Docker and Kubernetes, modern platforms like Window Server 2016, it's a modern operating system, modern platforms like Clouds, that's where you can create a lot of value out of existing application assets, reduce your costs, make these apps agile, even though they're thirteen years old and it's a way for the organization to start to get comfortable with the technology, to adopt it in a surface area that's very well known, to see results very, very quickly and then they gain the confidence to then spread it further into new applications, to spread it further into IOT, to spread it further into big data. But you've got to start it somewhere, right? So the MTA, Modernized Traditional Apps, is a very practical, pragmatic but also high, very quick, return way to get started. >> Oh, go ahead. >> Well I just, the other big announcement involving Kubernetes was managing Kubernetes in the Cloud and I wanted to make sure we hit that. >> That's right, that's right. >> Because I think if people aren't paying attention, they're just going to hear multi Cloud and they're going to go on and say, "Well everybody does multi Cloud, Docker's no different, Docker's just kind of catching up." Actually, this tech preview, I think, is a step forward. I think it's something- >> Thank you. >> I haven't actually seen in practice, so I'm kind of curious, again, how you as an engineering leader make those trade offs. Kind of talk a little bit about what you did and how deciding, "Well there's multi Cloud but the devil's in the details." You actually have integrated now with the native Kubernetes in these three Clouds, EKS, AKS and GKE. >> GKE, no that's right. No, it's a great question, John. The wonderful and fascinating but double edged sword of technology is that the race is always moving the abstraction up, right? You're always moving the abstraction up and you're always having to stay ahead and find where you can create real value for your customers. There was two factors that were going on, that you saw us kind of lean in to that and realize there's an opportunity here. One is, the Cloud providers are doing a wonderful job investing in Kubernetes and making it a manage service on their platforms, great. Now, let's take advantage of that because that's a horizontal infrastructure piece. At parallel we were seeing customers want to take advantage of these different Clouds but getting frustrated that every time they went to a different Cloud they were setting up another stack of process and tooling and automation and management and they're like, "Wait a minute. This is going to slow us down if we have to maintain these stacks." So we leaned in to that and said, "Okay, great. Let's take advantage of commoditized infrastructure, hosting Kubernetes. Let's also then take advantage of our ability to ingest and onboard them into Docker Enterprise Edition, and provide a consistent experience user based APIs, so that the enterprise doesn't get tied into these individual silos of tools, processes and stacks." Really, it's the combination of those two that you see a product opportunity emerged that we leaned heavily into and you saw the fruits of this morning. >> I saw a stat on the docker.com website that said that customers migrating to EE containers can reduce total cost by around 50%? >> Yes. >> That's a significant number. >> It's huge, right? You're reducing your cost of maintaining a ten year old app by 50% and you've made it Cloud portable, and you've made it more secure by putting it in the Docker container than outside and so it's like, "Why wouldn't you invest in that?" It shows a way to get comfortable with the technology, free up some cashflow that then you can pour back into additional innovation, so it's really a wonderful formula. That again, is why we start a lot of customers with their legacy applications because it has these types of benefits that gets them going in other parts of their business. >> And as you mentioned, 90% of an enterprise IT budget is spent keeping the lights on. >> That's right. >> Which means 10% for innovation and as we've talked about before, John, it's the aggressively innovating organizations that are the winners. >> That's exactly right and we're giving them tools, we're giving them a road map even, on how they can become an aggressively innovating organization. >> What about the visibility, in terms of, you know, an organization that's got eight different IT platforms, on prem, public Cloud, hybrid- >> Right. >> What are you doing with respect to being able to deliver visibility across containers and multiple clusters? >> That's right. Well that's a big part of today's announcement, was being able ... Every time we ingest one of these clusters, whether it's on prem, whether it's in the Cloud, whether it's a hosted Kubernetes cluster, that gives us that visibility of now we can manage applications across that, we can aggregate the logging, aggregate the monitoring. You can see, are your apps up, down, are they running out of resources? Do you need to load balance them to another cluster? So it's very much aligned with the vision that we shared on stage, which is fully federated management of the applications across clusters which includes visibility and all the tools necessary for that. >> Scott, I wanted to ask about culture and engineering culture >> Thank you. >> The DockerCon here is very, I think we called it humane in our intro, right? There's childcare on site, there's spoustivities, there's other places to take care of the people who are here and give them a great experience and a lot of training, of course, and things like that. But internally, engineering, there's a war for talent. Docker is very small compared to the Googles of the world but yet you have a very ambitious agenda. The theme of choice today, CLI versus GUI, Kubernetes versus Swarm, Lennox and Windows, not versus, Lennox and Windows, you know and, and, and, and now all these different Clouds and on prem. That's very ambitious and each "and" there takes engineering resources, so I'm kind of curious how the engineering team is growing, how you want to build the culture internally and how you use that to attract the right people? >> Well it certainly helps to be the start up that kicked off this entire movement, right? So a lot of credit to Solomon Hykes, our founder, and the original crew that ... Docker was a Skunkworks project in the previous version of our company and they had the vision to bring it forward and bring it to the world in an opensource model which at the time was a brand new language, go language. That was a catalyst that really got the company off and running in 2013/2014. We're staying true to that in that there's still a very strong opensource culture in the company and that attracts a lot of talent, as well as continuing to balance enterprise features and innovation and you see a combination of that on stage. You're also going to see a wonderful combination of that on the show floor, both from our own employees but also from the community. And I think that's the third dimension, John, which is being humble and call it "aware" that innovation doesn't just come from inside our four walls but that we give our engineers license to bring things in from the outside that add value to their projects. The Kubernetes is a great example of that, right? Our team saw the need for orchestration, we had our own IP in the form of Swarm, but they saw the capabilities of Kubernetes is very complimentary to that, or some customers were preferring to deploy that. So, no ifs, ands or buts, let's take advantage of that innovation, bring it inside the four walls and go. So, it's that kind of flexibility and awareness to attract great engineers who want to work on cutting edge, industry building technologies but also who are aware enough of, there's exciting things happening outside with the community and partnering with that community to bring those into the platform as well. >> So Scott, you guys are doing a lot of collaboration internally, but you're also doing a lot of collaboration with customers. How influential are customers to the development of Docker technologies? >> At ground zero, literally and we have at DockerCon, we call it a customer advisor group, where the customers who have been with us, who have deployed with us in production, we have them. And it's a very select group, it's about twelve to sixteen, and they tell us straight talk in terms of where it's working, where we need to improve. They give us feedback on the road map and so that happens every DockerCon, so that's once every six months. But then we actually have targets inside engineering and product management to be out in the field on a regular basis to make sure we're continuing to get that customer feedback. Innovation's a tricky balance, right? Because you want to be out in front and go where folks aren't asking you to, but you know there's opportunity, at the same time here, where they are today, and make sure you're not getting too far ahead. It's the old joke, Henry Ford, where if he's just listened to his customers, he would have made faster horses but instead he was listening to their problems, their real problems which was transportation and his genius, or his innovation, was to give them the Model T, right? We're trying to balance that ourselves inside Docker. Listen to customers but also know where the innovation, where the technology can take you to give you new solutions, hopefully many of which you saw on stage today. >> We did, well Scott, thanks so much for stopping by theCUBE again and sharing some of the exciting announcements that Docker has made and what you're doing to innovate internally and for the external enterprise community. We appreciate your time. >> Thank you, Lisa. Thank you, John. >> We want to thank you for watching theCUBE. Again, Lisa Martin with John Troyer, live in San Francisco at DockerCon 2018. Stick around, John and I will be right back with our next guest. (upbeat techno music)
SUMMARY :
brought to you by Docker John Troyer, and we're very pleased That's Scott Johnston but that's okay. That's okay. and six thousand people. of the keynote this morning, grown the community, grown the products, You gave some great and that was a really interesting. and death type of data. with Docker and containers of the solution and very quickly, and they did point it out on stage, And kind of transform and that through line and the Lily Mason play, was they probably know they have to do it and core to helping these very large, for a few years now. you talked about, Kubernetes now. Can you talk a little bit that's actually very cool. to get comfortable with the technology, and I wanted to make sure we hit that. and they're going to go on and say, but the devil's in the details." of technology is that the race I saw a stat on the docker.com website in the Docker container than outside is spent keeping the lights on. that are the winners. map even, on how they and all the tools and how you use that to of that on the show floor, a lot of collaboration with customers. and so that happens every DockerCon, and for the external enterprise community. We want to thank you
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Lisa Martin | PERSON | 0.99+ |
John Troyer | PERSON | 0.99+ |
Scott | PERSON | 0.99+ |
Steve | PERSON | 0.99+ |
Steve Johnston | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
John | PERSON | 0.99+ |
Scott Johnston | PERSON | 0.99+ |
Solomon Hykes | PERSON | 0.99+ |
Henry Ford | PERSON | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
San Francisco | LOCATION | 0.99+ |
90% | QUANTITY | 0.99+ |
Docker | ORGANIZATION | 0.99+ |
10% | QUANTITY | 0.99+ |
PayPal | ORGANIZATION | 0.99+ |
50% | QUANTITY | 0.99+ |
Window Server 2016 | TITLE | 0.99+ |
both | QUANTITY | 0.99+ |
two factors | QUANTITY | 0.99+ |
2013/2014 | DATE | 0.99+ |
DockerCon '18 | EVENT | 0.98+ |
DockerCon 2018 | EVENT | 0.98+ |
Skunkworks | ORGANIZATION | 0.98+ |
Docker Enterprise Edition | TITLE | 0.98+ |
MET Life | ORGANIZATION | 0.98+ |
One | QUANTITY | 0.98+ |
CUBE | ORGANIZATION | 0.98+ |
300 | QUANTITY | 0.98+ |
Docker | TITLE | 0.98+ |
eight | QUANTITY | 0.98+ |
300 people | QUANTITY | 0.98+ |
six thousand people | QUANTITY | 0.98+ |
around 50% | QUANTITY | 0.98+ |
today | DATE | 0.97+ |
McKesson | PERSON | 0.97+ |
DockerCon | EVENT | 0.97+ |
Gareth | PERSON | 0.97+ |
June 2014 | DATE | 0.97+ |
five | QUANTITY | 0.97+ |
Visa | ORGANIZATION | 0.97+ |
Windows | TITLE | 0.97+ |
two | QUANTITY | 0.97+ |
three | QUANTITY | 0.97+ |
four years ago | DATE | 0.97+ |
third dimension | QUANTITY | 0.96+ |
over 5,000 | QUANTITY | 0.96+ |
CLI | TITLE | 0.96+ |
McKesson | ORGANIZATION | 0.94+ |
first | QUANTITY | 0.94+ |
a decade ago | DATE | 0.94+ |
183 years old | QUANTITY | 0.94+ |
Googles | ORGANIZATION | 0.94+ |
Kubernetes | TITLE | 0.93+ |
few years ago | DATE | 0.93+ |
ten year old | QUANTITY | 0.92+ |
about twelve | QUANTITY | 0.92+ |
about 500 customers | QUANTITY | 0.91+ |
Lennox | ORGANIZATION | 0.91+ |
each | QUANTITY | 0.9+ |
this morning | DATE | 0.89+ |
Kickoff | DockerCon 2018
>> Live from San Francisco, it's theCUBE, covering DockerCon 18, brought to you by Docker and its ecosystem partners. >> Welcome to theCUBE. We are live in San Francisco at DockerCon 2018. I am Lisa Martin with my co-host for the day, John Troyer. John, it is not only a stunning day in San Francisco, beautiful blue skies, this is a packed event. Their fifth DockerCon event and they've got between 5,000 and 6,000 people. We just came from the general session keynote, and it was standing room only as far as the eyes could see. >> Yeah, looks like a good crowd here, a lot of energy. Docker keynotes, always super interesting, they always do a lot of demos, they bring up a lot of employees. It's not just like a parade of middle-aged executives, always is super dynamic, a lot of demos. Really liked the keynote this morning. >> I did too. The energy you mentioned was great. It kicked off with... who's the name of that gentleman that is one of the rally guys for... >> Franco Finn. >> Franco Finn, who has worked for the Warriors, the 2018 Golden State Warriors, NBA Champs. So that was a great way to kick it off, but also Steve Singh had great energy, their CEO, we're gonna have him on shortly today. Scott Johnston, and as you talked about their employees and also customers. They have some really great numbers. They've got, I think, about 120 sessions this year at DockerCon. Nine big enterprise customers talking about how they are approaching containerization with DockerCon. One of them was McKesson, which is a 183 year old company with a lot of staff that gave a really compelling keynote or a, yeah, a keynote this morning about how they are moving and modernizing their data center with Docker. >> A really nice story, a really an emphasis on trust, an emphasis on developer usability, and I liked one of the points was, once we got the developers using it it became easier, and I think using the whole platform. Lisa, I think they hit a lot of familiar things for Docker: so, developer experience, really big for Docker. That's they way they started, that's what they're still counting on. When Steve Singh got up, he talked about community, their very first thing. Over half the people here, first time at DockerCon and over half of the folks are just using containers in the late last year. That means this whole journey is just starting. There's a lot of white space in the container world. So developer experience, a big announcement, preview announcement for Docker Desktop, being able to create apps off of templates and things like that but very developer-focused shows as opposed to some of the more IT-focused. There's a broad mix here but definitely a lot of developers here at the show. >> A lot of developers, as you said, but also, you're right, it is a mix. It's IT professionals, it's enterprise architects, and it's executives and that's one of the... one of the targeted audiences that, I think, both Steve Singh and Scott Johnston talked about, so it'll be great to explore. As the CEO and the Chief Product Officer respectfully, what are they hearing from enterprise customers who have a lot of challenges with legacy applications that are very difficult to manage and I also read some stats, they had some stats in the press release this morning, but 80% of enterprise IT budgets are spent keeping the lights on for enterprise apps which leaves about 20% for innovation and of course, as we know, organizations that can aggressively innovate are the ones that win. So I'm not only looking forward to hearing with Docker Desktop, what they're doing to make it easy, easier, for developers to get in there and play around on both on Mac and Windows but also the executive conversation. What are they hearing from the executives and where is containerization, you know, from the c-sweep to the board room. >> Yeah, modernizing enterprise apps also has been a Docker theme for the last few years. Microsoft, the big guest up on stage, they've been a multi-year partnership with Microsoft and Docker, putting Docker with Windows together. The big announcement today, pre-technology preview of Kubernetes and Windows Server and the big demo was, they took a very old .net application and, you know, put it up on Kubernetes on Windows with just a couple of clicks. So again, I think that message to the executives is, "You're very safe in Docker's hands "We've got the developer experience covered, "we've got the partnerships." And then going big on Windows, I think choice was another theme that I heard ... >> Yes, it was. Steve talked a lot about choice. >> Um, to the execs here as well, both GUI and CLI, right? A lot of the cloud is very CLI-focused, very Linux-focused. Docker says "We're in on Windows, we support Windows "just as well as Linux so don't hate on the GUI. "You can use a GUI or you can use a CLI." No religion actually too, in terms of Linux versus Windows but Kubernetes, I thought, was a very big. Got mentioned a lot in the keynote this morning, Lisa. >> It did and you talked about choice. One of the things that Steve Singh mentioned from an executive's perspective is, three things that Docker is aiming to deliver. That sounds to me, as a marketer, like competitive differentiation. Talked about choice so that organizations can run apps wherever it makes sense for them, managing applications on any infrastructure, and, as you said earlier, about a few clips, managing their container infrastructures across multiple clouds in just a few clicks. They also talked about being, they also talked a number of times, not just in the press release but also this morning in the keynote, about no vendor lock-in. John, we hear that a lot, it sounds like a marketing term. What are you expecting to hear? What does that mean for Docker? >> I'm not so sure that lock-in is always important for every enterprise, in that any choice you make, it has a certain element of lock-in but it's an active argument or debate online that I see a lot. "Are you locked in when you go to a certain cloud? "Are you locked in when you choose a certain provider," whether it's open-sourced or not. Certainly a lot of Docker is open source. A lot of your choices are protected and they are really trying to say "We're going to be a platform that's going to "service a lot of different abilities to deploy." The big announcement that finished off the keynote was Docker Enterprise Edition can now manage Kubernetes. Not only Kubernetes in the cloud. Kubernetes on Prim, Kubernetes in the cloud managed by Docker, but can actually work with the native Kubernetes cluster managers of the clouds, of the three major clouds: Google GKE, Azure AKS, and AWS EKS. I think I got all those names right. But that's big because a lot of folks say "run anywhere" but they mean "run within our environment anywhere" and what Docker has done in Tech Preview is to connect its platform with the native platforms, orchestration platforms, of the three different clouds so that you can run on Prim, manage via Docker, or you can connect into the cloud's own cluster orchestration. And if they can deliver on that, the devil is in the details, but if they can deliver on that, that's actually a very nice feature to avoid that sort of lock-in. >> And that also goes to, John, one of the major things which is agility and one of the things that they've talked about is, containers today are portable but one of the challenges is that management of containers has not been portable. I think they said that 85% approximately of enterprise I.T. organizations that they has surveyed are running a multi-cloud strategy so they've gotta be able to really deliver this single pane of glass management so they talked about federated application or federated management of containerized applications. I think that's kind of what you're referring to in terms of getting away from the silos and enabling organizations to have that portability and especially as multi-national organizations need to have different access, different security, policies may be maintained across multiple locations. >> Indeed, right. These are global organizations that are betting on container technology. They do need access to be running apps, either parts of apps or services on different clouds. You might be running a Google cloud in Europe, you might be running an AWS here or vice versa. You might have some on-Prim stuff. We've seen a lot of that. I think another theme that we'll hit on, Lisa, along with that multi-cloud portfolio aspect, is the time to value. It's been a theme of this conference season. This last month or two, you and I have both been at a lot of different conference centers and I think time to value, being able to spin up apps within weeks or months that actually work and have value versus the old way, which was years and I think the theme for 2018 is that it's real. People are actually doing it and we'll talk to a couple of customers, I hope, today. >> And that's essential because enterprises, while there's still trepidation with moving into the container journey, they don't really have a choice to be able to aggressively innovate to be able to be leaders and compete with these cloud-native organizations. They don't have the luxury of time to rip and replace old enterprise applications and put them on a container or a micro-service's space archicture, they've got to be able to leverage something like containers to maximize time to value to deliver differentiating services. >> Absolutely. I'm very interested in being here today and we'll see what the day brings us. >> I think we're gonna have a lot of fun today, John. I think they kicked off things with great energy. I loved how, you know, they always do demos, right, on main stage during general sessions, and we were at SAP last week and of course, one of the demos didn't work. That's just the nature of trying to do things live. I liked how they were very cheeky with the praying to the demo Gods with the fortune cookies. I thought that was really good but the demos were simple. They were very clearly presented and I'm excited with you to dig in to what are they doing. Also what is setting them apart and how are they enabling enterprise organizations like MetLife, like McKesson, PayPal, Splunk to be able to transform to compete. >> Absolutely. One last thing about the conference, Lisa, is I do want to call out. It's a very humane conference. Not only do they have kind of a cheeky sense of humor here at Docker, but there's child care onsite, and there's spouse-tivities, there are activities for if you bring your spouse or family to the conference. They're trying to do a lot of things to make the conference experience good and successful and friendly and humane for people here at the show which I really appreciate. >> I like that, humane conference. You're right. We don't always see that. Well, John and I are going to be here all day talking with Docker executives, customers, partners and we're excited to have you with us. Lisa Martin for John Troyer. You're watching theCUBE at DockerCon 2018. We'll be right back with our first guest. (techno music)
SUMMARY :
brought to you by Docker We just came from the Really liked the keynote this morning. that is one of the rally guys for... Scott Johnston, and as you and I liked one of the points was, from the c-sweep to the board room. and the big demo was, they took Yes, it was. A lot of the cloud is very One of the things that of the three major clouds: and one of the things that is the time to value. They don't have the luxury of time and we'll see what the day brings us. but the demos were simple. for people here at the show and we're excited to have you with us.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Steve | PERSON | 0.99+ |
Steve Singh | PERSON | 0.99+ |
Franco Finn | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
MetLife | ORGANIZATION | 0.99+ |
John Troyer | PERSON | 0.99+ |
John | PERSON | 0.99+ |
PayPal | ORGANIZATION | 0.99+ |
Europe | LOCATION | 0.99+ |
Scott Johnston | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
San Francisco | LOCATION | 0.99+ |
2018 | DATE | 0.99+ |
80% | QUANTITY | 0.99+ |
Splunk | ORGANIZATION | 0.99+ |
McKesson | ORGANIZATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
85% | QUANTITY | 0.99+ |
San Francisco | LOCATION | 0.99+ |
Scott Johnston | PERSON | 0.99+ |
last week | DATE | 0.99+ |
Windows | TITLE | 0.99+ |
today | DATE | 0.99+ |
Docker | ORGANIZATION | 0.99+ |
Warriors | ORGANIZATION | 0.99+ |
Linux | TITLE | 0.99+ |
DockerCon 2018 | EVENT | 0.99+ |
one | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
first time | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
DockerCon 18 | EVENT | 0.99+ |
late last year | DATE | 0.98+ |
first thing | QUANTITY | 0.98+ |
Kubernetes | TITLE | 0.97+ |
ORGANIZATION | 0.97+ | |
6,000 people | QUANTITY | 0.97+ |
this year | DATE | 0.97+ |
first guest | QUANTITY | 0.97+ |
DockerCon | EVENT | 0.96+ |
183 year old | QUANTITY | 0.96+ |
Docker | TITLE | 0.96+ |
about 20% | QUANTITY | 0.96+ |
fifth | QUANTITY | 0.96+ |
5,000 | QUANTITY | 0.96+ |
about 120 sessions | QUANTITY | 0.95+ |
Over half the people | QUANTITY | 0.94+ |
CLI | TITLE | 0.94+ |
this morning | DATE | 0.94+ |
NBA Champs | EVENT | 0.93+ |
Bradley Wong, Docker & Kiran Kamity, Cisco - DockerCon 2017 - #theCUBE - #DockerCon
>> Narrator: From Austin, Texas, it's theCUBE covering DockerCon 2017, brought to you by Docker and support from it's ecosystem partners. (upbeat music) >> Hi, and we're back, I'm Stu Miniman, and this is SilconANGLES production of the Cube, here at DockerCon 2017, Austin, Texas. Happy to have on the program Kiran Kamity, who was CEO of ContainerX which was acquired by Cisco. And you're currently the senior director and head of container products at Cisco. And also joining us is Brad Wong, who is the director of product management at Docker. Gentlemen, thank you so much for joining us. >> Brad: Thanks for having us. [Kiran] Thank you, Stu. >> So Kiran, talk a little bit about ContainerX, you know, bring us back to, why containers, you know why you help start a company with containers, and when to be acquired by a big company like Cisco. >> Yeah, it was actually late 2014 is when Pradeep and I, my co-founder from ContainerX, we started brainstorming about, you know, what do we do in the space and the fact that the space was growing, and my previous company called RingCube, which has sold to Citrix, where we had actually built a container between 2006 and 2010. So we wanted to build a management platform for containers, and it was in a way there was little bit of an overlap with Docker Datacenter, but we were focusing on mostly tendency aspects of it. Bringing in concepts like viamordi rs into containers et cetera. And we were acquired by Cisco about eight months ago now, and the transition in the last eight months has been fantastic. >> Great, and Brad, you're first time on the cube, so give us your background, what brought you to Docker? >> Yeah, so actually before Docker I was at actually, a veteran of Cisco, interestingly enough. Many different ventures in Cisco, most recently I was actually part of the Insieme Networks team, focusing on the software defined networking, and Application Centric Infrastructure. Obviously I saw a pretty trend in the infrastructure space, that the future of infrastructure is being led by applications and developers. With that I actually got to start digging around with Docker quite a lot, found some good interest, and we started talking, and essentially that's how I ended up at Docker, to look at our partner ecosystem, how we can evolve that. Two years ago now, actually. >> I think two years ago Docker networking was a big discussion point. Cisco's been a partner there, but bring us up to speed if you would, both of you, on where you're engaging, on the engineering side, customer side, and the breadth and depth of what you're doing. >> You're right, two years ago, networking was in quite a different place. We kicked it off with acquiring a company back then called SocketPlane, which helped us really define-- >> Yeah and we know actually, ---- and ----, two alums, actually I know those guys, from the idea to starting the company, to doing acquisition was pretty quick for you and for them. >> Right, and we felt that we really needed to bring on board a good solid networking DNA into the company. We did that, and they helped us define what a successful model would be for networking which is why they came up with things like the container networking model, and live network, which then actually opened the door for our partners to then start creating extensions to that, and be able to ride on top of that to offer more advanced networking technologies like Contiv for example. >> Contiv was actually an open source project that was started within Cisco, even before the container was acquisitioned. Right after the acquisition happened, that team got blended into our team and we realized that there were some really crown jewels in Contiv that we wanted to productize. We've been working with Docker for the last six months now trying to productize that, and we went from alpha to beta to g a. Now Contiv is g a today, and it was announced in a blog post today, and it's actually 100% open-source networking product that Cisco TAC and Cisco advanced services have offered commercial support and services support. It's actually a unique moment, because this is the fist 100% open-source project that Cisco TAC has actually offered commercial support for, so it's a pretty interesting milestone I think. >> I think also with that, we also have it available on Docker store as well. It's actually the first Docker networking plug-in that it's been certified as well. We're pretty also happy to have that on there as well. >> Yeah. >> Anything else for the relationship we want to go in beyond those pieces? >> We also saw that there was a lot of other great synergies between the two companies as well. The first thing we wanted to do was to look at how we can also make it a lot better experience for joint customers to get Docker up and running, Docker Enterprise Edition up and running on infrastructure, specifically on Cisco infrastructure, so Cisco UCS. So we also kicked off a series of activities to test and validate and document how Docker Enterprise Edition can run on Cisco UCS, Nexus platforms, et cetera. We went ahead with that and a couple months later we brought out, jointly, to our Cisco validated designs for Docker Enterprise Edition. One on Cisco UCS infrastructure alone, and the other one jointly with NetApp as well, with the FlexPod Solution. So we're also very very happy with that as well. >> Great. Our community I'm sure knows the CVD's from what they are out there. UCS was originally designed to be the infrastructure for virtualized environments. Can you walk me through, what other significant differences there or anything kind of changing to move to containers versus what UCS for virtualized environment. >> The goal with that, UCS is esentially considered a premium kind of infrastructure server infrastructure for our customers. Not only can they run virtual environments today, but our goal is as containers become mainstreamed, containers evolved to being a first-class citizen alongside VM. We have to provide our customers with a solution that they need. And a turnkey solution from a Cisco standpoint is to take something like a Docker stack, or other stacks that our customer stopped, such as Kubernetes or other stacks as well, and offer them turnkey kind of experience. So with Docker Data Center what we have done is the CVD that we've announced so far has Docker Data Center, and the recipe provides an easy way for customers to get started with USC on Docker Data Center so that they get that turnkey experience. And with the MTA program that was announced, today at the key note. So that allows Cisco and Docker to work even more closely together to have not just the products, but also provide services to ensure that customers can completely sort of get started very very easily with support from advanced services and things like that. >> Great, I'm wondering if you have any customer examples that you can talk through. If you can't talk about a specific, logo, maybe you can talk about. Or if there are key verticals that you see that you're engaging first, or what can you share? >> We've been working joint customer evals, actually a couple of them. Once again I don't think we can point out the names yet. We haven't fully disclosed, or cleared it with their Prs Definitely into financials. Especially the online financials, a significant company that we've been working with jointly that has actually adopted both Contiv, and is actually seeing quite a lot of value in being able to take Docker, and also leverage the networking stack that Contiv provides. And be able to not just orchestrate networking policies for containers, but the other thing that they want to do is to have those same policies be able to run on cloud infrastructure, like EWS for example. So they obviously see that Docker is a great platform to be enable their affordability between on premises and also public cloud. But at the same time be able to leverage these kind of tools that makes that transition, and makes that move a lot easier so they don't have to re-think their security networking policies all over again. That's been actually a pretty used case I thought of the joint work that we did together with Contiv. >> Some of the customers that we've been talking to in fact we have one customer that I don't think I'm supposed say the name just yet, but we've drollled it out, has drolled out Contiv with the Docker on time. In five production data centers already. And these are the kind of customers that actually take to advanced networking capabilites that Contiv offers so that they can comprehensive L2 networking, L3 networking. Their monitoring pools that they currently use will be able to address the containers, because the L2, the L3 networking capabilities allows each container to have an IP address that is externally addressable, so that the current monitoring tools that you use for VMs et cetera can completely stay relevant, and be applicable in the container world. If you have an ACI fabric that continues to work with containers. So those are some of the reasons why these customers seem to like it. >> Kiran, you're relatively new into Cisco, and you were a software company. Many people they still think of Cisco as a networking company. I've heard people derogatory it's like, "Oh they made hardware define networking when they rolled out some of this stuff." Tell us about, you talk about an open source project that you guys are doing. I've talked to Lou Tucker a number of times. I know some of the software things you guys are doing. Give us your viewpoint as to your new employer, and how they might be different than people think of as the Cisco that we've known for decades. >> Cisco is, has of course it has, you know, several billion dollars of revenue coming in from hardware and infrastructure. And networking and security have been the bread and the butter for the company for many many years now But as the world moves to Cloud-Native becoming a first class citizen, the goal is really to provide complete solutions to our customers. And if you think of complete solutions, those solutions include things like networking, thing like security. Including analytics, and complete management platforms. At the same time, at the end of the day, the customers want to come to peace with the fact that this is a multi-cloud world Customers have data centers on premises, or on hosted private cloud environments. They have workloads that are running on public clouds. So with products like cloud center, our goal is to make sure that whatever they, the applications that they have, can be orchestrated across these multiple clouds. We want to make sure that the pain points the customers have around deploying whole solutions include easy set-up of products on infrastructure that they have, and that includes partnerships like UCS, or running on ACI or Nexus. We want to make sure that we give that turnkey experience to these customers. We want to make sure that those workloads can be moved across and run across these different clouds. That's where products like cloud center come in. We want to make sure that these customers have top grade analytics, which is completely software. That's were the app dynamics acquisition comes in. And we want to make sure that we provide that turnkey experience with support in terms of services. With our massive services organization, partners, et cetera. We view this as our job is to provide our customers what they need in terms of the end solution that they're looking for. And so it's not just hardware, it's just a part of it. Software, services, et cetera, complimented. >> Alright, Brad last question that I have for you in the keynote yesterday, I couldn't count how many times the word ecosystem was used. I think it was loud and clear that everybody there I think it was like, you know, Docker will not be successful unless it's partners are successful, kind of vice versa. When you look at kind of the product development piece of things, how does that resonate with you and the job that you're doing? >> We basically are seeing Docker become more of a, more and more of a platform as evidenced by yesterdays keynote. Every platform, the only way that platform's going to be successful is if we can do great, we have great options for our partners, like Cisco, to be able to integrate with us on multiple different levels, not just on one place. The networking plug-in is just one example. Many many other places as well Yesterday we announced two new open source initiatives. Lennox kit and also the movi project. You can imagine that there's probably lots of great places where partners like Cisco can actually play in there, not just only in the service fees, but maybe also in things like IOT as well, which is also a fast-emerging place for us to be. And all the way up until day two type of monitoring, type of environment as well where we think there's a lot of great places where once again, options like app dynamics, tetration analytics can fit in quite nicely with how do you take applications that have been migrated or modernized into containers, and start really tracking those using a common tool set. So we think that's really really good opportunities for our ecosystem partners to really innovate in those spaces, and to differentiate as well. >> Kiran, I want to give you the final word, take-aways that you want the users here, and those out watching the show to know about, you know, Cisco, and the Docker environment. >> I want to let everybody know that Cisco is not just hardware. Our goal is to provide turnkey complete solutions and experiences to our customers. And as they walk through this journey of embracing Cloud-Native workloads, and containerized workload there's various parts of the problem, that include all the way from hardware, to running analytics, to networking, to security, and services help, and Cisco as a company is here to offer that help, and make sure that the customers can walk away with turnkey solutions and experiences. >> Kiran and Brad, thank you so much for joining us. We'll be back with more coverage here. Day two, DockerCon 2017, you're watching theCube.
SUMMARY :
covering DockerCon 2017, brought to you by Docker and head of container products at Cisco. Brad: Thanks for having us. and when to be acquired by a big company like Cisco. and the fact that the space was growing, that the future of infrastructure and the breadth and depth of what you're doing. We kicked it off with acquiring a company back then from the idea to starting the company, and be able to ride on top of that and we realized that there were some really crown jewels in We're pretty also happy to have that on there as well. and the other one jointly with NetApp as well, there or anything kind of changing to move to containers and the recipe provides an easy way for customers that you can talk through. and also leverage the networking stack that Contiv provides. so that the current monitoring tools that you use for I know some of the software things you guys are doing. the goal is really to provide complete solutions and the job that you're doing? and to differentiate as well. take-aways that you want the users here, and make sure that the customers can walk away with Kiran and Brad, thank you so much for joining us.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Brad | PERSON | 0.99+ |
Cisco | ORGANIZATION | 0.99+ |
Kiran | PERSON | 0.99+ |
Kiran Kamity | PERSON | 0.99+ |
Brad Wong | PERSON | 0.99+ |
Lou Tucker | PERSON | 0.99+ |
100% | QUANTITY | 0.99+ |
two companies | QUANTITY | 0.99+ |
Contiv | ORGANIZATION | 0.99+ |
ContainerX | ORGANIZATION | 0.99+ |
2006 | DATE | 0.99+ |
Docker | ORGANIZATION | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
Austin, Texas | LOCATION | 0.99+ |
RingCube | ORGANIZATION | 0.99+ |
2010 | DATE | 0.99+ |
both | QUANTITY | 0.99+ |
late 2014 | DATE | 0.99+ |
Stu | PERSON | 0.99+ |
SocketPlane | ORGANIZATION | 0.99+ |
EWS | ORGANIZATION | 0.99+ |
first | QUANTITY | 0.99+ |
DockerCon 2017 | EVENT | 0.99+ |
today | DATE | 0.99+ |
#DockerCon | EVENT | 0.99+ |
each container | QUANTITY | 0.99+ |
Two years ago | DATE | 0.99+ |
one customer | QUANTITY | 0.99+ |
two years ago | DATE | 0.99+ |
UCS | ORGANIZATION | 0.99+ |
ACI | ORGANIZATION | 0.99+ |
yesterday | DATE | 0.98+ |
Citrix | ORGANIZATION | 0.98+ |
Cisco TAC | ORGANIZATION | 0.98+ |
USC | ORGANIZATION | 0.98+ |
one example | QUANTITY | 0.98+ |
Yesterday | DATE | 0.98+ |
Docker | TITLE | 0.98+ |
one place | QUANTITY | 0.98+ |
Docker Enterprise Edition | TITLE | 0.98+ |
Day two | QUANTITY | 0.97+ |
two alums | QUANTITY | 0.97+ |
Pradeep | PERSON | 0.97+ |
yesterdays | DATE | 0.97+ |
Insieme Networks | ORGANIZATION | 0.97+ |
five production data centers | QUANTITY | 0.97+ |
One | QUANTITY | 0.97+ |