Image Title

Search Results for Storage Day 2022:

Ashish Palekar & Cami Tavares, AWS | AWS Storage Day 2022


 

(upbeat music) >> Okay, we're back covering AWS Storage Day 2022 with Ashish Palekar. Who's the general manager of AWS EBS Snapshot and Edge and Cami Tavares. Who's the head of product at Amazon EBS. Thanks for coming back in theCube guys. Great to see you again. >> Great to see you as well, Dave. >> Great to see you, Dave. Ashish, we've been hearing a lot today about companies all kinds of applications to the cloud and AWS and using their data in new ways. Resiliency is always top of mind for companies when they think about just generally their workloads and specifically the clouds. How should they think about customers think about data resiliency? >> Yeah, when we think about data resiliency it's all about making sure that your application data, the data that your application needs is available when it needs it. It's really the ability for your workload to mitigate disruptions or recover from them. And to build that resilient architecture you really need to understand what kinds of disruptions your applications can experience. How broad the impact of those disruptions is, and then how quickly you need to recover. And a lot of this is a function of what the application does, how critical it is. And the thing that we constantly tell customers is, this works differently in the cloud than it does in a traditional on-premises environment. >> What's different about the cloud versus on-prem? Can you explain how it's different? >> Yeah, let me start with a video on-premises one. And in the on-premises one, building resilient architectures is really the customer's responsibility, and it's very challenging. You'll start thinking about what your single points of failure are. To avoid those, you have to build in redundancy, you might build in replication as an example for storage and doing this now means you have to have provision more hardware. And depending on what your availability requirements are, you may even have to start looking for multiple data centers, some in the same regions, some in different geographical locations. And you have to ensure that you're fully automated, so that your recovery processes can take place. And as you can see that's a lot of owners being placed on the customer. One other thing that we hear about is really elasticity and how elasticity plays into the resiliency for applications. As an example, if you experience a sudden spike in workloads, in a on-premises environment, that can lead to resource saturation. And so really you have two choices. One is to sort of throttle the workload and experience resiliency, or your second option becomes buying additional hardware and securing more capacity and keeping it fair low in case of experiencing such a spike. And so your two propositions that are either experiencing resiliency, challenges or paying really to have infrastructure that's lying around. And both of those are different really when you start thinking about the cloud. >> Yeah, there's a third option too, which is lose data, which is not an option. Go ahead- >> Which is not, yeah, I pretty much as a storage person, that is not an option. The reason about that that we think is reasonable for customers to take. The big contrast in the cloud really comes with how we think about capacity. And fundamentally the the cloud gives you that access to capacity so you are not managing that capacity. The infrastructure complexity and the cost associated with that are also just a function of how infrastructure is built really in the cloud. But all of that really starts with the bedrock of how we design for avoiding single points of failure. The best way to explain this is really to start thinking about our availability zones. Typically these availability zones consist of multiple data centers, located in the same regional area to enable high throughput and low latency for applications. But the availability zones themselves are physically independent. They have independent connections to utility power, standalone backup power resources, independent mechanical services and independent network connectivity. We take availability zone independence extremely seriously, so that when customers are building the availability of their workload, they can architect using these multiple zones. And that is something that when I'm talking to customers or Tami is talking to customers, we highly encourage customers to keep in mind as they're building resiliency for their applications. >> Right, so you can have within an availability zone, you can have, you know, instantaneous, you know when you're doing it right. You've got, you've captured that data and you can asynchronously move to outside of that in case there's, the very low probability, but it does happen, you get some disasters. You're minimizing that RPO. And I don't have to worry about that as a customer and figuring out how to do three site data centers. >> That's right. Like that even further, now imagine if you're expanding globally. All those things that we described about like creating new footprint and creating a new region and finding new data centers. As a customer in an on-premises environment, you take that on yourself. Whereas with AWS, because of our global presence, you can expand to a region and bring those same operational characteristics to those environments. And so again, bringing resiliency as you're thinking about expanding your workload, that's another benefit that you get from using the availability zone region architecture that AWS has. >> And as Charles Phillips, former CEO of Infor said, "Friends, don't let friends build data center," so I don't have to worry about building the data center. Let's bring Cami into the discussion here. Cami, think about elastic block storage, it gives, you know customers, you get persistent block storage for EC2 instances. So it's foundational for any mission critical or business critical application that you're building on AWS. How do you think about data resiliency in EBS specifically? I always ask the question, what happens if something goes wrong? So how should we think about data resiliency in EBS specifically? >> Yeah, you're right Dave, block storage is a really foundational piece. When we talk to customers about building in the cloud or moving an application to the cloud, and data resiliency is something that comes up all the time. And with EBS, you know EBS is a very large distributed system with many components. And we put a lot of thought and effort to build resiliency into EBS. So we design those components to operate and fail independently. So when customers create an EBS volume for example, we'll automatically choose the best storage nodes to address the failure domain and the data protection strategy for each of our different volume types. And part of our resiliency strategy also includes separating what we call a volume life cycle control plane. Which are things like creating a volume, or attaching a volume to an EC2 instance. So we separate that control plane, from the storage data plane, which includes all the components that are responsible for serving IO to your instance, and then persisting it to durable media. So what that means is once a volume is created and attached to the instance, the operations on that volume they're independent from the control point function. So even in the case of an infrastructure event, like a power issue, for example, you can recreate an EBS volume from a snapshot. And speaking of snapshots, that's the other core pillar of resiliency in EBS. Snapshots are point in time copies of EBS volumes that would store in S3. And snapshots are actually a regional service. And that means internally we use multiple of the availability zones that Ashish was talking about to replicate your data so that the snapshots can withstand the failure of an availability zone. And so thanks to that availability zone independence, and then this builtin component independence, customers can use that snapshot and recreate an EBS following another AZO or even in another region if they need to. >> Great so, okay, so you touched on some of the things EBS does to build resiliency into the service. Now thinking about over your right shoulders, you know, Joan Deviva, so what can organizations do to build more resilience into their applications on EBS so they can enjoy life without anxiety? >> (laughs) That is a great question. Also something that we love to talk to customers about. And the core thing to think about here is that we don't believe in a one size fits all approach. And so what we are doing in EBS is we give customers different tools so that they can design a resiliency strategy that is custom tailored for their data. And so to do this, this resiliency assessment, you have to think about the context of this specific workload and ask questions like what other critical services depend on this data and what will break if this data's not available and how long can can those systems withstand that, for example. And so the most important step I'll mention it again, snapshots, that is a very important step in a recovery plan. Make sure you have a backup of your data. And so we actually recommend that customers take the snapshots at least daily. And we have features that make that easier for you. For example, Data Lifecycle Manager which is a feature that is entirely free. It allows you to create backup policies, and then you can automate the process of creating the snapshot, so it's very low effort. And then when you want to use that backup to recreate a volume, we have a feature called Fast Snapshot Restore, that can expedite the creation of the volume. So if you have a more, you know a shorter recovery time objective you can use that feature to expedite the recovery process. So that's backup. And then the other pillar we talked to customers about is data replication. Just another very important step when you're thinking about your resiliency and your recovery plans. So with EBS, you can use replication tools that work at the level of the operating system. So that's something like DRBD for example. Or you can use AWS Elastic Disaster Recovery, and that will replicate your data across availability zones or nearby regions too. So we talked about backup and replication, and then the last topic that we recommend customers think about is having a workload monitoring solution in place. And you can do that in EBS, using cloud watch metrics. So you can monitor the health of your EBS volume using those metrics. We have a lot of tips in our documentation on how to measure that performance. And then you can use those performance metrics as triggers for automated recovery workflows that you can build using tools like auto scaling groups for example. >> Great, thank you for that advice. Just quick follow up. So you mentioned your recommendation, at least daily, what kind of granularity, if I want to compress my RPO can I go at a more granular level? >> Yes, you can go more granular and you can use again the daily lifecycle manager to define those policies. >> Great, thank you. Before we go, I want to just quickly cover what's new with EBS. Ashish, maybe you could talk about, I understand you've got something new today. You've got an announcement, take us through that. >> Yeah, thanks for checking in and I'm so glad you asked. We talked about how snapshots help resilience and are a critical part of building resilient architectures. So customers like the simplicity of backing up their EC2 instances, using multi volume snapshots. And what they're looking for is the ability to back up only to exclude specific volumes from the backup, especially those that don't need backup. So think of applications that have cash data, or applications that have temporary data that really doesn't need backup. So today we are adding a new parameter to the create snapshots API, which creates a crash consistent set of snapshots for volumes attached to an EC2 instance. Where customers can now exclude specific volumes from an instance backup. So customers using data life cycle manager that can be touched on, can automate their backups. And again they also get to exclude these specific volumes. So really the feature is not just about convenience, but it's also to help customers save on cost. As many of these customers are managing tens of thousands of snapshots. And so we want to make sure they can take it at the granularity that they need it. So super happy to bring that into the hands of customers as well. >> Yeah, that's a nice option. Okay, Ashish, Cami thank you so much for coming back in theCube, helping us learn about what's new and what's cool and EBS, appreciate your time. >> Thank you for having us Dave. >> Thank you for having us Dave. >> You're very welcome now, if you want to learn more about EBS resilience, stay right here because coming up, we've got a session which is a deep dive on protecting mission critical workloads with Amazon EBS. Stay right there, you're watching theCube's coverage of AWS Storage Day 2022. (calm music)

Published Date : Aug 12 2022

SUMMARY :

Great to see you again. and specifically the clouds. And the thing that we And so really you have two choices. option too, which is lose data, to capacity so you are not and you can asynchronously that you get from using so I don't have to worry about And with EBS, you know EBS is a very large of the things EBS does And the core thing to So you mentioned your and you can use again the Ashish, maybe you could is the ability to back up only you so much for coming back if you want to learn more

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
AshishPERSON

0.99+

Ashish PalekarPERSON

0.99+

AWSORGANIZATION

0.99+

Charles PhillipsPERSON

0.99+

Joan DevivaPERSON

0.99+

DavePERSON

0.99+

CamiPERSON

0.99+

third optionQUANTITY

0.99+

EBSORGANIZATION

0.99+

two propositionsQUANTITY

0.99+

second optionQUANTITY

0.99+

OneQUANTITY

0.99+

InforORGANIZATION

0.99+

Cami TavaresPERSON

0.99+

bothQUANTITY

0.99+

todayDATE

0.98+

two choicesQUANTITY

0.98+

EBSTITLE

0.97+

EC2TITLE

0.97+

TamiPERSON

0.96+

tens of thousands of snapshotsQUANTITY

0.95+

eachQUANTITY

0.95+

AZOTITLE

0.93+

Amazon EBSORGANIZATION

0.91+

theCubeORGANIZATION

0.89+

AshishORGANIZATION

0.89+

single pointsQUANTITY

0.86+

three siteQUANTITY

0.83+

single pointsQUANTITY

0.82+

DRBDTITLE

0.8+

Storage Day 2022EVENT

0.78+

one sizeQUANTITY

0.76+

Elastic DisasterTITLE

0.7+

EdgeORGANIZATION

0.68+

CEOPERSON

0.63+

LifecycleTITLE

0.59+

thingQUANTITY

0.57+

SnapshotTITLE

0.49+

S3TITLE

0.46+

Edward Naim, AWS | AWS Storage Day 2022


 

[Music] welcome back to aws storage day 2022 i'm dave vellante and we're pleased to have back on thecube edname the gm of aws file storage ed how you doing good to see you i'm good dave good good to see you as well you know we've been tracking aws storage for a lot of years 16 years actually we we've seen the evolution of services of course we started with s3 and object and saw that expand the block and file and and now the pace is actually accelerating and we're seeing aws make more moves again today and block an object but what about file uh it's one format in the world and the day wouldn't really be complete without talking about file storage so what are you seeing from customers in terms of let's start with data growth how are they dealing with the challenges what are those challenges if you could address you know specifically some of the issues that they're having that would be great and then later we're going to get into the role that cloud file storage plays take it away well dave i'm definitely increasingly hearing customers talk about the challenges in managing ever-growing data sets and they're especially challenged in doing that on-premises when we look at the data that's stored on premises zettabytes of data the fastest growing data sets consist of unstructured data that are stored as files and many cups have tens of petabytes or hundreds of petabytes or even exabytes of file data and this data is typically growing 20 30 percent a year and in reality on-premises models really designed to handle this amount of data in this type of growth and i'm not just talking about keeping up with hardware purchases and hardware floor space but a big part of the challenge is labor and talent to keep up with the growth they're seeing companies managing storage on-prem they really need an unprecedented number of skilled resources to manage the storage and these skill sets are in really high demand and they're in short supply and then another big part of the challenge that customers tell me all the time is that that operating at scale dealing with these ever-growing data sets at scale is really hard and it's not just hard in terms of the the people you need and the skill sets that you need but operating at scale presents net new challenges so for example it becomes increasingly hard to know what data you have and what storage media your data stored on when you have a massive amount of data that's spanning hundreds of thousands or uh thousands of applications and users and it's growing super fast each year and at scale you start seeing edge technical issues get triggered more commonly impacting your availability or your resiliency or your security and you start seeing processes that used to work when you were a much smaller scale no longer work it's just scale is hard it's really hard and then finally companies are wanting to do more with their fast growing data sets to get insights from it and they look at the machine learning and the analytics and the processing services and the compute power that they have at their fingertips on the cloud and having that data be in silos on-prem can really limit how they get the most out of their data you know i've been covering glad you brought up the skills gap i've been covering that quite extensively with my colleagues at etr you know our survey partner so that's a really important topic and we're seeing it across the board i mean really acute in cyber security but for sure just generally in i.t and frankly ceos they don't want to invest in training people to manage storage i mean it wasn't that long ago that managing loans was a was a talent and that's of course nobody does that anymore but they'd executives would much rather apply skills to get value from data so my specific question is what can be done what is aws doing to address this problem well with the growth of data that that we're seeing it it's just it's really hard for a lot of it teams to keep up with just the infrastructure management part that's needed so things like deploying capacity and provisioning resources and patching and conducting compliance reviews and that stuff is just table stakes the asks on these teams to your point are growing to be much bigger than than those pieces so we're really seeing fast uptake of our amazon fsx service because it's such an easy path for helping customers with these scaling challenges fsx enables customers to launch and to run and to scale feature rich and highly performant network attached file systems on aws and it provides fully managed file storage which means that we handle all of the infrastructure so all of that provisioning and that patching and ensuring high availability and customers simply make api calls to do things like scale up their storage or change their performance level at any point or change a backup policy and a big part of why fsx has been so feeling able to customers is it really enables them to to choose the file system technology that powers their storage so we provide four of the most popular file system technologies we provide windows file server netapp ontap open zfs and luster so that storage and application admins can use what they're familiar with so they essentially get the full capabilities and even the management clis that they're used to and that they've built workflows and applications around on-premises but they get along with that of course the benefits of fully managed elastic cloud storage that can be spin up and spun spin down and scaled on demand and performance changed on demand etc and what storage and application admins are seeing is that fsx not only helps them keep up with their scale and growth but it gives them the bandwidth to do more of what they want to do supporting strategic decision making helping their end customers figure out how they can get more value from their data identifying opportunities to reduce cost and what we realize is that for for a number of storage and application admins the cloud is is a different environment from what they're used to and we're making it a priority to help educate and train folks on cloud storage earlier today we talked about aws storage digital badges and we announced a dedicated file badge that helps storage admins and professionals to learn and demonstrate their aws skills in our aws storage badges you can think of them as credentials that represent cloud computing learning that customers can add to their repertoire add to their resume as they're embarking on this cloud journey and we'll be talking more in depth on this later today especially around the file badge which i'm very excited about so a couple things there that i wanted to comment on i mean i was there for the netapp you know your announcement we've covered that quite extensively this is just shows that it's not a zero-sum game necessarily right it's a win-win-win for customers you've got your you know specific aws services you've got partner services you know customers want want choice and then the managed service model you know to me is a no-brainer for most customers we learned this in the hadoop years i mean it just got so complicated then you saw what happened with the managed services around you know data lakes and lake houses it's just really simplified things for customers i mean there's still some customers that want to do it yourself but a managed service for the file storage sounds like a really easy decision especially for those it teams that are overburdened as we were talking about before and i also like you know the education component is nice touch too you get the badge thing so that's kind of cool so i'm hearing that if the fully managed file storage service is a catalyst for cloud adoption so the question is which workloads should people choose to move into the cloud where's the low friction low risk sweet spot ed well that's one of the first questions that customers ask when they're about to embark on their cloud journey and i wish i could give a simple or a single answer but the answer is really it varies and it varies per customer and i'll give you an example for some customers the cloud journey begins with what we call extending on-premises workloads into the cloud so an example of that is compute bursting workloads where customers have data on premises and they have some compute on premises but they want to burst their processing of that data to the cloud because they really want to take advantage of the massive amount of compute that they get on aws and that's common with workloads like visual effects ringer chip design simulation genomics analysis and so that's an example of extending to the cloud really leveraging the cloud first for your workloads another example is disaster recovery and that's a really common example customers will use a cloud for their secondary or their failover site rather than maintaining their their second on-prem location and so that's a lot of customers start with some of those workloads by extending to the cloud and then there's there's a lot of other customers where they've made the decision to migrate most or all of their workloads and they're not they're skipping the whole extending step they aren't starting there they're instead focused on going all in as fast as possible because they really want to get to the full benefits of the cloud as fast as possible and for them the migration journey is really it's a matter of sequencing sequencing which specific workloads to move and when and what's interesting is we're increasingly seeing customers prioritizing their most important and their most mission-critical applications ahead of their other workloads in terms of timing and they're they're doing that to get their workloads to benefit from the added resilience they get from running on the cloud so really it really does uh depend dave yeah thank you i mean that's pretty pretty good description of the options there and i i just come something you know bursting obviously i love those examples you gave around genomics chip design visual effects rendering the dr piece is again very common sort of cloud you know historical you know sweet spots for cloud but then the point about mission critical is interesting because i hear a lot of customers especially with the digital transformation push wanting to change their operating model i mean on the one hand not changing things put it in the cloud the lift and shift you have to change things low friction but then once they get there they're like wow we can do a lot more with the cloud so that was really helpful those those examples now last year at storage day you released a new file service and then you followed that up at re-event with another file service introduction sometimes i can admit i get lost in the array of services so help us understand when a customer comes to aws with like an nfs or an smb workload how do you steer them to the right managed service you know the right horse for the right course yeah well i'll start by saying uh you know a big part of our focus has been in providing choice to customers and what customers tell us is that the spectrum of options that we provide to them really helps them in their cloud journey because there really isn't a one-size-fits-all file system for all workloads and so having these options actually really helps them to to be able to move pretty easily to the cloud um and so my answer to your question about uh where do we steer a customer when they have a file workload is um it really depends on what the customer is trying to do and uh in many cases where they're coming from so i'll walk you through a little bit of of of how we think about this with customers so for storage and application admins who are extending existing workloads to the cloud or migrating workloads to aws the easiest path generally is to move to an fsx file system that provides the same or really similar underlying file system engine that they use on premises so for example if you're running a netapp appliance on premises or a windows file server on premises choosing that option within fsx provides the least effort for a customer to lift their application and their data set and they'll get the full safe set of capabilities that they're used to they'll get the performance profiles that they're used to but of course they'll get all the benefits of the cloud that i was talking about earlier like spin up and spin down and fully managed and elastic capacity then we also provide open source file systems within the fsx family so if you're a customer and you're used to those or if you aren't really wedded to a particular file system technology these are really good options and they're built on top of aws's latest infrastructure innovations which really allows them to provide pretty significant price and performance benefits to customers so for example the file system file servers for these offerings are powered by aws's graviton family of processors and under the hood we use storage technology that's built on top of aws's scalable reliable datagram transport protocol which really optimizes for for speed on the cloud and so for those two open source file systems we have open zfs and that provides a really powerful highly performant nfs v3 and v4 and 4.1 and 4.2 file system built on a fast and resilient open source linux file system it has a pretty rich set of capabilities it has things like point-to-time snapshots and in-place data cloning and our customers are really using it because of these capabilities and because of its performance for a pretty broad set of enterprise i.t workloads and vertically focused workloads like within the financial services space and the healthcare life sciences space and then luster is a scale-out file system that's built on the world's most popular high-performance file system which is the luster open source file system and customers are using it for compute intensive workloads where they're throwing tons of compute at massive data sets and they need to drive tens or hundreds of gigabytes per second of throughput it's really popular for things like machine learning training and high performance computing big data analytics video rendering and transcoding so really those scale out compute intensive workloads and then we have a very different type of customer very different persona and this is the individual that we call the aws builder and these are folks who are running cloud native workloads they leverage a broad spectrum of aws's compute and analytic services and they have really no history of on-prem examples are data scientists who require a file share for training sets research scientists who are performing analysis on lab data developers who are building containerized or serverless workloads and cloud practitioners who need a simple solution for storing assets for their cloud workflows and and these these folks are building and running a wide range of data focused workloads and they've grown up using services like lambda and building containerized workloads so most of these individuals generally are not storage experts and they look for storage that just works s3 and consumer file shares uh like dropbox are their reference point for how cloud storage works and they're indifferent to or unaware of bio protocols like smb or nfs and performing typical nas administrative tasks is just not it's not a natural experience for them it's not something they they do and we built amazon efs to meet the needs of that group it's fully elastic it's fully serverless spreads data across multiple availability zones by default it scales infinitely it works very much like s3 so for example you get the same durability and availability profile of s3 you get intelligent tiering of colder data just like you do on s3 so that service just clicks with cloud native practitioners it's it's intuitive and it just works there's mind-boggling the number of use cases you just went through and this is where it's so you know it's you know a lot of times people roll their eyes oh here's amazon talking about you know customer obsession again but if you don't stay close to your customers there's no way you could have predicted when you're building these services how they were going to be put to use the only way you can understand it is watch what customers do with it i loved the conversation about graviton we've written about that a lot i mean nitro we've written about that how it's you've completely rethought virtualization the security components in there the hpc luster piece and and the efs for data scientists so really helpful there thank you i'm going to change uh topics a little bit because there's been this theme that you've been banging on at storage day putting data to work and i tell you it's a bit of a passion of mine ed because frankly customers have been frustrated with the return on data initiatives it's been historically complicated very time consuming and expensive to really get value from data and often the business lines end up frustrated so let's talk more about that concept and i understand you have an announcement that fits with this scene can you tell us more about that absolutely today we're announcing a new service called amazon file cache and it's a service on aws that accelerates and simplifies hybrid workflows and specifically amazon file cache provides a high speed cache on aws that makes it easier to process file data regardless of where the data is stored and amazon file cache serves as a temporary high performance storage location and it's for data that's stored in on-premise file servers or in file systems or object stores in aws and what it does is it enables enterprises to make these dispersed data sets available to file based applications on aws with a unified view and at high speeds so think of sub millisecond latencies and and tens or hundreds of gigabytes per second of throughput and so a really common use case it supports is if you have data stored on premises and you want to burst the processing workload to the cloud you can set up this cache on aws and it allows you to have the working set for your compute workload be cached near your aws compute so what you would do as a customer when you want to use this is you spin up this cache you link it to one or more on-prem nfs file servers and then you mount this cache to your compute instances on aws and when you do this all of your on-prem data will appear up automatically as folders and files on the cache and when your aws compute instances access a file for the first time the cache downloads the data that makes up that file in real time and that data then would reside on the cache as you work with it and when it's in the cache your application has access to that data at those sub millisecond latencies and at up to hundreds of gigabytes per second of throughput and all of this data movement is done automatically and in the background completely transparent to your application that's running on the compute instances and then when you're done with your workload with your data processing job you can export the changes and all the new data back to your on-premises file servers and then tear down the cache another common use case is if you have a compute intensive file-based application and you want to process a data set that's in one or more s3 buckets you can have this cache serve as a really high speed layer that your compute instances mount as a network file system you can also place this cache in front of a mix of on-prem file servers and s3 buckets and even fsx file systems that are on aws all of the data from these locations will appear within a single name space that clients that mount the cache have access to and those clients get all the performance benefits of the cache and also get a unified view of their data sets and and to your point about listening to customers and really paying attention to customers dave we built this service because customers asked us to a lot of customers asked us to actually it's a really helpful enable enabler for a pretty wide variety of cloud bursting workloads and hybrid workflows ranging from media rendering and transcoding to engineering design simulation to big data analytics and it really aligns with that theme of extend that we were talking about earlier you know i often joke that uh aws has the best people working on solving the speed of light problem so okay but so this idea of bursting as i said has been a great cloud use case from the early days and and bringing it to file storage is very sound and approach with file cache looks really practical um when is the service available how can i get started you know bursting to aws give us the details there yeah well stay tuned we we announced it today at storage day and it will be generally available later this year and once it becomes available you can create a cache via the the aws management console or through the sdks or the cli and then within minutes of creating the cache it'll be available to your linux instances and your instances will be able to access it using standard file system mount commands and the pricing model is going to be a pretty familiar one to cloud customers customers will only pay for the cash storage and the performance they need and they can spin a cash up and use it for the duration of their compute burst workload and then tear it down so i'm really excited that amazon file cache will make it easier for customers to leverage the agility and the performance and the cost efficiency of aws for processing data no matter where the data is stored yeah cool really interested to see how that gets adopted ed always great to catch up with you as i said the pace is mind-boggling it's accelerating in the cloud overall but storage specifically so by asking us can we take a little breather here can we just relax for a bit and chill out uh not as long as customers are asking us for more things so there's there's more to come for sure all right ed thanks again great to see you i really appreciate your time thanks dave great catching up okay and thanks for watching our coverage of aws storage day 2022 keep it right there for more in-depth conversations on thecube your leader in enterprise and emerging tech coverage [Music] you

Published Date : Aug 12 2022

SUMMARY :

and then you mount this cache to your

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Edward NaimPERSON

0.99+

tensQUANTITY

0.99+

tens of petabytesQUANTITY

0.99+

hundreds of petabytesQUANTITY

0.99+

AWSORGANIZATION

0.99+

amazonORGANIZATION

0.99+

awsORGANIZATION

0.99+

hundreds of thousandsQUANTITY

0.99+

last yearDATE

0.98+

16 yearsQUANTITY

0.98+

todayDATE

0.98+

first timeQUANTITY

0.97+

each yearQUANTITY

0.97+

davePERSON

0.97+

secondQUANTITY

0.97+

dave vellantePERSON

0.97+

20 30 percent a yearQUANTITY

0.97+

later this yearDATE

0.96+

oneQUANTITY

0.96+

awsTITLE

0.95+

windowsTITLE

0.94+

thousands of applicationsQUANTITY

0.94+

later todayDATE

0.93+

one formatQUANTITY

0.93+

hundreds of gigabytes per secondQUANTITY

0.93+

first questionsQUANTITY

0.93+

hundreds of gigabytes per secondQUANTITY

0.92+

two open sourceQUANTITY

0.92+

s3TITLE

0.92+

fsxTITLE

0.89+

4.1TITLE

0.88+

firstQUANTITY

0.88+

a lot of yearsQUANTITY

0.87+

earlier todayDATE

0.84+

linuxTITLE

0.84+

four of the most popular fileQUANTITY

0.79+

nitroORGANIZATION

0.79+

netappTITLE

0.78+

4.2TITLE

0.74+

single answerQUANTITY

0.74+

gravitonTITLE

0.74+

zettabytesQUANTITY

0.73+

dayEVENT

0.73+

lot of customersQUANTITY

0.73+

exabytesQUANTITY

0.72+

a lot of other customersQUANTITY

0.71+

2022DATE

0.71+

v4TITLE

0.71+

single nameQUANTITY

0.68+

tons of computeQUANTITY

0.64+

couple thingsQUANTITY

0.63+

minutesQUANTITY

0.56+

DayEVENT

0.54+

Mat Mathews & Randy Boutin, AWS | AWS Storage Day 2022


 

(upbeat music) >> Welcome to theCube's coverage of AWS Storage Day. We're here with a couple of AWS product experts. Covering AWS's migration and transfer services, Randy Boutin is the general manager of AWS DataSync, and Mat Matthews, GM of AWS Transfer Family. Guys, good to see you again. Thanks for coming on. >> Dave, thanks. >> So look, we saw during the pandemic, the acceleration to cloud migration. We've tracked that, we've quantified that. What's driving that today? >> Yeah, so Dave, great to be back here. Saw you last year at Storage Day. >> Nice to be in studio too, isn't it? Thanks, guys, for coming in. >> We've conquered COVID. >> So yeah, I mean, this is a great question. I think digital transformation is really what's driving a lot of the focus right now from companies, and it's really not about just driving down costs. It's also about what are the opportunities available once you get into the cloud in terms of, what does that unlock in terms of innovation? So companies are focused on the usual things, optimizing costs, but ensuring they have the right security and agility. You know, a lot has happened over the last year, and companies need to be able to react, right? They need to be able to react quickly, so cloud gives them a lot of these capabilities, but the real benefit that we see is that once your data's in the cloud, it opens up the power of the cloud for analytics, for new application development, and things of that sort, so what we're seeing is that companies are really just focused on understanding cloud migration strategy, and how they can get their data there, and then use that to unlock that data for the value. >> I mean, if I've said it once, I've said it 100 times, if you weren't a digital business during the pandemic, you were out of business. You know, migration historically is a bad word in IT. Your CIOs see it and go, "Ugh." So what's the playbook for taking years of data on-prem, and moving it into the cloud? What are you seeing as best practice there? >> Yeah, so as you said, the migration historically has been painful, right? And it's a daunting task for any business or any IT executive, but fortunately, AWS has a broad suite of capabilities to help enable these migrations. And by that, I mean, we have tools to help you understand your existing on-prem workloads, understand what services in the AWS offering align to those needs, but also help you estimate the cost, right? Cost is a big part of this move. We can help you estimate that cost, and predict that cost, and then use tools like DataSync to help you move that data when that time comes. >> So you're saying you help predict the cost of the migration, or the cost of running in the cloud? >> Running in the cloud, right. Yeah, we can help estimate the run time. Based on the performance that we assess on-prem, we can then project that into a cloud service, and estimate that cost. >> So can you guys explain DataSync? Sometimes I get confused, DataSync, what's the difference between DataSync and Storage Gateway? And I want to get into when we should use each, but let's start there if we could. >> Yeah, sure, I'll take that. So Storage Gateway is primarily a means for a customer to access their data in the cloud from on-prem. All right, so if you have an application that you want to keep on-prem, you're not ready yet to migrate that application to the cloud, Gateway is a strong solution, because you can move a lot of that data, a lot of your cold or long tail data into something like S3 or EFS, but still access it from your on-prem location. DataSync's all about data movement, so if you need to move your data from A to B, DataSync is your optimized solution to do that. >> Are you finding that people, that's ideally a one time move, or is it actually, sometimes you're seeing customers do it more? Again, moving data, if I don't- Move as much data as you need to, but no more, to paraphrase Einstein. >> What we're seeing in DataSync is that customers do use DataSync for their initial migration. They'll also, as Matt was mentioning earlier, once you get your data into the cloud, that flywheel of potential starts to take hold, and customers want to ultimately move that data within the cloud to optimize its value. So you might move from service to service. You might move from EFS to S3, et cetera, to enable the cloud flywheel to benefit you. DataSync does that as well, so customers use us to initially migrate, they use us to move within the cloud, and also we just recently announced service for other clouds, so you can actually bring data in now from Google and Azure as well. >> Oh, how convenient. So okay, so that's cool. So you helped us understand the use cases, but can we dig one more layer, like what protocols are supported? I'm trying to understand really the right fit for the right job. >> Yeah, so that's really important. So for transfer specifically, one of the things that we see with customers is you've got obviously a lot of internal data within your company, but today it's a very highly interconnected world, so companies deal with lots of business partners, and historically they've used, there's a big prevalence of using file transfer to exchange data with business partners, and as you can imagine, there's a lot of value in that data, right? Sometimes it's purchase orders, inventory data from suppliers, or things like that. So historically customers have had protocols like SFTP or FTP to help them interface with or exchange data or files with external partners. So for transfer, that's what we focus on is helping customers exchange data over those existing protocols that they've used for many years. And the real focus is it's one thing to migrate your own data into the cloud, but you can't force thousands or tens of thousands sometimes of partners to also work in a different way to get you their data, so we want to make that very seamless for customers using the same exact protocols like SFTP that they've used for years. We just announced AS2 protocol, which is very heavily used in supply chains to exchange inventory and information across multi-tiers of partners, and things of that nature. So we're really focused on letting customers not have to impact their partners, and how they work and how they exchange, but also take advantage of the data, so get that data into the cloud so they can immediately unlock the value with analytics. >> So AS2 is specifically in the context of supply chain, and I'm presuming it's secure, and kind of governed, and safe. Can you explain that a little bit? >> Yeah, so AS2 has a lot of really interesting features for transactional type of exchanges, so it has signing and encryption built in, and also has notification so you can basically say, "Hey, I sent you this purchase order," and to prove that you received it, it has capability called non-repudiation, which means it's actually a legal transaction. So those things are very important in transactional type of exchanges, and allows customers in supply chains, whether it's vendors dealing with their suppliers, or transportation partners, or things like that to leverage file transfer for those types of exchanges. >> So encryption, providence of transactions, am I correct, without having to use the blockchain, and all the overhead associated with that? >> It's got some built in capabilities. >> I mean, I love blockchain, but there's drawbacks. >> Exactly, and that's why it's been popular. >> That's really interesting, 'cause Andy Jassy one day, I was on a phone call with him and John Furrier, and we were talking up crypto and blockchain. He said, "Well, why do, explain to me." You know Jassy, right? He always wants to go deeper. "Explain why I can't do this with some other approach." And so I think he was recognizing some of the drawbacks. So that's kind of a cool thing, and it leads me- We're running this obviously today, August 10th. Yesterday we had our Supercloud event in Palo Alto on August 9th, and it's all about the ecosystem. One of the observations we made about the 2020s is the cloud is totally different now. People are building value on top of the infrastructure that you guys have built out over the last 15 years. And so once an organization's data gets into the cloud, how does it affect, and it relates to AS2 somewhat, how does it affect the workflows in terms of interacting with external partners, and other ecosystem players that are also in the cloud? >> Yeah, great, yeah, again, we want to try and not have to affect those workflows, take them as they are as much as possible, get the data exchange working. One of the things that we focus on a lot is, how do you process this data once it comes in? Every company has governance requirements, security requirements, and things like that, so they usually have a set of things that they need to automate and orchestrate for the data as it's coming in, and a lot of these companies use something called Managed File Transfer Solutions that allow them to automate and orchestrate those things. We also see that many times this is very customer specific, so a bank might have a certain set of processes they have to follow, and it needs to be customized. As you know, AWS is a great solution for building custom solutions, and actually today, we're just announcing a new set of of partners in a program called the Service Delivery Program with AWS Transfer Family that allows customers to work with partners that are very well versed in transfer family and related services to help build a very specific solution that allows them to build that automation orchestration, and keep their partners kind of unaware that they're interfacing in a different way. >> And once this data is in the cloud, or actually, maybe stays on-prem in some cases, but it basically plugs in to the AWS services portfolio, the whole security model, the governance model, shared responsibility comes in, is that right? It's all, sort of all in there? >> Yeah, that's right, that's exactly right, and we're working with it's all about the customer's needs, and making sure that their investment in AWS doesn't disrupt their existing workflows and their relationships with their customers and their partners, and that's exactly what Matt's been describing is we're taking a close look at how we can extend the value of AWS, integrate into our customer's workflows, and bring that value to them with minimal investment or disruption. >> So follow up on that. So I love that, because less disruption means it's easier, less friction, and I think of like, trying to think of examples. Think about data de-duplication like purpose-built backup appliances, right? Data domain won that battle, because they could just plug right in. Avamar, they were trying to get you to redo everything, okay, and so we saw that movie play out. At the same time, I've talked to CIOs that say, "I love that, but the cloud opens up all these cool new opportunities for me to change my operating model." So are you seeing that as well? Where okay, we make it easy to get in. We're not disrupting workflows, and then once they get in, they say, "Well if we did it this way, we'd take out a bunch of costs. We'd accelerate our business." What's that dynamic like? >> Exactly that, right. So that moved to the Cloud Continuum. We don't think it's going to be binary. There's always going to be something on-prem. We accept that, but there's a continuum there, so day one, they'll migrate a portion of that workload into the cloud, start to extract and see value there, but then they'll continue, as you said, they'll continue to see opportunities. With all of the various capabilities that AWS has to offer, all the value that represents, they'll start to see that opportunity, and then start to engage and consume more of those features over time. >> Great, all right, give us the bumper sticker. What's next in transfer services from your perspectives? >> Yeah, so we're obviously always going to listen to our customers, that's our focus. >> You guys say that a lot. (all laughing) We say it a lot. But yeah, so we're focused on helping customers again increase that level of automation orchestration, again that suite of capability, generally, in our industry, known as managed file transfer, when a file comes in, it needs to get maybe encrypted, or decrypted, or compressed, or decompressed, scanned for viruses, those kind of capabilities, make that easier for customers. If you remember last year at Storage Day, we announced a low code workflow framework that allows customers to kind of build those steps. We're continuing to add built-in capabilities to that so customers can easily just say, "Okay, I want these set of activities to happen when files come in and out." So that's really what's next for us. >> All right, Randy, we'll give you the last word. Bring us home. >> I'm going to surprise you with the customer theme. >> Oh, great, love it. >> Yeah, so we're listening to customers, and what they're asking for our support for more sources, so we'll be adding support for more cloud sources, more on-prem sources, and giving the customers more options, also performance and usability, right? So we want to make it easier, as the enterprise continues to consume the cloud, we want to make DataSync and the movement of their data as easy as possible. >> I've always said it starts with the data. S3, that was the first service, and the other thing I've said a lot is the cloud is expanding. We're seeing connections to on-prem. We're seeing connections out to the edge. It's just becoming this massive global system, as Werner Vogels talks about all the time. Thanks, guys, really appreciate it. >> Dave, thank you very much. >> Thanks, Dave. >> All right, keep it right there for more coverage of AWS Storage Day 2022. You're watching theCube. (upbeat music)

Published Date : Aug 12 2022

SUMMARY :

Guys, good to see you again. the acceleration to cloud migration. Yeah, so Dave, great to be back here. Nice to be in studio too, isn't it? and companies need to and moving it into the cloud? in the AWS offering align to those needs, Running in the cloud, right. So can you guys explain DataSync? All right, so if you have an application but no more, to paraphrase Einstein. for other clouds, so you can for the right job. so get that data into the cloud and kind of governed, and safe. and to prove that you received it, but there's drawbacks. Exactly, and that's One of the observations we made that they need to automate and orchestrate and making sure that their investment for me to change my operating model." So that moved to the Cloud Continuum. services from your perspectives? always going to listen that allows customers to give you the last word. I'm going to surprise the movement of their data We're seeing connections out to the edge. of AWS Storage Day 2022.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Randy BoutinPERSON

0.99+

AWSORGANIZATION

0.99+

JassyPERSON

0.99+

Mat MatthewsPERSON

0.99+

MattPERSON

0.99+

Palo AltoLOCATION

0.99+

thousandsQUANTITY

0.99+

DavePERSON

0.99+

100 timesQUANTITY

0.99+

Andy JassyPERSON

0.99+

EinsteinPERSON

0.99+

August 9thDATE

0.99+

first serviceQUANTITY

0.99+

RandyPERSON

0.99+

last yearDATE

0.99+

John FurrierPERSON

0.99+

GoogleORGANIZATION

0.99+

todayDATE

0.99+

Werner VogelsPERSON

0.99+

YesterdayDATE

0.99+

tens of thousandsQUANTITY

0.99+

S3TITLE

0.98+

one timeQUANTITY

0.98+

one more layerQUANTITY

0.98+

Mat MathewsPERSON

0.98+

OneQUANTITY

0.97+

2020sDATE

0.97+

AWS DataSyncORGANIZATION

0.97+

Storage DayEVENT

0.97+

pandemicEVENT

0.97+

AWS Transfer FamilyORGANIZATION

0.97+

one thingQUANTITY

0.97+

eachQUANTITY

0.96+

oneQUANTITY

0.94+

EFSTITLE

0.94+

DataSyncTITLE

0.93+

SupercloudEVENT

0.89+

COVIDOTHER

0.85+

AzureORGANIZATION

0.82+

August 10thDATE

0.79+

last 15 yearsDATE

0.79+

today,DATE

0.77+

AS2OTHER

0.75+

yearsQUANTITY

0.7+

Transfer FamilyTITLE

0.69+

Storage Day 2022EVENT

0.66+

Service Delivery ProgramOTHER

0.64+

AS2ORGANIZATION

0.63+

thingsQUANTITY

0.58+

onceQUANTITY

0.57+

DataSyncORGANIZATION

0.57+

AWSEVENT

0.56+

GatewayORGANIZATION

0.52+

theCubePERSON

0.52+

AS2TITLE

0.5+

Storage GatewayORGANIZATION

0.48+

AvamarORGANIZATION

0.47+

dayQUANTITY

0.47+

SFTPOTHER

0.45+

SFTPTITLE

0.44+

2022DATE

0.43+

Storage GatewayTITLE

0.31+

Kevin Miller, AWS | Modernize, unify, and innovate with data | AWS Storage Day 2022


 

(upbeat music) >> We're here on theCube covering AWS Storage Day 2022. Kevin Miller joins us. He's the vice president and general manager of Amazon S3. Hello, Kevin, good to see you again. >> Hey Dave, it's great to see you as always. >> It seems like just yesterday we were celebrating the 15th anniversary of S3, and of course the launch of the modern public cloud, which started there. You know, when you think back Kevin, over the past year, what are some of the trends that you're seeing and hearing from customers? What do they want to see AWS focus more on? What's the direction that you're setting? >> Yeah, well Dave, really I think there's probably three trends that we're seeing really pop this year. I think one just given the kind of macroeconomic situation right now is cost optimization. That's not a surprise. Everyone's just taking a closer look at what they're using, and where they might be able to pair back. And you know, I think that's a place that obviously S3 has a long history of helping customers save money. Whether it's through our new storage classes, things like our Glacier Instant Retrieval, storage class that we launched to reinvent last year. Or things like our S3 storage lens capability to really dig in and help customers identify where their costs are are being spent. But so certainly every, you know, a lot of customers are focused on that right now, and for obvious reasons. I think the second thing that we're seeing is, just a real focus on simplicity. And it kind of goes hand in hand with cost optimization, because what a lot of customers are looking for is, how do I take the staff that I have, and do more this year. Right, continue to innovate, continue to bring new applications or top line generating revenue applications to the market, but not have to add a lot of extra headcount to do that. And so, what they're looking for is management and simplicity. How do I have all of this IT infrastructure, and not have to have people spending a lot of their time going into kind of routine maintenance and operations. And so that's an area that we're spending a lot of time. We think we have a lot of capability today, but looking at ways that we can continue to simplify, make it easier for customers to manage their infrastructure. Things like our S3 intelligent tiering storage class, which just automatically gives cost savings for data that's not routinely accessed. And so that's a big focus for us this year as well. And then I think the last and probably third thing I would highlight is an emerging theme or it's been a theme, but really continuing to increase in volume, is all around sustainability. And you know, our customers are looking for us to give them the data and the assurances for them, for their own reports and their own understanding of how sustainable is my infrastructure. And so within AWS, of course, you know we're on a path towards operating with 100% renewable energy by 2025. As well as helping the overall Amazon goal of achieving net zero carbon by 2040. So those are some big lofty goals. We've been giving customers greater insights with our carbon footprint tool. And we think that, you know the cloud continues to be just a great place to run and reduce customer's carbon footprint for the similar you know, storage capacity or similar compute capacity. But that's just going to continue to be a trend and a theme that we're looking at ways that we can continue to help customers do more to aggressively drive down their carbon footprint. >> I mean, it makes sense. It's like you're partnering up with the cloud, you know, you did same thing on security, you know, there's that shared responsibility model, same thing now with ESG. And on the macro it's interesting Kevin, this is the first time I can remember where, you know it used to be, if there's a downturn it's cost optimization, you go to simplicity. But at the same time with digital, you know, the rush to digital, people still are thinking about, okay how do I invest in the future? So but let's focus on cost for a moment then we'll come back to sort of the data value. Can you tell us how AWS helps customers save on storage, you know, beyond just the price per terabyte actions that you could take. I mean I love that, you guys should keep doing that. >> Absolutely. >> But what other knobs are you turning? >> Yeah, right and we've had obviously something like 15 cost reductions or price reductions over the years, and we're just going to continue to use that lever where we can, but it's things like the launch of our Glacier Instant Retrieval storage class that we did last year at Reinvent, where that's now you know, 4/10ths of a cent per gigabyte month. For data that customers access pretty infrequently maybe a few times a year, but they can now access that data immediately and just pay a small retrieval fee when they access that data. And so that's an example of a new capability that reduces customer's total cost of ownership, but is not just a straight up price reduction. I mentioned S3 Intelligent-Tiering, that's another case where, you know, when we launch Glacier Instant Retrieval, we integrated that with Intelligent-Tiering as well. So we have the archive instant access tier within Intelligent-Tiering. And so now data that's not accessed for 90 days is just automatically put into AIA and and then results in a reduced storage cost to customers. So again, leaning into this idea that customers are telling us, "Just do, you know what should be done "for my data to help me reduce cost, can you just do it, "and sort of give me the right defaults." And that's what we're trying to do with things like Intelligent-Tiering. We've also, you know, outside of the S3 part of our portfolio, we've been adding similar kinds of capabilities within some of our file services. So things like our, you know elastic file service launched a one zone storage class as well as an intelligent tiering capability to just automatically help customers save money. I think in some cases up to 92% on their their EFS storage costs with this automatic intelligent tiering capability. And then the last thing I would say is that we also are just continuing to help customers in other ways, like I said, our storage lens is a great way for customers to really dig in and figure out. 'Cause you know, often customers will find that they may have, you know, certain data sets that someone's forgotten about or, they're capturing more data than they expected perhaps in a logging application or something that ends up generating a lot more data than they expected. And so storage lens helps them really zoom in very quickly on, you know this is the data, here's how frequently it's being accessed and then they can make decisions about use that data I keep, how long do I keep it? Maybe that's good candidates to move down into one of our very cold storage classes like Glacier Deep Archive, where they they still have the data, but they don't expect to need to actively retrieve it on a regular basis. >> SDL bromide, if you can measure it, you can manage it. So if I can see it, visualize it, that I can take actions. When you think about S3- >> That's right. it's always been great for archival workloads but you made some updates to Glacier that changed the way that we maybe think about archive data. Can you talk about those changes specifically, what it means for how customers should leverage AWS services going forward? >> Yeah, and actually, you know, Glacier's coming up on its 10 year anniversary in August, so we're pretty excited about that. And you know, but there's just been a real increase in the pace of innovation, I think over the last three or four years there. So we launched the Glacier Deep Archive capability in 2019, 2018, I guess it was. And then we launched Glacier Instant Retrieval of course last year. So really what we're seeing is we now have three storage classes that cover are part of the Glacier family. So everything from millisecond retrieval for that data, that needs to be accessed quickly when it is accessed, but isn't being accessed, you know, regularly. So maybe a few times a year. And there's a lot of use cases that we're seeing really quickly emerge for that. Everything from, you know, user generated content like photos and videos, to big broadcaster archives and particularly in media and entertainment segment. Seeing a lot of interest in Glaciers Instant Retrieval because that data is pretty cold on a regular basis. But when they want to access it, they want a huge amount of data, petabytes of data potentially back within seconds, and that's the capability we can provide with Glacier Instant Retrieval. And then on the other end of the spectrum, with Glacier Deep Archive, again we have customers that have huge archives of data that they be looking to have that 3-AZ durability that we provide with Glacier, and make sure that data is protected. But really, you know expect to access it once a year if ever. Now it could be a backup copy of data or secondary or tertiary copy of data, could be data that they just don't have an active use for it. And I think that's one of the things we're starting to see grow a lot, is customers that have shared data sets where they may not need that data right now but they do want to keep it because as they think about, again these like new applications that can drive top line growth, they're finding that they may go back to that data six months or nine months from now and start to really actively use it. So if they want that option value to keep that data so they can use it down the road, Glacier Deep Archive, or Glacier Flexible Retrieval, which is kind of our storage class right in the middle of the road. Those are great options for customers to keep the data, keep it safe and secure, but then have it, you know pretty accessible when they're ready to get it back. >> Got it, thank you for that. So, okay, so customers have choices. I want to get into some of the competitive differentiators. And of course we were talking earlier about cost optimization, which is obviously an important topic given the macro environment you know, but there's more. And so help us understand what's different about AWS in terms of helping customers get value from their data, cost reduction as a component of value, part of the TCO, for sure. But just beyond being a cloud bit bucket, you know just a storage container in the cloud, what are some of the differentiators that you can talk to? >> Yeah, well Dave, I mean, I think that when it comes to value, I think there's tremendous benefits in AWS, well beyond just cost reduction. I think, you know, part of it is S3 now has built, I think, an earned reputation for being resilient, for storing, you know, at massive scale giving customers that confidence that they will be able to scale up. You know, we store more than 200 trillion objects. We regularly peak at over 100 million requests per second. So customers can build on S3 and Glacier with the confidence that we're going to be there to help their applications grow and scale over time. And then I think that in all of the applications both first party and third party, the customers can use, and services that they can use to build modern applications is an incredible benefit. So whether it's all of our serverless offerings, things like Lambda or containers and everything we have to manage that. Or whether it's the deep analytics and machine learning capabilities we have to help really extract, you know value and insight from data in near real time. You know, we're just seeing an incredible number of customers build those kinds of applications where they're processing data and feeding their results right back into their business right away. So I'm just going to briefly mention a couple, like, you know one example is ADP that really helps their customers measure, compare and sort of analyze their workforce. They have a couple petabytes of data, something like 25 billion individual data points and they're just processing that data continuously through their analytics and machine learning applications to then again, give those insights back to their customers. Another good example is AstraZeneca. You know, they are processing petabytes and petabytes of genomic sequencing data. And they have a goal to analyze 2 million genomes over the next four years. And so they're just really scaling up on AWS, both from a pure storage point of view, but more importantly, from all of the compute and analytics capability on top that is really critical to achieving that goal. And then, you know, beyond the first party services we have as I mentioned, it's really our third party, right? The AWS partner network provides customers an incredible range of choice in off the shelf applications that they can quickly provision and make use of the data to drive those business insights. And I think today the APN has something like 100,000 partners over in 150 countries. And we specifically have a storage competency partner where customers can go to get those applications that directly work, you know, on top of their data. And really, like I said, drive some of that insight. So, you know, I think it's that overall benefit of being able to really do a lot more with their data than just have it sit idle. You know, that's where I think we see a lot of customers interested in driving additional value. >> I'm glad you mentioned the ecosystem, and I'm glad you mentioned the storage competency as well. So there are other storage partners that you have, even though you're a head of a big storage division. And then I think there's some other under the cover things too. I've recently wrote, actually have written about this a lot. Things like nitro and rethinking virtualization and how to do, you know offloads. The security that comes, you know fundamentally as part of the platform is, I think architecturally is something that leads the way in the industry for sure. So there's a lot we could unpack, but you've fundamentally changed the storage market over the last 16 years. And again, I've written about this extensively. We used to think about storage in blocks or you got, you know, somebody who's really good in files, there were companies that dominated each space with legacy on-prem storage. You know, when you think about object storage Kevin, it was a niche, right? It was something used for archival, it was known for its simple, get put syntax, great for cheap and deep storage, and S3 changed that. Why do you think that's happened and S3 has evolved, the object has evolved the way it has, and what's the future hold for S3? >> Yeah I mean, you know, Dave, I think that probably the biggest overall trend there is that customers are looking to build cloud native applications. Where as much of that application is managed as they can have. They don't want to have to spend time managing the underlying infrastructure, the compute and storage and everything that goes around it. And so a fully managed service like S3, where there's no provisioning storage capacity, there's, you know we provide the resiliency and the durability that just really resonates with customers. And I think that increasingly, customers are seeing that they want to innovate across the entire range of business. So it's not about a central IT team anymore, it's about engineers that are embedded within lines of business, innovating around what is critical to achieve their business results. So, you know, if they're in a manufacturing segment, how can we pull data from sensors and other instrumentation off of our equipment and then make better decisions about when we need to do predictive maintenance, how quickly we can run our manufacturing line, looking for inefficiencies. And so we've developed around our managed offerings like S3, we've just developed, you know, customers who are investing and executing on plans and you know transformations. That really give them, you know put digital technology directly into the line of business that they're looking for. And I think that trend is just going to continue. People sometimes ask me, well "I mean, 16 years, you know, isn't S3 done?" And I would say, "By no stretcher are we done." We have plenty of feedback from customers on ways that we can continue to simplify, reduce the kinds of things they need to do, when they're looking for example and rolling out new security policies and parameters across their entire organization. So raising the bar there, finding, you know, raising the bar on how they can efficiently manage their storage and reduce costs. So I think we have plenty of innovation ahead of us to continue to help customers provide that fully managed capability. >> Yeah I often say Kevin, the next 10 years ain't going to be like the last in cloud. So I really thank you for coming on theCube and sharing your insights, really appreciate it. >> Absolutely Dave, thanks for having me. >> You're welcome. Okay keep it right there for more coverage of AWS Storage Day 2022 in theCube. (calm bright music)

Published Date : Aug 10 2022

SUMMARY :

Hello, Kevin, good to see you again. to see you as always. and of course the launch And we think that, you know that you could take. that they may have, you When you think about S3- Glacier that changed the way And you know, but there's that you can talk to? And then, you know, beyond the and how to do, you know offloads. and you know transformations. So I really thank you of AWS Storage Day 2022 in theCube.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Kevin MillerPERSON

0.99+

KevinPERSON

0.99+

AWSORGANIZATION

0.99+

DavePERSON

0.99+

AmazonORGANIZATION

0.99+

2019DATE

0.99+

90 daysQUANTITY

0.99+

2025DATE

0.99+

AugustDATE

0.99+

2018DATE

0.99+

2040DATE

0.99+

last yearDATE

0.99+

16 yearsQUANTITY

0.99+

100,000 partnersQUANTITY

0.99+

2 million genomesQUANTITY

0.99+

S3TITLE

0.99+

yesterdayDATE

0.99+

more than 200 trillion objectsQUANTITY

0.99+

todayDATE

0.99+

GlacierORGANIZATION

0.99+

ReinventORGANIZATION

0.98+

bothQUANTITY

0.98+

15th anniversaryQUANTITY

0.98+

first timeQUANTITY

0.98+

second thingQUANTITY

0.98+

nine monthsQUANTITY

0.98+

this yearDATE

0.98+

APNORGANIZATION

0.97+

LambdaTITLE

0.97+

oneQUANTITY

0.97+

150 countriesQUANTITY

0.96+

AstraZenecaORGANIZATION

0.96+

each spaceQUANTITY

0.95+

15 cost reductionsQUANTITY

0.95+

S3COMMERCIAL_ITEM

0.95+

three storage classesQUANTITY

0.95+

25 billion individual data pointsQUANTITY

0.94+

one exampleQUANTITY

0.94+

three trendsQUANTITY

0.93+

10 year anniversaryQUANTITY

0.93+

up to 92%QUANTITY

0.91+

Instant RetrievalCOMMERCIAL_ITEM

0.9+

six monthsQUANTITY

0.9+

over 100 million requests per secondQUANTITY

0.89+

past yearDATE

0.87+

four yearsQUANTITY

0.87+

last 16 yearsDATE

0.87+

firstQUANTITY

0.87+

third thingQUANTITY

0.86+

Glacier InstantCOMMERCIAL_ITEM

0.85+

100% renewable energyQUANTITY

0.83+

one zoneQUANTITY

0.83+

gigabyteQUANTITY

0.83+

Glacier Deep ArchiveORGANIZATION

0.83+

Glacier Flexible RetrievalORGANIZATION

0.81+

Glacier Instant RetrievalCOMMERCIAL_ITEM

0.79+

Storage Day 2022EVENT

0.78+

few times a yearQUANTITY

0.75+

Wayne Duso & Nancy Wang | AWS Storage Day 2022


 

>>Okay, we're back. My name is Dave Valante and this is the Cube's coverage of AWS storage day. You know, coming off of reinforc I wrote the, the cloud was a new layer of defense. In fact, the first line of defense in a cyber security strategy. And that brings new thinking and models for protecting data, data protection, specifically, traditionally thought of as backup and recovery, it's become a critical adjacency to security and a component of a comprehensive cybersecurity strategy. We're here in our studios outside of Boston with two cube alums, and we're gonna discuss this in other topics. Wayne do so is the vice president for AWS storage edge and data services, and Nancy Wong as general manager of AWS backup and data protection services, guys. Welcome. Great to see you again. Thanks for coming on. Of >>Course, always a pleasure, Dave. Good to >>See you, Dave. All right. So Wayne, let's talk about how organizations should be thinking about this term data protection. It's an expanding definition, isn't >>It? It is an expanding definition. They, last year we talked about data and the importance of data to companies. Every company is becoming a data company, you know, da the amount of data they generate, the amount of data they can use to create models, to do predictive analytics. And frankly, to find ways of innovating is, is grown rapidly. And, you know, there's this tension between access to all that data, right? Getting the value out of that data. And how do you secure that data? And so this is something we think about with customers all the time. So data durability, data protection, data resiliency, and, you know, trust in their data. If you think about running your organization on your data, trust in your data is so important. So, you know, you gotta trust where you're putting your data. You know, people who are putting their data on a platform need to trust that platform will in fact, ensure it's durability, security, resiliency. >>And, you know, we see ourselves AWS as a partner in securing their data, making their data dur durable, making their data resilient, right? So some of that responsibility is on us. Some of that is on so shared responsibility around data protection, data resiliency. And, you know, we think about forever, you know, the notion of, you know, compromise of your infrastructure, but more and more people think about the compromise of their data as data becomes more valuable. And in fact, data is a company's most valuable asset. We've talked about this before. Only second to their people. You know, the people that are most valuable asset, but right next to that is their data. So really important stuff. >>So Nancy, you talked to a lot of customers, but by the way, it always comes back to the data. We've saying this for years, haven't we? So you've got this expanding definition of data protection, you know, governance is in there. You, you think about access cetera. When you talk to customers, what are you hearing from them? How are they thinking about data protection? >>Yeah. So a lot of the customers that Wayne and I have spoken to often come to us seeking thought leadership about, you know, how do I solve this data challenge? How do I solve this data sprawl challenge, but also more importantly, tying it back to data protection and data resiliency is how do I make sure that data is secure, that it's protected against, let's say ransomware events, right. And continuously protected. So there's a lot of mental frameworks that come to mind and a very popular one that comes up in quite a few conversations is this cybersecurity framework, right? And from a data protection perspective is just as important to protect and recover your data as it is to be able to detect different events or be able to respond to those events. Right? So recently I was just having a conversation with a regulatory body of financial institutions in Europe, where we're designing a architecture that could help them make their data immutable, but also continuously protected. So taking a step back, that's really where I see AWS's role in that we provide a wide breadth of primitives to help customers build secure platforms and scaffolding so that they can focus on building the data protection, the data governance controls, and guardrails on top of that platform. >>And, and that's always been AWS's philosophy, you know, make sure that developers have access to those primitives and APIs so that they can move fast and, and essentially build their own if that that's in fact what they wanna do. And as you're saying, when data protection is now this adjacency to cyber security, but there's disaster recoveries in there, business continuance, cyber resilience, et cetera. So, so maybe you could pick up on that and sort of extend how you see AWS, helping customers build out those resilient services. >>Yeah. So, you know, two core pillars to a data protection strategy is around their data durability, which is really an infrastructure element. You know, it's, it's, it's, it's by and large the responsibility of the provider of that infrastructure to make sure that data's durable, cuz if it's not durable, everything else doesn't matter. And then the second pillar is really about data resiliency. So in terms of security, controls and governance, like these are really important, but these are shared responsibility. Like the customers working with us with the services that we provide are there to architect the design, it's really human factors and design factors that get them resiliency, >>Nancy, anything you would add to what Wayne just said. >>Yeah, absolutely. So customers tell us that they want always on data resiliency and data durability, right? So oftentimes in those conversations, three common themes come up, which is they want a centralized solution. They want to be able to transcribe their intent into what they end up doing with their data. And number three, they want something that's policy driven because once you centralize your policies, it's much better and easier to establish control and governance at an organizational level. So keeping that in mind with policy as our interface, there's two managed AWS solutions that I recommend you all check out in terms of data resiliency and data durability. Those are AWS backup, which is our centralized solution for managing protection recovery, and also provides an audit audit capability of how you protect your data across 15 different AWS services, as well as on-premises VMware and for customers whose mission critical data is contained entirely on disk. We also offer AWS elastic disaster recovery services, especially for customers who want to fail over their workloads from on premises to the cloud. >>So you can essentially centralize as a quick follow up, centralize the policy. And like I said, the intent, but you can support a federated data model cuz you're building out this massive, you know, global system, but you can take that policy and essentially bring it anywhere on the AWS cloud. Is that >>Right? Exactly. And actually one powerful integration I want to touch upon is that AWS backup is natively integrated with AWS organizations, which is our defacto multi account federated organization model for how AWS services work with customers, both in the cloud, on the edge, at the edge and on premises. >>So that's really important because as, as we talk about all the time on the cube, this notion of a, a decentralized data architecture data mesh, but the problem is how do you ensure governance and a federated model? So we're clearly moving in that direction. Wayne, I want to ask you about cyber as a board level discussion years ago, I interviewed Dr. Robert Gates, you know, former defense secretary and he sat on a number of boards and I asked him, you know, how important and prominent is security at the board level? Is it really a board level discussion? He said, absolutely. Every time we meet, we talk about cyber security, but not every company at the time, this was kind of early last decade was doing that. That's changed now. Ransomware is front and center. Hear about it all the time. What's AWS. What's your thinking on cyber as a board level discussion and specifically what are you guys doing around ran ransomware? >>Yeah. So, you know, malware in general, ransomware being a particular type of malware. Sure. It's a hot topic and it continues to be a hot topic. And whether at the board level, the C-suite level, I had a chance to listen to Dr. Gates a couple months ago and super motivational, but we think about ransomware and the same way that our customers do. Right? Cause all of us are subject to an incident. Nobody is immune to a ransomware incident. So we think very much the same way. And you, as Nancy said, along the lines of the, this framework, we really think about, you know, how do customers identify their critical access? How do they plan for protecting those assets, right? How do they make sure that they are in fact protected? And if they do detect the ransomware event and ransomware events come from a lot of different places, like there's not one signature, there's not one thumbprint, if you would for ransomware. >>So it's, it's, there's really a lot of vigilance that needs to be put in place, but a lot of planning that needs to be put in place. And once that's detected and a, a, we have to recover, you know, we know that we have to take an action and recover having that plan in place, making sure that your assets are fully protected and can be restored. As you know, ransomware is a insidious type of malware. You know, it sits in your system for a long time. It figures out what's going on, including your backup policies, your protection policies, and figures out how to get around those with some of the things that Nancy talked about in terms of air gaping, your capabilities, being able to, if you would scan your secondary, your backup storage for malware, knowing that it's a good copy. And then being able to restore from that known good copy in the event of an incident is critical. So we think about this for ourselves and the same way that we think about these for our customers. You gotta have a great plan. You gotta have great protection and you gotta be ready to restore in the case of an incident. And we wanna make sure we provide all the capabilities to do >>That. Yeah. So I'll glad you mentioned air gaping. So at the recent re reinforce, I think it was Kurt kufeld was speaking about ransomware and he didn't specifically mention air gaping. I had to leave. So I might have, I might have missed it cause I was doing the cube, but that's a, that's a key aspect. I'm sure there were, were things on the, on the deep dives that addressed air gaping, but Nancy look, AWS has the skills. It has the resources, you know, necessary to apply all these best practices and, you know, share those with customers. But, but what specific investments is AWS making to make the CISO's life easier? Maybe you could talk about that. >>Sure. So following on to your point about the reinforced keynote, Dave, right? CJ Boes talked about how the events of a ransomware, for example, incident or event can take place right on stage where you go from detect to respond and to recover. And specifically on the recovery piece, you mentioned AWS backup, the managed service that protects across 15 different AWS services, as well as on-premises VMware as automated recovery. And that's in part why we've decided to continue that investment and deliver AWS backup audit manager, which helps customers actually prove their posture against how their protection policies are actually mapping back to their organizational controls based on, for example, how they TA tag their data for mission criticality or how sensitive that data is. Right. And so turning to best practices, especially for ransomware events. Since this is very top of mind for a lot of customers these days is I will, will always try to encourage customers to go through game day simulations, for example, identifying which are those most critical applications in their environment that they need up and running for their business to function properly, for example, and actually going through the recovery plan and making sure that their staff is well trained or that they're able to go through, for example, a security orchestration automation, recovery solution, to make sure that all of their mission critical applications are back up and running in case of a ransomware event. >>Yeah. So I love the game day thing. I mean, we know, well just the, in the history of it, you couldn't even test things like disaster recovery, right? Because it was too dangerous with the cloud. You can test these things safely and actually plan out, develop a blueprint, test your blueprint. I love the, the, the game day >>Analogy. Yeah. And actually one thing I'd love to add is, you know, we talked about air gaping. I just wanna kind of tie up that statement is, you know, one thing that's really interesting about the way that the AWS cloud is architected is the identity access and management platform actually allows us to create identity constructs, that air gap, your data perimeter. So that way, when attackers, for example, are able to gain a foothold in your environment, you're still able to air gap your most mission critical and also crown jewels from being infiltrated. >>Mm that's key. Yeah. We've learned, you know, when paying the ransom is not a good strategy, right? Cuz most of the time, many times you don't even get your data back. Okay. So we, we're kind of data geeks here. We love data and we're passionate about it on the cube AWS and you guys specifically are passionate about it. So what excites you, Wayne, you start and then Nancy, you bring us home. What excites you about data and data protection and why? >>You know, we are data nerds. So at the end of the day, you know, there's this expressions we use all the time, but data is such a rich asset for all of us. And some of the greatest innovations that come out of AWS comes out of our analysis of our own data. Like we collect a lot of data on our operations and some of our most critical features for our customers come out of our analysis, that data. So we are data nerds and we understand how businesses view their data cuz we view our data the same way. So, you know, Dave security really started in the data center. It started with the enterprises. And if we think about security, often we talk about securing compute and securing network. And you know, if you, if you secured your compute, you secured your data generally, but we've separated data from compute so that people can get the value from their data no matter how they want to use it. And in doing that, we have to make sure that their data is durable and it's resilient to any sort of incident and event. So this is really, really important to us. And what do I get excited about? You know, again, thinking back to this framework, I know that we as thought leaders alongside our customers who also thought leaders in their space can provide them with the capabilities. They need to protect their data, to secure their data, to make sure it's compliant and always, always, always durable. >>You know, it's funny, you'd say funny it's it's serious actually. Steven Schmidt at reinforc he's the, the, the chief security officer at Amazon used to be the C C ISO of AWS. He said that Amazon sees quadrillions of data points a month. That's 15 zeros. Okay. So that's a lot of data. Nancy bring us home. What's what excites you about data and data protection? >>Yeah, so specifically, and this is actually drawing from conversations that I had with multiple ISV partners at AWS reinforc is the ability to derive value from secondary data, right? Because traditionally organizations have really seen that as a call center, right? You're producing secondary data because most likely you're creating backups of your mission critical workloads. But what if you're able to run analytics and insights and derive insights from that, that secondary data, right? Then you're actually able to let AWS do the undifferentiated heavy lifting of analyzing that secondary data state. So that way us customers or ISV partners can build value on the security layers above. And that is how we see turning cost into value. >>I love it. As you're taking the original premise of the cloud, taking away the under heavy lifting for, you know, D deploying, compute, storage, and networking now bringing up to the data level, the analytics level. So it continues. The cloud continues to expand. Thank you for watching the cubes coverage of AWS storage day 2022.

Published Date : Aug 10 2022

SUMMARY :

Great to see you again. So Wayne, let's talk about how organizations should be thinking about this term data So data durability, data protection, data resiliency, and, you know, And, you know, we think about forever, you know, the notion of, you know, So Nancy, you talked to a lot of customers, but by the way, it always comes back to the data. about, you know, how do I solve this data challenge? And, and that's always been AWS's philosophy, you know, make sure that developers have access it's, it's, it's by and large the responsibility of the provider of that infrastructure to make sure that data's durable, how you protect your data across 15 different AWS services, as well as on-premises VMware And like I said, the intent, but you can support a federated data model cuz you're building both in the cloud, on the edge, at the edge and on premises. data mesh, but the problem is how do you ensure governance and a federated model? along the lines of the, this framework, we really think about, you know, how do customers identify you know, we know that we have to take an action and recover having that plan in place, you know, necessary to apply all these best practices and, And specifically on the recovery piece, you mentioned AWS backup, you couldn't even test things like disaster recovery, right? I just wanna kind of tie up that statement is, you know, one thing that's really interesting Cuz most of the time, many times you don't even get your data back. So at the end of the day, you know, there's this expressions we use What's what excites you about data and data protection? at AWS reinforc is the ability to derive value from secondary data, you know, D deploying, compute, storage, and networking now bringing up to the data level,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
NancyPERSON

0.99+

Nancy WongPERSON

0.99+

DavePERSON

0.99+

Steven SchmidtPERSON

0.99+

AWSORGANIZATION

0.99+

Dave ValantePERSON

0.99+

AmazonORGANIZATION

0.99+

EuropeLOCATION

0.99+

WaynePERSON

0.99+

BostonLOCATION

0.99+

15QUANTITY

0.99+

Kurt kufeldPERSON

0.99+

CJ BoesPERSON

0.99+

Nancy WangPERSON

0.99+

Robert GatesPERSON

0.99+

twoQUANTITY

0.99+

last yearDATE

0.99+

GatesPERSON

0.99+

first lineQUANTITY

0.99+

second pillarQUANTITY

0.99+

oneQUANTITY

0.99+

Wayne DusoPERSON

0.99+

bothQUANTITY

0.98+

15 zerosQUANTITY

0.98+

one thumbprintQUANTITY

0.98+

one signatureQUANTITY

0.97+

two core pillarsQUANTITY

0.96+

early last decadeDATE

0.96+

three common themesQUANTITY

0.95+

a monthQUANTITY

0.9+

secondQUANTITY

0.88+

couple months agoDATE

0.85+

Dr.PERSON

0.84+

two cubeQUANTITY

0.77+

VMwareTITLE

0.71+

Day 2022EVENT

0.71+

threeQUANTITY

0.66+

yearsDATE

0.65+

gameEVENT

0.57+

dayEVENT

0.52+

2022DATE

0.45+

CubeORGANIZATION

0.35+

Wayne Durso & Nancy Wang | AWS Storage Day 2022


 

[Music] okay we're back my name is dave vellante and this is thecube's coverage of aws storage day you know coming off of reinforce i wrote that the cloud was a new layer of defense in fact the first line of defense in a cyber security strategy that brings new thinking and models for protecting data data protection specifically traditionally thought of as backup and recovery it's become a critical adjacency to security and a component of a comprehensive cyber security strategy we're here in our studios outside of boston with two cube alums and we're going to discuss this and other topics wayne dusso is the vice president for aws storage edge and data services and nancy wong as general manager of aws backup and data protection services guys welcome great to see you again thanks for coming on of course always a pleasure dave good to see you dave all right so wayne let's talk about how organizations should be thinking about this term data protection it's an expanding definition isn't it it is an expanded definition dave last year we talked about uh data and the importance of data to companies every company um is becoming a data company uh you know the amount of data they generate uh the amount of data they can use to uh create models to do predictive analytics and frankly uh to find ways of innovating uh is is growing uh rapidly and you know there's this tension between access to all that data right getting the value out of that data and how do you secure that data and so this is something we think about with customers all the time so data durability data protection data resiliency and you know trust in their data if you think about running your organization on your data trust in your data is so important so you know you got to trust where you're putting your data you know people who are putting their data on a platform need to trust that platform will in fact ensure its durability security resiliency and you know we see ourselves uh aws as a partner uh in securing their data making their data they're built durable making their data resilient all right so some of that responsibility is on us some of that is on amazon responsibility around data protection data resiliency and you know um we think about forever you know the notion of um you know compromise of your infrastructure but more and more people think about the compromise of their data as data becomes more valuable in fact data is a company's most valuable asset we've talked about this before only second to their people you know the people who are the most valuable asset but right next to that is their data so really important stuff so nancy you talk to a lot of customers but by the way it always comes back to the data we've been saying this for years haven't we so you've got this expanding definition of data protection you know governance is in there you think about access etc when you talk to customers what are you hearing from them how are they thinking about data protection yeah so a lot of the customers that wayne and i have spoken to often come to us seeking thought leadership about you know how do i solve this data challenge how do i solve this data sprawl challenge but also more importantly tying it back to data protection and data resiliency is how do i make sure that data is secure that it's protected against let's say ransomware events right and continuously protected so there's a lot of mental frameworks that come to mind and a very popular one that comes up in quite a few conversations is in this cyber security framework right and from a data protection perspective it's just as important to protect and recover your data as it is to be able to detect different events or be able to respond to those events right so recently i was just having a conversation with a regulatory body of financial institutions in europe where we're designing a architecture that could help them make their data immutable but also continuously protected so taking a step back that's really where i see aws's role in that we provide a wide breadth of primitives to help customers build secure platforms and scaffolding so that they can focus on building the data protection the data governance controls and guardrails on top of that platform and that's always been aws philosophy make sure that developers have access to those primitives and apis so that they can move fast and essentially build their own if that that's in fact what they want to do and as you're saying when data protection is now this adjacency to cyber security but there's disaster recoveries in there business continuance cyber resilience etc so so maybe you could pick up on that and sort of extend how you see aws helping customers build out those resilient services yeah so you know two uh core pillars to a data protection strategy is around their data durability which is really an infrastructural element you know it's it's it's by and large the responsibility of the provided that infrastructure to make sure that data is durable because if it's not durable and everything else doesn't matter um and the second pillar is really about data resiliency so in terms of security controls and governance like these are really important but these are a shared responsibility like the customers working with us with the services that we provide are there to architect the design it's really human factors and design factors that get them resiliency nancy anything you would add to what wayne just said yeah absolutely so customers tell us that they want always on data resiliency and data durability right so oftentimes in those conversations three common themes come up which is they want a centralized solution they want to be able to transcribe their intent into what they end up doing with their data and number three they want something that's policy driven because once you centralize your policies it's much better and easier to establish control and governance at an organizational level so keeping that in mind with policy as our interface there's two managed aws solutions that i recommend you all check out in terms of data resiliency and data durability those are aws backup which is our centralized solution for managing protection recovery and also provides an audit audit capability of how you protect your data across 15 different aws services as well as on-premises vmware and for customers whose mission-critical data is contained entirely on disk we also offer aws elastic disaster recovery services especially for customers who want to fail over their workloads from on-premises to the cloud so you can essentially centralize as a quick follow-up centralize the policy and as you said the intent but you can support a federated data model because you're building out this massive you know global system but you can take that policy and essentially bring it anywhere on the aws cloud is that right exactly and actually one powerful integration i want to touch upon is that aws backup is natively integrated with aws organizations which is our de facto multi-account federated organization model for how aws services work with customers both in the cloud on the edge at the edge and on premises so that's really important because as we talk about all the time on the cube this notion of a decentralized data architecture data mesh but the problem is how do you ensure governance in a federated model so we're clearly moving in that direction when i want to ask you about cyber as a board level discussion years ago i interviewed dr robert gates you know former defense secretary and he sat on a number of boards and i asked him you know how important and prominent is security at the board level is it really a board level discussion he said absolutely every time we meet we talk about cyber security but not every company at the time this was kind of early last decade was doing that that's changed um now ransomware is front and center hear about it all the time what's aws what's your thinking on cyber as a board level discussion and specifically what are you guys doing around ransomware yeah so you know malware in general ransomware being a particular type of malware um it's a hot topic and it continues to be a hot topic and whether at the board level the c-suite level um i had a chance to listen to uh dr gates a couple months ago and uh it was super motivational um but we think about ransomware in the same way that our customers do right because all of us are subject to an incident nobody is uh uh immune to a ransomware incident so we think very much the same way and as nancy said along the lines of the nist framework we really think about you know how do customers identify their critical access how do they plan for protecting those assets right how do they make sure that they are in fact protected and if they do detect a ransomware event and ransomware events come from a lot of different places like there's not one signature there's not one thumb print if you would for ransomware so it's it's there's really a lot of vigilance uh that needs to be put in place but a lot of planning that needs to be put in place and once that's detected and a we have to recover you know we know that we have to take an action and recover having that plan in place making sure that your assets are fully protected and can be restored as you know ransomware is a insidious uh type of malware you know it sits in your system for a long time it figures out what's going on including your backup policies your protection policies and figures out how to get around those with some of the things that nancy talked about in terms of air gapping your capabilities being able to if you would scan your secondary your backup storage for malware knowing that it's a good copy and then being able to restore from that known good copy in the event of an incident is critical so we think about this for ourselves in the same way that we think about these for our customers you've got to have a great plan you've got to have great protection and you've got to be ready to restore in the case of an incident and we want to make sure we provide all the capabilities to do that yeah so i'm glad you mentioned air gapping so at the recent reinforce i think it was kurt kufeld was speaking about ransomware and he didn't specifically mention air gapping i had to leave so i might i might have missed it because i'm doing the cube but that's a that's a key aspect i'm sure there were things in the on the deep dives that addressed air gapping but nancy look aws has the skills it has the resources you know necessary to apply all these best practices and you know share those as customers but but what specific investments is aws making to make the cso's life easier maybe you could talk about that sure so following on to your point about the reinforced keynote dave right cj moses talked about how the events of a ransomware for example incident or event can take place right on stage where you go from detect to respond and to recover and specifically on the recover piece he mentioned aws backup the managed service that protects across 15 different aws services as well as on-premises vmware as automated recovery and that's in part why we've decided to continue that investment and deliver aws backup audit manager which helps customers actually prove their posture against how their protection policies are actually mapping back to their organizational controls based on for example how they tag their data for mission criticality or how sensitive that data is right and so turning to best practices especially for ransomware events since this is very top of mind for a lot of customers these days is i will always try to encourage customers to go through game day simulations for example identifying which are those most critical applications in their environment that they need up and running for their business to function properly for example and actually going through the recovery plan and making sure that their staff is well trained or that they're able to go through for example a security orchestration automation recovery solution to make sure that all of their mission critical applications are back up and running in case of a ransomware event yeah so i love the game date thing i mean we know well just in the history of it you couldn't even test things like disaster recovery be right because it was too dangerous with the cloud you can test these things safely and actually plan out develop a blueprint test your blueprint i love the the game day analogy yeah and actually one thing i love to add is you know we talked about air gapping i just want to kind of tie up that statement is you know one thing that's really interesting about the way that the aws cloud is architected is the identity access and management platform actually allows us to create identity constructs that air gap your data perimeter so that way when attackers for example are able to gain a foothold in your environment you're still able to air gap your most mission critical and also crown jewels from being infiltrated that's key yeah we've learned you know when paying the ransom is not a good strategy right because most of the time many times you don't even get your data back okay so we we're kind of data geeks here we love data um and we're passionate about it on the cube aws and you guys specifically are passionate about it so what excites you wayne you start and then nancy you bring us home what excites you about data and data protection and why you know we are data nerds uh so at the end of the day um you know there's there's expressions we use all the time but data is such a rich asset for all of us some of the greatest innovations that come out of aws comes out of our analysis of our own data like we collect a lot of data on our operations and some of our most critical features for our customers come out of our analysis that data so we are data nerds and we understand how businesses uh view their data because we view our data the same way so you know dave security really started in the data center it started with the enterprises and if we think about security often we talk about securing compute and securing network and you know if you if you secured your compute you secured your data generally but we've separated data from compute so that people can get the value from their data no matter how they want to use it and in doing that we have to make sure that their data is durable and it's resilient to any sort of incident event so this is really really important to us and what do i get excited about um you know again thinking back to this framework i know that we as thought leaders alongside our customers who also thought leaders in their space can provide them with the capabilities they need to protect their data to secure their data to make sure it's compliant and always always always durable you know it's funny you'd say it's not funny it's serious actually steven schmidt uh at reinforce he's the the chief security officer at amazon used to be the c c iso of aws he said that amazon sees quadrillions of data points a month that's 15 zeros okay so that's a lot of data nancy bring us home what's what excites you about data and data protection yeah so specifically and this is actually drawing from conversations that i had with multiple isv partners at aws reinforce is the ability to derive value from secondary data right because traditionally organizations have really seen that as a cost center right you're producing secondary data because most likely you're creating backups of your mission critical workloads but what if you're able to run analytics and insights and derive insights from that secondary data right then you're actually able to let aws do the undifferentiated heavy lifting of analyzing that secondary data as state so that way you as customers or isv partners can build value on the security layers above and that is how we see turning cost into value i love it you're taking the original premise of the cloud taking away the undifferentiated heavy lifting for you know deploying compute storage and networking now bringing up to the data level the analytics level so it continues the cloud continues to expand thank you for watching thecube's coverage of aws storage day 2022

Published Date : Aug 5 2022

**Summary and Sentiment Analysis are not been shown because of improper transcript**

ENTITIES

EntityCategoryConfidence
amazonORGANIZATION

0.99+

awsORGANIZATION

0.99+

kurt kufeldPERSON

0.99+

europeLOCATION

0.99+

last yearDATE

0.99+

bostonLOCATION

0.99+

wayne dussoPERSON

0.99+

steven schmidtPERSON

0.99+

Nancy WangPERSON

0.99+

twoQUANTITY

0.98+

Wayne DursoPERSON

0.98+

uh awsORGANIZATION

0.98+

first lineQUANTITY

0.97+

AWSORGANIZATION

0.97+

dave vellantePERSON

0.97+

davePERSON

0.97+

one signatureQUANTITY

0.97+

three common themesQUANTITY

0.96+

one thumbQUANTITY

0.96+

waynePERSON

0.96+

nancyPERSON

0.95+

second pillarQUANTITY

0.94+

15 zerosQUANTITY

0.94+

oneQUANTITY

0.92+

15 differentQUANTITY

0.92+

bothQUANTITY

0.92+

dr robert gatesPERSON

0.91+

secondQUANTITY

0.91+

a monthQUANTITY

0.9+

one thingQUANTITY

0.88+

vmwareTITLE

0.81+

a couple months agoDATE

0.81+

early last decadeDATE

0.8+

years agoDATE

0.78+

lot of customersQUANTITY

0.76+

lotQUANTITY

0.76+

15 differentQUANTITY

0.74+

a lot of customersQUANTITY

0.74+

dr gatesPERSON

0.67+

day 2022EVENT

0.65+

dataQUANTITY

0.63+

cubeORGANIZATION

0.63+

ransomwareTITLE

0.62+

nancyORGANIZATION

0.59+

threeQUANTITY

0.54+

Day 2022EVENT

0.53+

yearsQUANTITY

0.48+

coreQUANTITY

0.48+

nancy wongPERSON

0.47+

thecubePERSON

0.47+

cloudTITLE

0.36+

AWS Storage Day 2022 Intro


 

(upbeat music) >> Welcome to theCUBE's coverage of AWS Storage Day 2022. My name is Dave Vellante. In 2021, theCUBE team was in Seattle covering Storage Day. And after that event, I wrote a breaking analysis piece on Wikibon and SiliconANGLE called "Thinking Outside The Box, "AWS Signals A New Era For Storage." And the point of that post was that the cloud's impact was clearly moving into the storage realm in a big way. And the days of consuming storage as a box were numbered. And I projected, AWS doesn't share these numbers but I projected that AWS's storage business was on track to hit $10 billion, making it the second largest purveyor of storage with a gross trajectory that by mid-decade would make AWS the number one storage player in the market. Now, a lot of people didn't like that post, particularly the fact that I was mixing AWS storage service, OpEx, with what generally were CapEx purchases. But I didn't really care to argue the nuance of CapEx versus OpEx. Rather, the point I was really making was, and I was looking at the spending data from ETR and estimating the revenue for the players, and the message was clear. Data was moving to and being created in the cloud much faster than on-prem and the spending patterns were following data growth. Now, fast forward almost 12 months and the picture is even more clear to me. The number of cloud storage services from AWS is expanding as is their consequent adoption. The pace of delivery is accelerating. And very importantly, the optionality of the ecosystem is exploding. Virtually every storage company, primary, secondary, data protection, archival, is partnering with AWS to run their services in the cloud and in many cases connect to their on-prem installations, expanding the cloud as we've talked about and written about extensively. Despite the narrative from some about repatriation and people moving out of the cloud back on-prem, such activity is a rounding error in the grand scheme of enterprise tech spending. The data is clear, cloud and cloud storage spending continues to grow at 30% plus per year, far ahead of any other markets. Now, the edge presents new opportunities and likely will bring novel architectures as we've predicted many times covering what AWS is doing with the Arm-based Graviton and others. Now, this is especially important at the far edge, like real-time AI inferencing and new workloads. You know, there's questions that remain about how much storage is going to persist at the edge, how much is going to go back into the cloud, and what requirements exist across the board. But in many respects, the edge is all incremental in terms of data growth and data creation. So the challenge is how do we harness the power of that data? So what can we expect going forward in storage? Well, the pace of service delivery from hyperscale providers generally and AWS specifically is going to continue to accelerate. AWS is likely going to lead the way. We've seen this, started with S3, expand storage portfolio into block and file, and then bringing cohort services like new compute architectures, we've talked about Nitro and Graviton and others, and a portfolio of database options and new machine intelligence, machine learning, and AI solutions. Storage in the cloud is moving from being a bit bucket to being a platform that is evolving as part of an emerging data mesh architecture where business users, those with context, gain secure, governed, and facile self-service access to data that they need when they need it so they can make better decisions and importantly create new data products and services. This is the vision for data generally in the 2020s and cloud storage specifically will be an underpinning of this new era. Thanks for watching theCUBE's coverage of AWS Storage Day. This is Dave Vellante. (upbeat music)

Published Date : Aug 5 2022

SUMMARY :

and the message was clear.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Dave VellantePERSON

0.99+

AWSORGANIZATION

0.99+

$10 billionQUANTITY

0.99+

SeattleLOCATION

0.99+

2021DATE

0.99+

2020sDATE

0.99+

Storage DayEVENT

0.97+

theCUBEORGANIZATION

0.96+

NitroORGANIZATION

0.95+

S3TITLE

0.94+

ETRORGANIZATION

0.93+

GravitonORGANIZATION

0.92+

almost 12 monthsQUANTITY

0.91+

SiliconANGLEORGANIZATION

0.9+

WikibonORGANIZATION

0.85+

Storage Day 2022EVENT

0.84+

30% plusQUANTITY

0.83+

number oneQUANTITY

0.82+

CapExTITLE

0.82+

second largest purveyorQUANTITY

0.81+

AWSEVENT

0.8+

Thinking Outside The BoxTITLE

0.78+

OpExTITLE

0.76+

AWS Storage DayEVENT

0.69+

OpExORGANIZATION

0.62+

midQUANTITY

0.55+

decadeQUANTITY

0.47+

GravitonCOMMERCIAL_ITEM

0.42+