Image Title

Search Results for each computing:

Breaking Analysis: Google's Point of View on Confidential Computing


 

>> From theCUBE studios in Palo Alto in Boston, bringing you data-driven insights from theCUBE and ETR. This is Breaking Analysis with Dave Vellante. >> Confidential computing is a technology that aims to enhance data privacy and security by providing encrypted computation on sensitive data and isolating data from apps in a fenced off enclave during processing. The concept of confidential computing is gaining popularity, especially in the cloud computing space where sensitive data is often stored and of course processed. However, there are some who view confidential computing as an unnecessary technology in a marketing ploy by cloud providers aimed at calming customers who are cloud phobic. Hello and welcome to this week's Wikibon CUBE Insights powered by ETR. In this Breaking Analysis, we revisit the notion of confidential computing, and to do so, we'll invite two Google experts to the show, but before we get there, let's summarize briefly. There's not a ton of ETR data on the topic of confidential computing. I mean, it's a technology that's deeply embedded into silicon and computing architectures. But at the highest level, security remains the number one priority being addressed by IT decision makers in the coming year as shown here. And this data is pretty much across the board by industry, by region, by size of company. I mean we dug into it and the only slight deviation from the mean is in financial services. The second and third most cited priorities, cloud migration and analytics, are noticeably closer to cybersecurity in financial services than in other sectors, likely because financial services has always been hyper security conscious, but security is still a clear number one priority in that sector. The idea behind confidential computing is to better address threat models for data in execution. Protecting data at rest and data and transit have long been a focus of security approaches, but more recently, silicon manufacturers have introduced architectures that separate data and applications from the host system. Arm, Intel, AMD, Nvidia and other suppliers are all on board, as are the big cloud players. Now the argument against confidential computing is that it narrowly focuses on memory encryption and it doesn't solve the biggest problems in security. Multiple system images updates different services and the entire code flow aren't directly addressed by memory encryption, rather to truly attack these problems, many believe that OSs need to be re-engineered with the attacker and hacker in mind. There are so many variables and at the end of the day, critics say the emphasis on confidential computing made by cloud providers is overstated and largely hype. This tweet from security researcher Rodrigo Branco sums up the sentiment of many skeptics. He says, "Confidential computing is mostly a marketing campaign for memory encryption. It's not driving the industry towards the hard open problems. It is selling an illusion." Okay. Nonetheless, encrypting data in use and fencing off key components of the system isn't a bad thing, especially if it comes with the package essentially for free. There has been a lack of standardization and interoperability between different confidential computing approaches. But the confidential computing consortium was established in 2019 ostensibly to accelerate the market and influence standards. Notably, AWS is not part of the consortium, likely because the politics of the consortium were probably a conundrum for AWS because the base technology defined by the the consortium is seen as limiting by AWS. This is my guess, not AWS's words, and but I think joining the consortium would validate a definition which AWS isn't aligned with. And two, it's got a lead with this Annapurna acquisition. This was way ahead with Arm integration and so it probably doesn't feel the need to validate its competitors. Anyway, one of the premier members of the confidential computing consortium is Google, along with many high profile names including Arm, Intel, Meta, Red Hat, Microsoft, and others. And we're pleased to welcome two experts on confidential computing from Google to unpack the topic, Nelly Porter is head of product for GCP confidential computing and encryption, and Dr. Patricia Florissi is the technical director for the office of the CTO at Google Cloud. Welcome Nelly and Patricia, great to have you. >> Great to be here. >> Thank you so much for having us. >> You're very welcome. Nelly, why don't you start and then Patricia, you can weigh in. Just tell the audience a little bit about each of your roles at Google Cloud. >> So I'll start, I'm owning a lot of interesting activities in Google and again security or infrastructure securities that I usually own. And we are talking about encryption and when encryption and confidential computing is a part of portfolio in additional areas that I contribute together with my team to Google and our customers is secure software supply chain. Because you need to trust your software. Is it operate in your confidential environment to have end-to-end story about if you believe that your software and your environment doing what you expect, it's my role. >> Got it. Okay. Patricia? >> Well, I am a technical director in the office of the CTO, OCTO for short, in Google Cloud. And we are a global team. We include former CTOs like myself and senior technologists from large corporations, institutions and a lot of success, we're startups as well. And we have two main goals. First, we walk side by side with some of our largest, more strategic or most strategical customers and we help them solve complex engineering technical problems. And second, we are devise Google and Google Cloud engineering and product management and tech on there, on emerging trends and technologies to guide the trajectory of our business. We are unique group, I think, because we have created this collaborative culture with our customers. And within OCTO, I spend a lot of time collaborating with customers and the industry at large on technologies that can address privacy, security, and sovereignty of data in general. >> Excellent. Thank you for that both of you. Let's get into it. So Nelly, what is confidential computing? From Google's perspective, how do you define it? >> Confidential computing is a tool and it's still one of the tools in our toolbox. And confidential computing is a way how we would help our customers to complete this very interesting end-to-end lifecycle of the data. And when customers bring in the data to cloud and want to protect it as they ingest it to the cloud, they protect it at rest when they store data in the cloud. But what was missing for many, many years is ability for us to continue protecting data and workloads of our customers when they running them. And again, because data is not brought to cloud to have huge graveyard, we need to ensure that this data is actually indexed. Again, there is some insights driven and drawn from this data. You have to process this data and confidential computing here to help. Now we have end to end protection of our customer's data when they bring the workloads and data to cloud, thanks to confidential computing. >> Thank you for that. Okay, we're going to get into the architecture a bit, but before we do, Patricia, why do you think this topic of confidential computing is such an important technology? Can you explain, do you think it's transformative for customers and if so, why? >> Yeah, I would maybe like to use one thought, one way, one intuition behind why confidential commuting matters, because at the end of the day, it reduces more and more the customer's thresh boundaries and the attack surface. That's about reducing that periphery, the boundary in which the customer needs to mind about trust and safety. And in a way, is a natural progression that you're using encryption to secure and protect the data. In the same way that we are encrypting data in transit and at rest, now we are also encrypting data while in use. And among other beneficials, I would say one of the most transformative ones is that organizations will be able to collaborate with each other and retain the confidentiality of the data. And that is across industry, even though it's highly focused on, I wouldn't say highly focused, but very beneficial for highly regulated industries. It applies to all of industries. And if you look at financing for example, where bankers are trying to detect fraud, and specifically double finance where you are, a customer is actually trying to get a finance on an asset, let's say a boat or a house, and then it goes to another bank and gets another finance on that asset. Now bankers would be able to collaborate and detect fraud while preserving confidentiality and privacy of the data. >> Interesting. And I want to understand that a little bit more but I'm going to push you a little bit on this, Nelly, if I can because there's a narrative out there that says confidential computing is a marketing ploy, I talked about this upfront, by cloud providers that are just trying to placate people that are scared of the cloud. And I'm presuming you don't agree with that, but I'd like you to weigh in here. The argument is confidential computing is just memory encryption and it doesn't address many other problems. It is over hyped by cloud providers. What do you say to that line of thinking? >> I absolutely disagree, as you can imagine, with this statement, but the most importantly is we mixing multiple concepts, I guess. And exactly as Patricia said, we need to look at the end-to-end story, not again the mechanism how confidential computing trying to again, execute and protect a customer's data and why it's so critically important because what confidential computing was able to do, it's in addition to isolate our tenants in multi-tenant environments the cloud covering to offer additional stronger isolation. They called it cryptographic isolation. It's why customers will have more trust to customers and to other customers, the tenant that's running on the same host but also us because they don't need to worry about against threats and more malicious attempts to penetrate the environment. So what confidential computing is helping us to offer our customers, stronger isolation between tenants in this multi-tenant environment, but also incredibly important, stronger isolation of our customers, so tenants from us. We also writing code, we also software providers will also make mistakes or have some zero days. Sometimes again us introduced, sometimes introduced by our adversaries. But what I'm trying to say by creating this cryptographic layer of isolation between us and our tenants and amongst those tenants, we're really providing meaningful security to our customers and eliminate some of the worries that they have running on multi-tenant spaces or even collaborating to gather this very sensitive data knowing that this particular protection is available to them. >> Okay, thank you. Appreciate that. And I think malicious code is often a threat model missed in these narratives. Operator access, yeah, maybe I trust my clouds provider, but if I can fence off your access even better, I'll sleep better at night. Separating a code from the data, everybody's, Arm, Intel, AMD, Nvidia, others, they're all doing it. I wonder if, Nelly, if we could stay with you and bring up the slide on the architecture. What's architecturally different with confidential computing versus how operating systems and VMs have worked traditionally. We're showing a slide here with some VMs, maybe you could take us through that. >> Absolutely. And Dave, the whole idea for Google and now industry way of dealing with confidential computing is to ensure that three main property is actually preserved. Customers don't need to change the code. They can operate on those VMs exactly as they would with normal non-confidential VMs, but to give them this opportunity of lift and shift or no changing their apps and performing and having very, very, very low latency and scale as any cloud can, something that Google actually pioneer in confidential computing. I think we need to open and explain how this magic was actually done. And as I said, it's again the whole entire system have to change to be able to provide this magic. And I would start with we have this concept of root of trust and root of trust where we will ensure that this machine, when the whole entire post has integrity guarantee, means nobody changing my code on the most low level of system. And we introduce this in 2017 called Titan. It was our specific ASIC, specific, again, inch by inch system on every single motherboard that we have that ensures that your low level former, your actually system code, your kernel, the most powerful system is actually proper configured and not changed, not tampered. We do it for everybody, confidential computing included. But for confidential computing, what we have to change, we bring in AMD, or again, future silicon vendors and we have to trust their former, their way to deal with our confidential environments. And that's why we have obligation to validate integrity, not only our software and our former but also former and software of our vendors, silicon vendors. So we actually, when we booting this machine, as you can see, we validate that integrity of all of the system is in place. It means nobody touching, nobody changing, nobody modifying it. But then we have this concept of AMD secure processor, it's special ASICs, best specific things that generate a key for every single VM that our customers will run or every single node in Kubernetes or every single worker thread in our Hadoop or Spark capability. We offer all of that. And those keys are not available to us. It's the best keys ever in encryption space because when we are talking about encryption, the first question that I'm receiving all the time, where's the key, who will have access to the key? Because if you have access to the key then it doesn't matter if you encrypted or not. So, but the case in confidential computing provides so revolutionary technology, us cloud providers, who don't have access to the keys. They sitting in the hardware and they head to memory controller. And it means when hypervisors that also know about these wonderful things saying I need to get access to the memories that this particular VM trying to get access to, they do not decrypt the data, they don't have access to the key because those keys are random, ephemeral and per VM, but the most importantly, in hardware not exportable. And it means now you would be able to have this very interesting role that customers or cloud providers will not be able to get access to your memory. And what we do, again, as you can see our customers don't need to change their applications, their VMs are running exactly as it should run and what you're running in VM, you actually see your memory in clear, it's not encrypted, but God forbid is trying somebody to do it outside of my confidential box. No, no, no, no, no, they would not be able to do it. Now you'll see cyber and it's exactly what combination of these multiple hardware pieces and software pieces have to do. So OS is also modified. And OS is modified such way to provide integrity. It means even OS that you're running in your VM box is not modifiable and you, as customer, can verify. But the most interesting thing, I guess, how to ensure the super performance of this environment because you can imagine, Dave, that encrypting and it's additional performance, additional time, additional latency. So we were able to mitigate all of that by providing incredibly interesting capability in the OS itself. So our customers will get no changes needed, fantastic performance and scales as they would expect from cloud providers like Google. >> Okay, thank you. Excellent. Appreciate that explanation. So, again, the narrative on this as well, you've already given me guarantees as a cloud provider that you don't have access to my data, but this gives another level of assurance, key management as they say is key. Now humans aren't managing the keys, the machines are managing them. So Patricia, my question to you is, in addition to, let's go pre confidential computing days, what are the sort of new guarantees that these hardware-based technologies are going to provide to customers? >> So if I am a customer, I am saying I now have full guarantee of confidentiality and integrity of the data and of the code. So if you look at code and data confidentiality, the customer cares and they want to know whether their systems are protected from outside or unauthorized access, and that recovered with Nelly, that it is. Confidential computing actually ensures that the applications and data internals remain secret, right? The code is actually looking at the data, the only the memory is decrypting the data with a key that is ephemeral and per VM and generated on demand. Then you have the second point where you have code and data integrity, and now customers want to know whether their data was corrupted, tampered with or impacted by outside actors. And what confidential computing ensures is that application internals are not tampered with. So the application, the workload as we call it, that is processing the data, it's also, it has not been tampered and preserves integrity. I would also say that this is all verifiable. So you have attestation and these attestation actually generates a log trail and the log trail guarantees that, provides a proof that it was preserved. And I think that the offer's also a guarantee of what we call ceiling, this idea that the secrets have been preserved and not tampered with, confidentiality and integrity of code and data. >> Got it. Okay, thank you. Nelly, you mentioned, I think I heard you say that the applications, it's transparent, you don't have to change the application, it just comes for free essentially. And we showed some various parts of the stack before. I'm curious as to what's affected, but really more importantly, what is specifically Google's value add? How do partners participate in this, the ecosystem, or maybe said another way, how does Google ensure the compatibility of confidential computing with existing systems and applications? >> And a fantastic question by the way. And it's very difficult and definitely complicated world because to be able to provide these guarantees, actually a lot of work was done by community. Google is very much operate in open, so again, our operating system, we working with operating system repository OSs, OS vendors to ensure that all capabilities that we need is part of the kernels, are part of the releases and it's available for customers to understand and even explore if they have fun to explore a lot of code. We have also modified together with our silicon vendors a kernel, host kernel to support this capability and it means working this community to ensure that all of those patches are there. We also worked with every single silicon vendor as you've seen, and that's what I probably feel that Google contributed quite a bit in this whole, we moved our industry, our community, our vendors to understand the value of easy to use confidential computing or removing barriers. And now I don't know if you noticed, Intel is pulling the lead and also announcing their trusted domain extension, very similar architecture. And no surprise, it's, again, a lot of work done with our partners to, again, convince, work with them and make this capability available. The same with Arm this year, actually last year, Arm announced their future design for confidential computing. It's called Confidential Computing Architecture. And it's also influenced very heavily with similar ideas by Google and industry overall. So it's a lot of work in confidential computing consortiums that we are doing, for example, simply to mention, to ensure interop, as you mentioned, between different confidential environments of cloud providers. They want to ensure that they can attest to each other because when you're communicating with different environments, you need to trust them. And if it's running on different cloud providers, you need to ensure that you can trust your receiver when you are sharing your sensitive data workloads or secret with them. So we coming as a community and we have this attestation sig, the, again, the community based systems that we want to build and influence and work with Arm and every other cloud providers to ensure that we can interrupt and it means it doesn't matter where confidential workloads will be hosted, but they can exchange the data in secure, verifiable and controlled by customers way. And to do it, we need to continue what we are doing, working open, again, and contribute with our ideas and ideas of our partners to this role to become what we see confidential computing has to become, it has to become utility. It doesn't need to be so special, but it's what we want it to become. >> Let's talk about, thank you for that explanation. Let's talk about data sovereignty because when you think about data sharing, you think about data sharing across the ecosystem and different regions and then of course data sovereignty comes up. Typically public policy lags, the technology industry and sometimes is problematic. I know there's a lot of discussions about exceptions, but Patricia, we have a graphic on data sovereignty. I'm interested in how confidential computing ensures that data sovereignty and privacy edicts are adhered to, even if they're out of alignment maybe with the pace of technology. One of the frequent examples is when you delete data, can you actually prove that data is deleted with a hundred percent certainty? You got to prove that and a lot of other issues. So looking at this slide, maybe you could take us through your thinking on data sovereignty. >> Perfect. So for us, data sovereignty is only one of the three pillars of digital sovereignty. And I don't want to give the impression that confidential computing addresses it all. That's why we want to step back and say, hey, digital sovereignty includes data sovereignty where we are giving you full control and ownership of the location, encryption and access to your data. Operational sovereignty where the goal is to give our Google Cloud customers full visibility and control over the provider operations, right? So if there are any updates on hardware, software stack, any operations, there is full transparency, full visibility. And then the third pillar is around software sovereignty where the customer wants to ensure that they can run their workloads without dependency on the provider's software. So they have sometimes is often referred as survivability, that you can actually survive if you are untethered to the cloud and that you can use open source. Now let's take a deep dive on data sovereignty, which by the way is one of my favorite topics. And we typically focus on saying, hey, we need to care about data residency. We care where the data resides because where the data is at rest or in processing, it typically abides to the jurisdiction, the regulations of the jurisdiction where the data resides. And others say, hey, let's focus on data protection. We want to ensure the confidentiality and integrity and availability of the data, which confidential computing is at the heart of that data protection. But it is yet another element that people typically don't talk about when talking about data sovereignty, which is the element of user control. And here, Dave, is about what happens to the data when I give you access to my data. And this reminds me of security two decades ago, even a decade ago, where we started the security movement by putting firewall protections and login accesses. But once you were in, you were able to do everything you wanted with the data. An insider had access to all the infrastructure, the data and the code. And that's similar because with data sovereignty we care about whether it resides, where, who is operating on the data. But the moment that the data is being processed, I need to trust that the processing of the data will abide by user control, by the policies that I put in place of how my data is going to be used. And if you look at a lot of the regulation today and a lot of the initiatives around the International Data Space Association, IDSA, and Gaia-X, there is a movement of saying the two parties, the provider of the data and the receiver of the data are going to agree on a contract that describes what my data can be used for. The challenge is to ensure that once the data crosses boundaries, that the data will be used for the purposes that it was intended and specified in the contract. And if you actually bring together, and this is the exciting part, confidential computing together with policy enforcement, now the policy enforcement can guarantee that the data is only processed within the confines of a confidential computing environment, that the workload is cryptographically verified that there is the workload that was meant to process the data and that the data will be only used when abiding to the confidentiality and integrity safety of the confidential computing environment. And that's why we believe confidential computing is one necessary and essential technology that will allow us to ensure data sovereignty, especially when it comes to user control. >> Thank you for that. I mean it was a deep dive, I mean brief, but really detailed. So I appreciate that, especially the verification of the enforcement. Last question, I met you two because as part of my year end prediction post, you guys sent in some predictions and I wasn't able to get to them in the predictions post. So I'm thrilled that you were able to make the time to come on the program. How widespread do you think the adoption of confidential computing will be in 23 and what's the maturity curve look like, this decade in your opinion? Maybe each of you could give us a brief answer. >> So my prediction in five, seven years, as I started, it'll become utility. It'll become TLS as of, again, 10 years ago we couldn't believe that websites will have certificates and we will support encrypted traffic. Now we do and it's become ubiquity. It's exactly where confidential computing is getting and heading, I don't know we deserve yet. It'll take a few years of maturity for us, but we will be there. >> Thank you. And Patricia, what's your prediction? >> I will double that and say, hey, in the future, in the very near future, you will not be able to afford not having it. I believe as digital sovereignty becomes evermore top of mind with sovereign states and also for multi national organizations and for organizations that want to collaborate with each other, confidential computing will become the norm. It'll become the default, if I say, mode of operation. I like to compare that today is inconceivable. If we talk to the young technologists, it's inconceivable to think that at some point in history, and I happen to be alive that we had data at rest that was not encrypted, data in transit that was not encrypted, and I think that will be inconceivable at some point in the near future that to have unencrypted data while in use. >> And plus I think the beauty of the this industry is because there's so much competition, this essentially comes for free. I want to thank you both for spending some time on Breaking Analysis. There's so much more we could cover. I hope you'll come back to share the progress that you're making in this area and we can double click on some of these topics. Really appreciate your time. >> Anytime. >> Thank you so much. >> In summary, while confidential computing is being touted by the cloud players as a promising technology for enhancing data privacy and security, there are also those, as we said, who remain skeptical. The truth probably lies somewhere in between and it will depend on the specific implementation and the use case as to how effective confidential computing will be. Look, as with any new tech, it's important to carefully evaluate the potential benefits, the drawbacks, and make informed decisions based on the specific requirements in the situation and the constraints of each individual customer. But the bottom line is silicon manufacturers are working with cloud providers and other system companies to include confidential computing into their architectures. Competition, in our view, will moderate price hikes. And at the end of the day, this is under the covers technology that essentially will come for free. So we'll take it. I want to thank our guests today, Nelly and Patricia from Google, and thanks to Alex Myerson who's on production and manages the podcast. Ken Schiffman as well out of our Boston studio, Kristin Martin and Cheryl Knight help get the word out on social media and in our newsletters. And Rob Hof is our editor-in-chief over at siliconangle.com. Does some great editing for us, thank you all. Remember all these episodes are available as podcasts. Wherever you listen, just search Breaking Analysis podcast. I publish each week on wikibon.com and siliconangle.com where you can get all the news. If you want to get in touch, you can email me at david.vellante@siliconangle.com or dm me @DVellante. And you can also comment on my LinkedIn post. Definitely you want to check out etr.ai for the best survey data in the enterprise tech business. I know we didn't hit on a lot today, but there's some amazing data and it's always being updated, so check that out. This is Dave Vellante for theCUBE Insights, powered by ETR. Thanks for watching and we'll see you next time on Breaking Analysis. (upbeat music)

Published Date : Feb 11 2023

SUMMARY :

bringing you data-driven and at the end of the day, Just tell the audience a little and confidential computing Got it. and the industry at large for that both of you. in the data to cloud into the architecture a bit, and privacy of the data. people that are scared of the cloud. and eliminate some of the we could stay with you and they head to memory controller. So, again, the narrative on this as well, and integrity of the data and of the code. how does Google ensure the compatibility and ideas of our partners to this role One of the frequent examples and that the data will be only used of the enforcement. and we will support encrypted traffic. And Patricia, and I happen to be alive beauty of the this industry and the constraints of

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
NellyPERSON

0.99+

PatriciaPERSON

0.99+

International Data Space AssociationORGANIZATION

0.99+

Alex MyersonPERSON

0.99+

AWSORGANIZATION

0.99+

IDSAORGANIZATION

0.99+

Rodrigo BrancoPERSON

0.99+

Dave VellantePERSON

0.99+

DavePERSON

0.99+

MicrosoftORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

NvidiaORGANIZATION

0.99+

2019DATE

0.99+

2017DATE

0.99+

Kristin MartinPERSON

0.99+

Nelly PorterPERSON

0.99+

Ken SchiffmanPERSON

0.99+

Rob HofPERSON

0.99+

Cheryl KnightPERSON

0.99+

last yearDATE

0.99+

Palo AltoLOCATION

0.99+

Red HatORGANIZATION

0.99+

two partiesQUANTITY

0.99+

AMDORGANIZATION

0.99+

Patricia FlorissiPERSON

0.99+

IntelORGANIZATION

0.99+

oneQUANTITY

0.99+

fiveQUANTITY

0.99+

second pointQUANTITY

0.99+

david.vellante@siliconangle.comOTHER

0.99+

MetaORGANIZATION

0.99+

secondQUANTITY

0.99+

thirdQUANTITY

0.99+

OneQUANTITY

0.99+

twoQUANTITY

0.99+

ArmORGANIZATION

0.99+

eachQUANTITY

0.99+

two expertsQUANTITY

0.99+

FirstQUANTITY

0.99+

first questionQUANTITY

0.99+

Gaia-XORGANIZATION

0.99+

two decades agoDATE

0.99+

bothQUANTITY

0.99+

this yearDATE

0.99+

seven yearsQUANTITY

0.99+

OCTOORGANIZATION

0.99+

zero daysQUANTITY

0.98+

10 years agoDATE

0.98+

each weekQUANTITY

0.98+

todayDATE

0.97+

Google's PoV on Confidential Computing NO PUB


 

>> Welcome Nelly and Patricia, great to have you. >> Great to be here. >> Thank you so much for having us. >> You're very welcome. Nelly, why don't you start, and then Patricia you can weigh in. Just tell the audience a little bit about each of your roles at Google Cloud. >> So I'll start, I'm honing a lot of interesting activities in Google and again, security or infrastructure securities that I usually hone, and we're talking about encryption, Antware encryption, and confidential computing is a part of portfolio. In additional areas that I contribute to get with my team to Google and our customers is secure software supply chain. Because you need to trust your software. Is it operating your confidential environment to have end to end story about if you believe that your software and your environment doing what you expect, it's my role. >> Got it, okay. Patricia? >> Well I am a technical director in the office of the CTO, OCTO for short, in Google Cloud. And we are a global team. We include former CTOs like myself and senior technologies from large corporations, institutions, and a lot of success for startups as well. And we have two main goals. First, we work side by side with some of our largest, more strategic or most strategic customers and we help them solve complex engineering technical problems. And second, we are device Google and Google Cloud engineering and product management on emerging trends in technologies to guide the trajectory of our business. We are unique group, I think, because we have created this collaborative culture with our customers. And within OCTO I spend a lot of time collaborating with customers in the industry at large on technologies that can address privacy, security, and sovereignty of data in general. >> Excellent, thank you for that both of you. Let's get into it. So Nelly, what is confidential computing from Google's perspective? How do you define it? >> Confidential computing is a tool. And it's one of the tools in our toolbox. And confidential computing is a way how would help our customers to complete this very interesting end to end lifecycle of their data. And when customers bring in the data to Cloud and want to protect it, as they ingest it to the Cloud, they protect it address when they store data in the Cloud. But what was missing for many, many years is ability for us to continue protecting data and workloads of our customers when they running them. And again, because data is not brought to Cloud to have huge graveyard, we need to ensure that this data is actually indexed. Again there is some insights driven and drawn from this data. You have to process this data and confidential computing here to help. Now we have end to end protection of our customer's data when they bring the workloads and data to Cloud, thanks to confidential computing. >> Thank you for that. Okay, we're going to get into the architecture a bit but before we do Patricia, why do you think this topic of confidential computing is such an important technology? Can you explain, do you think it's transformative for customers and if so, why? >> Yeah, I would maybe like to use one thought, one way, one intuition behind why confidential matters. Because at the end of the day it reduces more and more the customers thrush boundaries and the attack surface, that's about reducing that periphery, the boundary, in which the customer needs to mind about trust and safety. And in a way is a natural progression that you're using encryption to secure and protect data in the same way that we are encrypting data in transit and at rest. Now we are also encrypting data while in use. And among other beneficial I would say one of the most transformative ones is that organizations will be able to collaborate with each other and retain the confidentiality of the data. And that is across industry. Even though it's highly focused on, I wouldn't say highly focused, but very beneficial for highly regulated industries. It applies to all of industries. And if you look at financing for example, where bankers are trying to detect fraud and specifically double finance where you are a customer is actually trying to get a finance on an asset, let's say a boat or a house and then it goes to another bank and gets another finance on that asset. Now bankers would be able to collaborate and detect fraud while preserving confidentiality and privacy of the of the data. >> Interesting, and I want to understand that a little bit more but I'm going to push you a little bit on this, Nelly, if I can, because there's a narrative out there that says confidential computing is a marketing ploy. I talked about this upfront, by Cloud providers that are just trying to placate people that are scared of the Cloud. And I'm presuming you don't agree with that but I'd like you to weigh in here. The argument is confidential computing is just memory encryption, it doesn't address many other problems, it is overhyped by Cloud providers. What do you say to that line of thinking? >> I absolutely disagree as you can imagine, it's a crazy statement. But the most importantly is we mixing multiple concepts I guess. And exactly as Patricia said, we need to look at the end-to-end story not again the mechanism of how confidential computing trying to again execute and protect customer's data, and why it's so critically important. Because what confidential computing was able to do it's in addition to isolate our tenants in multi-tenant environments the Cloud over. To offer additional stronger isolation, we called it cryptographic isolation. It's why customers will have more trust to customers and to other customers, the tenants that's running on the same host but also us, because they don't need to worry about against threats and more malicious attempts to penetrate the environment. So what confidential computing is helping us to offer our customers, stronger isolation between tenants in this multi-tenant environment but also incredibly important, stronger isolation of our customers. So tenants from us, we also writing code, we also software providers will also make mistakes or have some zero days sometimes again us introduced, sometimes introduced by our adversaries. But what I'm trying to say by creating this cryptographic layer of isolation between us and our tenants, and amongst those tenants, they're really providing meaningful security to our customers and eliminate some of the worries that they have running on multi-tenant spaces or even collaborating together this very sensitive data, knowing that this particular protection is available to them. >> Okay, thank you, appreciate that. And I, you know, I think malicious code is often a threat model missed in these narratives. You know, operator access, yeah, could maybe I trust my Clouds provider, but if I can fence off your access even better I'll sleep better at night. Separating a code from the data, everybody's arm Intel, AM, Invidia, others, they're all doing it. I wonder if Nell, if we could stay with you and bring up the slide on the architecture. What's architecturally different with confidential computing versus how operating systems and VMs have worked traditionally? We're showing a slide here with some VMs, maybe you could take us through that. >> Absolutely, and Dave, the whole idea for Google and industry way of dealing with confidential computing is to ensure as it's three main property is actually preserved. Customers don't need to change the code. They can operate in those VMs exactly as they would with normal non-confidential VMs. But to give them this opportunity of lift and shift or no changing their apps and performing and having very, very, very low latency and scale as any Cloud can, something that Google actually pioneered in confidential computing. I think we need to open and explain how this magic was actually done. And as I said, it's again the whole entire system have to change to be able to provide this magic. And I would start with we have this concept of root of trust and root of trust where we will ensure that this machine, the whole entire post has integrity guarantee, means nobody changing my code on the most low level of system. And we introduce this in 2017 code Titan. Those our specific ASIC specific, again inch by inch system on every single motherboard that we have, that ensures that your low level former, your actually system code, your kernel, the most powerful system, is actually proper configured and not changed, not tempered. We do it for everybody, confidential computing concluded. But for confidential computing what we have to change we bring in a MD again, future silicon vendors, and we have to trust their former, their way to deal with our confidential environments. And that's why we have obligation to validate integrity not only our software and our firmware but also firmware and software of our vendors, silicon vendors. So we actually, when we booting this machine as you can see, we validate that integrity of all of this system is in place. It means nobody touching, nobody changing, nobody modifying it. But then we have this concept of the secure processor. It's special Asics best, specific things that generate a key for every single VM that our customers will run or every single node in Kubernetes, or every single worker thread in our Spark capability. We offer all of that, and those keys are not available to us. It's the best keys ever in encryption space. Because when we are talking about encryption the first question that I'm receiving all the time, where's the key, who will have access to the key? Because if you have access to the key then it doesn't matter if you encrypt it enough. But the case in confidential computing quite so revolutionary technology, ask Cloud providers who don't have access to the keys. They're sitting in the hardware and they fed to memory controller. And it means when Hypervisors that also know about these wonderful things, saying I need to get access to the memories that this particular VM I'm trying to get access to. They do not encrypt the data, they don't have access to the key. Because those keys are random, ephemeral and VM, but the most importantly in hardware not exportable. And it means now you will be able to have this very interesting role that customers all Cloud providers, will not be able to get access to your memory. And what we do, again, as you can see our customers don't need to change their applications. Their VMs are running exactly as it should run. And what you're running in VM you actually see your memory in clear, it's not encrypted. But God forbid is trying somebody to do it outside of my confidential box. No, no, no, no, no, you will not be able to do it. Now you'll see cybernet. And it's exactly what combination of these multiple hardware pieces and software pieces have to do. So OS is also modified, and OS is modified such way to provide integrity. It means even OS that you're running in UVM bucks is not modifiable and you as customer can verify. But the most interesting thing I guess how to ensure the super performance of this environment because you can imagine, Dave, that's increasing it's additional performance, additional time, additional latency. So we're able to mitigate all of that by providing incredibly interesting capability in the OS itself. So our customers will get no changes needed, fantastic performance, and scales as they would expect from Cloud providers like Google. >> Okay, thank you. Excellent, appreciate that explanation. So you know again, the narrative on this is, well you know you've already given me guarantees as a Cloud provider that you don't have access to my data but this gives another level of assurance. Key management as they say is key. Now you're not, humans aren't managing the keys the machines are managing them. So Patricia, my question to you is in addition to, you know, let's go pre-confidential computing days what are the sort of new guarantees that these hardware-based technologies are going to provide to customers? >> So if I am a customer, I am saying I now have full guarantee of confidentiality and integrity of the data and of the code. So if you look at code and data confidentiality the customer cares then they want to know whether their systems are protected from outside or unauthorized access. And that we covered with Nelly that it is. Confidential computing actually ensures that the applications and data antennas remain secret, right? The code is actually looking at the data only the memory is decrypting the data with a key that is ephemeral, and per VM, and generated on demand. Then you have the second point where you have code and data integrity and now customers want to know whether their data was corrupted, tempered, with or impacted by outside actors. And what confidential computing insures is that application internals are not tampered with. So the application, the workload as we call it, that is processing the data it's also it has not been tempered and preserves integrity. I would also say that this is all verifiable. So you have attestation, and this attestation actually generates a log trail and the log trail guarantees that provides a proof that it was preserved. And I think that the offers also a guarantee of what we call ceiling, this idea that the secrets have been preserved and not tempered with. Confidentiality and integrity of code and data. >> Got it, okay, thank you. You know, Nelly, you mentioned, I think I heard you say that the applications, it's transparent,you don't have to change the application it just comes for free essentially. And I'm, we showed some various parts of the stack before. I'm curious as to what's affected but really more importantly what is specifically Google's value add? You know, how do partners, you know, participate in this? The ecosystem or maybe said another way how does Google ensure the compatibility of confidential computing with existing systems and applications? >> And a fantastic question by the way. And it's very difficult and definitely complicated world because to be able to provide these guarantees actually a lot of works was done by community. Google is very much operate and open. So again, our operating system we working in this operating system repository OS vendors to ensure that all capabilities that we need is part of their kernels, are part of their releases, and it's available for customers to understand and even explore if they have fun to explore a lot of code. We have also modified together with our silicon vendors, kernel, host kernel, to support this capability and it means working this community to ensure that all of those patches are there. We also worked with every single silicon vendor as you've seen, and that's what I probably feel that Google contributed quite a bit in this role. We moved our industry, our community, our vendors to understand the value of easy to use confidential computing or removing barriers. And now I don't know if you noticed Intel is pulling the lead and also announcing the trusted domain extension very similar architecture and no surprise, it's again a lot of work done with our partners to again, convince, work with them, and make this capability available. The same with ARM this year, actually last year, ARM unknowns are future design for confidential computing. It's called confidential computing architecture. And it's also influenced very heavily with similar ideas by Google and industry overall. So it's a lot of work in confidential computing consortiums that we are doing. For example, simply to mention to ensure interop, as you mentioned, between different confidential environments of Cloud providers. We want to ensure that they can attest to each other. Because when you're communicating with different environments, you need to trust them. And if it's running on different Cloud providers you need to ensure that you can trust your receiver when you are sharing your sensitive data workloads or secret with them. So we coming as a community and we have this at the station, the community based systems that we want to build and influence and work with ARM and every other Cloud providers to ensure that they can interrupt. And it means it doesn't matter where confidential workloads will be hosted but they can exchange the data in secure, verifiable, and controlled by customers way. And to do it, we need to continue what we are doing. Working open again and contribute with our ideas and ideas of our partners to this role to become what we see confidential computing has to become, it has to become utility. It doesn't need to be so special but it's what what we've wanted to become. >> Let's talk about, thank you for that explanation. Let talk about data sovereignty, because when you think about data sharing you think about data sharing across, you know, the ecosystem and different regions and then of course data sovereignty comes up. Typically public policy lags, you know, the technology industry and sometimes is problematic. I know, you know, there's a lot of discussions about exceptions, but Patricia, we have a graphic on data sovereignty. I'm interested in how confidential computing ensures that data sovereignty and privacy edicts are adhered to even if they're out of alignment maybe with the pace of technology. One of the frequent examples is when you you know, when you delete data, can you actually prove the data is deleted with a hundred percent certainty? You got to prove that and a lot of other issues. So looking at this slide, maybe you could take us through your thinking on data sovereignty. >> Perfect, so for us, data sovereignty is only one of the three pillars of digital sovereignty. And I don't want to give the impression that confidential computing addresses at all. That's why we want to step back and say, hey, digital sovereignty includes data sovereignty where we are giving you full control and ownership of the location, encryption, and access to your data. Operational sovereignty where the goal is to give our Google Cloud customers full visibility and control over the provider operations, right? So if there are any updates on hardware, software, stack, any operations, that is full transparency, full visibility. And then the third pillar is around software sovereignty where the customer wants to ensure that they can run their workloads without dependency on the provider's software. So they have sometimes is often referred as survivability that you can actually survive if you are untethered to the Cloud and that you can use open source. Now let's take a deep dive on data sovereignty, which by the way is one of my favorite topics. And we typically focus on saying, hey, we need to care about data residency. We care where the data resides because where the data is at rest or in processing it typically abides to the jurisdiction, the regulations of the jurisdiction where the data resides. And others say, hey, let's focus on data protection. We want to ensure the confidentiality and integrity and availability of the data which confidential computing is at the heart of that data protection. But it is yet another element that people typically don't talk about when talking about data sovereignty, which is the element of user control. And here Dave, is about what happens to the data when I give you access to my data. And this reminds me of security two decades ago, even a decade ago, where we started the security movement by putting firewall protections and login accesses. But once you were in, you were able to do everything you wanted with the data, an insider had access to all the infrastructure, the data, and the code. And that's similar because with data sovereignty we care about whether it resides, who is operating on the data. But the moment that the data is being processed, I need to trust that the processing of the data will abide by user control, by the policies that I put in place of how my data is going to be used. And if you look at a lot of the regulation today and a lot of the initiatives around the International Data Space Association, IDSA, and Gaia X, there is a movement of saying the two parties, the provider of the data and the receiver of the data going to agree on a contract that describes what my data can be used for. The challenge is to ensure that once the data crosses boundaries, that the data will be used for the purposes that it was intended and specified in the contract. And if you actually bring together, and this is the exciting part, confidential computing together with policy enforcement. Now the policy enforcement can guarantee that the data is only processed within the confines of a confidential computing environment. That the workload is cryptographically verified that there is the workload that was meant to process the data and that the data will be only used when abiding to the confidentiality and integrity, safety of the confidential computing environment. And that's why we believe confidential computing is one, necessary and essential technology that will allow us to ensure data sovereignty especially when it comes to user control. >> Thank you for that. I mean it was a deep dive, I mean brief, but really detailed, so I appreciate that, especially the verification of the enforcement. Last question, I met you two because as part of my year end prediction post you guys sent in some predictions, and I wasn't able to get to them in the predictions post. So I'm thrilled that you were able to make the time to come on the program. How widespread do you think the adoption of confidential computing will be in '23 and what's the maturity curve look like, you know, this decade in, in your opinion? Maybe each of you could give us a brief answer. >> So my prediction in five, seven years as I started, it'll become utility. It'll become TLS. As of, again, 10 years ago we couldn't believe that websites will have certificates and we will support encrypted traffic. Now we do, and it's become ubiquity. It's exactly where our confidential computing is heading and heading, I don't know if we are there yet yet. It'll take a few years of maturity for us, but we'll do that. >> Thank you, and Patricia, what's your prediction? >> I would double that and say, hey, in the future, in the very near future you will not be able to afford not having it. I believe as digital sovereignty becomes ever more top of mind with sovereign states and also for multinational organizations and for organizations that want to collaborate with each other, confidential computing will become the norm. It'll become the default, If I say mode of operation, I like to compare that, today is inconceivable if we talk to the young technologists. It's inconceivable to think that at some point in history and I happen to be alive that we had data at address that was not encrypted. Data in transit, that was not encrypted. And I think that we will be inconceivable at some point in the near future that to have unencrypted data while we use. >> You know, and plus, I think the beauty of the this industry is because there's so much competition this essentially comes for free. I want to thank you both for spending some time on Breaking Analysis. There's so much more we could cover. I hope you'll come back to share the progress that you're making in this area and we can double click on some of these topics. Really appreciate your time. >> Anytime. >> Thank you so much.

Published Date : Feb 10 2023

SUMMARY :

Patricia, great to have you. and then Patricia you can weigh in. In additional areas that I contribute to Got it, okay. of the CTO, OCTO for Excellent, thank you in the data to Cloud into the architecture a bit and privacy of the of the data. but I'm going to push you a is available to them. we could stay with you and they fed to memory controller. So Patricia, my question to you is and integrity of the data and of the code. that the applications, and ideas of our partners to this role is when you you know, and that the data will be only used of the enforcement. and we will support encrypted traffic. and I happen to be alive and we can double click

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
NellyPERSON

0.99+

PatriciaPERSON

0.99+

International Data Space AssociationORGANIZATION

0.99+

DavePERSON

0.99+

GoogleORGANIZATION

0.99+

IDSAORGANIZATION

0.99+

last yearDATE

0.99+

2017DATE

0.99+

two partiesQUANTITY

0.99+

oneQUANTITY

0.99+

twoQUANTITY

0.99+

second pointQUANTITY

0.99+

FirstQUANTITY

0.99+

ARMORGANIZATION

0.99+

first questionQUANTITY

0.99+

fiveQUANTITY

0.99+

bothQUANTITY

0.99+

IntelORGANIZATION

0.99+

two decades agoDATE

0.99+

AsicsORGANIZATION

0.99+

secondQUANTITY

0.99+

Gaia XORGANIZATION

0.99+

OneQUANTITY

0.99+

eachQUANTITY

0.98+

seven yearsQUANTITY

0.98+

OCTOORGANIZATION

0.98+

one thoughtQUANTITY

0.98+

a decade agoDATE

0.98+

this yearDATE

0.98+

10 years agoDATE

0.98+

InvidiaORGANIZATION

0.98+

'23DATE

0.98+

todayDATE

0.98+

CloudTITLE

0.98+

three pillarsQUANTITY

0.97+

one wayQUANTITY

0.97+

hundred percentQUANTITY

0.97+

zero daysQUANTITY

0.97+

three main propertyQUANTITY

0.95+

third pillarQUANTITY

0.95+

two main goalsQUANTITY

0.95+

CTOORGANIZATION

0.93+

NellPERSON

0.9+

KubernetesTITLE

0.89+

every single VMQUANTITY

0.86+

NellyORGANIZATION

0.83+

Google CloudTITLE

0.82+

every single workerQUANTITY

0.77+

every single nodeQUANTITY

0.74+

AMORGANIZATION

0.73+

doubleQUANTITY

0.71+

single motherboardQUANTITY

0.68+

single siliconQUANTITY

0.57+

SparkTITLE

0.53+

kernelTITLE

0.53+

inchQUANTITY

0.48+

Breaking Analysis: Google's PoV on Confidential Computing


 

>> From theCUBE Studios in Palo Alto in Boston, bringing you data-driven insights from theCUBE and ETR. This is Breaking Analysis with Dave Vellante. >> Confidential computing is a technology that aims to enhance data privacy and security, by providing encrypted computation on sensitive data and isolating data, and apps that are fenced off enclave during processing. The concept of, I got to start over. I fucked that up, I'm sorry. That's not right, what I said was not right. On Dave in five, four, three. Confidential computing is a technology that aims to enhance data privacy and security by providing encrypted computation on sensitive data, isolating data from apps and a fenced off enclave during processing. The concept of confidential computing is gaining popularity, especially in the cloud computing space, where sensitive data is often stored and of course processed. However, there are some who view confidential computing as an unnecessary technology in a marketing ploy by cloud providers aimed at calming customers who are cloud phobic. Hello and welcome to this week's Wikibon Cube Insights powered by ETR. In this Breaking Analysis, we revisit the notion of confidential computing, and to do so, we'll invite two Google experts to the show. But before we get there, let's summarize briefly. There's not a ton of ETR data on the topic of confidential computing, I mean, it's a technology that's deeply embedded into silicon and computing architectures. But at the highest level, security remains the number one priority being addressed by IT decision makers in the coming year as shown here. And this data is pretty much across the board by industry, by region, by size of company. I mean we dug into it and the only slight deviation from the mean is in financial services. The second and third most cited priorities, cloud migration and analytics are noticeably closer to cybersecurity in financial services than in other sectors, likely because financial services has always been hyper security conscious, but security is still a clear number one priority in that sector. The idea behind confidential computing is to better address threat models for data in execution. Protecting data at rest and data in transit have long been a focus of security approaches, but more recently, silicon manufacturers have introduced architectures that separate data and applications from the host system, ARM, Intel, AMD, Nvidia and other suppliers are all on board, as are the big cloud players. Now, the argument against confidential computing is that it narrowly focuses on memory encryption and it doesn't solve the biggest problems in security. Multiple system images, updates, different services and the entire code flow aren't directly addressed by memory encryption. Rather to truly attack these problems, many believe that OSs need to be re-engineered with the attacker and hacker in mind. There are so many variables and at the end of the day, critics say the emphasis on confidential computing made by cloud providers is overstated and largely hype. This tweet from security researcher Rodrigo Bronco, sums up the sentiment of many skeptics. He says, "Confidential computing is mostly a marketing campaign from memory encryption. It's not driving the industry towards the hard open problems. It is selling an illusion." Okay. Nonetheless, encrypting data in use and fencing off key components of the system isn't a bad thing, especially if it comes with the package essentially for free. There has been a lack of standardization and interoperability between different confidential computing approaches. But the confidential computing consortium was established in 2019 ostensibly to accelerate the market and influence standards. Notably, AWS is not part of the consortium, likely because the politics of the consortium were probably a conundrum for AWS because the base technology defined by the consortium is seen as limiting by AWS. This is my guess, not AWS' words. But I think joining the consortium would validate a definition which AWS isn't aligned with. And two, it's got to lead with this Annapurna acquisition. It was way ahead with ARM integration, and so it's probably doesn't feel the need to validate its competitors. Anyway, one of the premier members of the confidential computing consortium is Google, along with many high profile names, including Aem, Intel, Meta, Red Hat, Microsoft, and others. And we're pleased to welcome two experts on confidential computing from Google to unpack the topic. Nelly Porter is Head of Product for GCP Confidential Computing and Encryption and Dr. Patricia Florissi is the Technical Director for the Office of the CTO at Google Cloud. Welcome Nelly and Patricia, great to have you. >> Great to be here. >> Thank you so much for having us. >> You're very welcome. Nelly, why don't you start and then Patricia, you can weigh in. Just tell the audience a little bit about each of your roles at Google Cloud. >> So I'll start, I'm owning a lot of interesting activities in Google and again, security or infrastructure securities that I usually own. And we are talking about encryption, end-to-end encryption, and confidential computing is a part of portfolio. Additional areas that I contribute to get with my team to Google and our customers is secure software supply chain because you need to trust your software. Is it operate in your confidential environment to have end-to-end security, about if you believe that your software and your environment doing what you expect, it's my role. >> Got it. Okay, Patricia? >> Well, I am a Technical Director in the Office of the CTO, OCTO for short in Google Cloud. And we are a global team, we include former CTOs like myself and senior technologies from large corporations, institutions and a lot of success for startups as well. And we have two main goals, first, we walk side by side with some of our largest, more strategic or most strategical customers and we help them solve complex engineering technical problems. And second, we advice Google and Google Cloud Engineering, product management on emerging trends and technologies to guide the trajectory of our business. We are unique group, I think, because we have created this collaborative culture with our customers. And within OCTO I spend a lot of time collaborating with customers in the industry at large on technologies that can address privacy, security, and sovereignty of data in general. >> Excellent. Thank you for that both of you. Let's get into it. So Nelly, what is confidential computing from Google's perspective? How do you define it? >> Confidential computing is a tool and one of the tools in our toolbox. And confidential computing is a way how we would help our customers to complete this very interesting end-to-end lifecycle of the data. And when customers bring in the data to cloud and want to protect it as they ingest it to the cloud, they protect it at rest when they store data in the cloud. But what was missing for many, many years is ability for us to continue protecting data and workloads of our customers when they run them. And again, because data is not brought to cloud to have huge graveyard, we need to ensure that this data is actually indexed. Again, there is some insights driven and drawn from this data. You have to process this data and confidential computing here to help. Now we have end-to-end protection of our customer's data when they bring the workloads and data to cloud thanks to confidential computing. >> Thank you for that. Okay, we're going to get into the architecture a bit, but before we do Patricia, why do you think this topic of confidential computing is such an important technology? Can you explain? Do you think it's transformative for customers and if so, why? >> Yeah, I would maybe like to use one thought, one way, one intuition behind why confidential computing matters because at the end of the day, it reduces more and more the customer's thrush boundaries and the attack surface. That's about reducing that periphery, the boundary in which the customer needs to mind about trust and safety. And in a way is a natural progression that you're using encryption to secure and protect data in the same way that we are encrypting data in transit and at rest. Now, we are also encrypting data while in the use. And among other beneficials, I would say one of the most transformative ones is that organizations will be able to collaborate with each other and retain the confidentiality of the data. And that is across industry, even though it's highly focused on, I wouldn't say highly focused but very beneficial for highly regulated industries, it applies to all of industries. And if you look at financing for example, where bankers are trying to detect fraud and specifically double finance where a customer is actually trying to get a finance on an asset, let's say a boat or a house, and then it goes to another bank and gets another finance on that asset. Now bankers would be able to collaborate and detect fraud while preserving confidentiality and privacy of the data. >> Interesting and I want to understand that a little bit more but I got to push you a little bit on this, Nellie if I can, because there's a narrative out there that says confidential computing is a marketing ploy I talked about this up front, by cloud providers that are just trying to placate people that are scared of the cloud. And I'm presuming you don't agree with that, but I'd like you to weigh in here. The argument is confidential computing is just memory encryption, it doesn't address many other problems. It is over hyped by cloud providers. What do you say to that line of thinking? >> I absolutely disagree as you can imagine Dave, with this statement. But the most importantly is we mixing a multiple concepts I guess, and exactly as Patricia said, we need to look at the end-to-end story, not again, is a mechanism. How confidential computing trying to execute and protect customer's data and why it's so critically important. Because what confidential computing was able to do, it's in addition to isolate our tenants in multi-tenant environments the cloud offering to offer additional stronger isolation, they called it cryptographic isolation. It's why customers will have more trust to customers and to other customers, the tenants running on the same host but also us because they don't need to worry about against rats and more malicious attempts to penetrate the environment. So what confidential computing is helping us to offer our customers stronger isolation between tenants in this multi-tenant environment, but also incredibly important, stronger isolation of our customers to tenants from us. We also writing code, we also software providers, we also make mistakes or have some zero days. Sometimes again us introduce, sometimes introduced by our adversaries. But what I'm trying to say by creating this cryptographic layer of isolation between us and our tenants and among those tenants, we really providing meaningful security to our customers and eliminate some of the worries that they have running on multi-tenant spaces or even collaborating together with very sensitive data knowing that this particular protection is available to them. >> Okay, thank you. Appreciate that. And I think malicious code is often a threat model missed in these narratives. You know, operator access. Yeah, maybe I trust my cloud's provider, but if I can fence off your access even better, I'll sleep better at night separating a code from the data. Everybody's ARM, Intel, AMD, Nvidia and others, they're all doing it. I wonder if Nell, if we could stay with you and bring up the slide on the architecture. What's architecturally different with confidential computing versus how operating systems and VMs have worked traditionally? We're showing a slide here with some VMs, maybe you could take us through that. >> Absolutely, and Dave, the whole idea for Google and now industry way of dealing with confidential computing is to ensure that three main property is actually preserved. Customers don't need to change the code. They can operate in those VMs exactly as they would with normal non-confidential VMs. But to give them this opportunity of lift and shift though, no changing the apps and performing and having very, very, very low latency and scale as any cloud can, some things that Google actually pioneer in confidential computing. I think we need to open and explain how this magic was actually done, and as I said, it's again the whole entire system have to change to be able to provide this magic. And I would start with we have this concept of root of trust and root of trust where we will ensure that this machine within the whole entire host has integrity guarantee, means nobody changing my code on the most low level of system, and we introduce this in 2017 called Titan. So our specific ASIC, specific inch by inch system on every single motherboard that we have that ensures that your low level former, your actually system code, your kernel, the most powerful system is actually proper configured and not changed, not tempered. We do it for everybody, confidential computing included, but for confidential computing is what we have to change, we bring in AMD or future silicon vendors and we have to trust their former, their way to deal with our confidential environments. And that's why we have obligation to validate intelligent not only our software and our former but also former and software of our vendors, silicon vendors. So we actually, when we booting this machine as you can see, we validate that integrity of all of this system is in place. It means nobody touching, nobody changing, nobody modifying it. But then we have this concept of AMD Secure Processor, it's special ASIC best specific things that generate a key for every single VM that our customers will run or every single node in Kubernetes or every single worker thread in our Hadoop spark capability. We offer all of that and those keys are not available to us. It's the best case ever in encryption space because when we are talking about encryption, the first question that I'm receiving all the time, "Where's the key? Who will have access to the key?" because if you have access to the key then it doesn't matter if you encrypted or not. So, but the case in confidential computing why it's so revolutionary technology, us cloud providers who don't have access to the keys, they're sitting in the hardware and they fed to memory controller. And it means when hypervisors that also know about this wonderful things saying I need to get access to the memories, that this particular VM I'm trying to get access to. They do not decrypt the data, they don't have access to the key because those keys are random, ephemeral and per VM, but most importantly in hardware not exportable. And it means now you will be able to have this very interesting world that customers or cloud providers will not be able to get access to your memory. And what we do, again as you can see, our customers don't need to change their applications. Their VMs are running exactly as it should run. And what you've running in VM, you actually see your memory clear, it's not encrypted. But God forbid is trying somebody to do it outside of my confidential box, no, no, no, no, no, you will now be able to do it. Now, you'll see cyber test and it's exactly what combination of these multiple hardware pieces and software pieces have to do. So OS is also modified and OS is modified such way to provide integrity. It means even OS that you're running in your VM box is not modifiable and you as customer can verify. But the most interesting thing I guess how to ensure the super performance of this environment because you can imagine Dave, that's increasing and it's additional performance, additional time, additional latency. So we're able to mitigate all of that by providing incredibly interesting capability in the OS itself. So our customers will get no changes needed, fantastic performance and scales as they would expect from cloud providers like Google. >> Okay, thank you. Excellent, appreciate that explanation. So you know again, the narrative on this is, well, you've already given me guarantees as a cloud provider that you don't have access to my data, but this gives another level of assurance, key management as they say is key. Now humans aren't managing the keys, the machines are managing them. So Patricia, my question to you is in addition to, let's go pre-confidential computing days, what are the sort of new guarantees that these hardware based technologies are going to provide to customers? >> So if I am a customer, I am saying I now have full guarantee of confidentiality and integrity of the data and of the code. So if you look at code and data confidentiality, the customer cares and they want to know whether their systems are protected from outside or unauthorized access, and that we covered with Nelly that it is. Confidential computing actually ensures that the applications and data antennas remain secret. The code is actually looking at the data, only the memory is decrypting the data with a key that is ephemeral, and per VM, and generated on demand. Then you have the second point where you have code and data integrity and now customers want to know whether their data was corrupted, tempered with or impacted by outside actors. And what confidential computing ensures is that application internals are not tempered with. So the application, the workload as we call it, that is processing the data is also has not been tempered and preserves integrity. I would also say that this is all verifiable, so you have attestation and this attestation actually generates a log trail and the log trail guarantees that provides a proof that it was preserved. And I think that the offers also a guarantee of what we call sealing, this idea that the secrets have been preserved and not tempered with, confidentiality and integrity of code and data. >> Got it. Okay, thank you. Nelly, you mentioned, I think I heard you say that the applications is transparent, you don't have to change the application, it just comes for free essentially. And we showed some various parts of the stack before, I'm curious as to what's affected, but really more importantly, what is specifically Google's value add? How do partners participate in this, the ecosystem or maybe said another way, how does Google ensure the compatibility of confidential computing with existing systems and applications? >> And a fantastic question by the way, and it's very difficult and definitely complicated world because to be able to provide these guarantees, actually a lot of work was done by community. Google is very much operate and open. So again our operating system, we working this operating system repository OS is OS vendors to ensure that all capabilities that we need is part of the kernels are part of the releases and it's available for customers to understand and even explore if they have fun to explore a lot of code. We have also modified together with our silicon vendors kernel, host kernel to support this capability and it means working this community to ensure that all of those pages are there. We also worked with every single silicon vendor as you've seen, and it's what I probably feel that Google contributed quite a bit in this world. We moved our industry, our community, our vendors to understand the value of easy to use confidential computing or removing barriers. And now I don't know if you noticed Intel is following the lead and also announcing a trusted domain extension, very similar architecture and no surprise, it's a lot of work done with our partners to convince work with them and make this capability available. The same with ARM this year, actually last year, ARM announced future design for confidential computing, it's called confidential computing architecture. And it's also influenced very heavily with similar ideas by Google and industry overall. So it's a lot of work in confidential computing consortiums that we are doing, for example, simply to mention, to ensure interop as you mentioned, between different confidential environments of cloud providers. They want to ensure that they can attest to each other because when you're communicating with different environments, you need to trust them. And if it's running on different cloud providers, you need to ensure that you can trust your receiver when you sharing your sensitive data workloads or secret with them. So we coming as a community and we have this at Station Sig, the community-based systems that we want to build, and influence, and work with ARM and every other cloud providers to ensure that they can interop. And it means it doesn't matter where confidential workloads will be hosted, but they can exchange the data in secure, verifiable and controlled by customers really. And to do it, we need to continue what we are doing, working open and contribute with our ideas and ideas of our partners to this role to become what we see confidential computing has to become, it has to become utility. It doesn't need to be so special, but it's what what we've wanted to become. >> Let's talk about, thank you for that explanation. Let's talk about data sovereignty because when you think about data sharing, you think about data sharing across the ecosystem in different regions and then of course data sovereignty comes up, typically public policy, lags, the technology industry and sometimes it's problematic. I know there's a lot of discussions about exceptions but Patricia, we have a graphic on data sovereignty. I'm interested in how confidential computing ensures that data sovereignty and privacy edicts are adhered to, even if they're out of alignment maybe with the pace of technology. One of the frequent examples is when you delete data, can you actually prove the data is deleted with a hundred percent certainty, you got to prove that and a lot of other issues. So looking at this slide, maybe you could take us through your thinking on data sovereignty. >> Perfect. So for us, data sovereignty is only one of the three pillars of digital sovereignty. And I don't want to give the impression that confidential computing addresses it at all, that's why we want to step back and say, hey, digital sovereignty includes data sovereignty where we are giving you full control and ownership of the location, encryption and access to your data. Operational sovereignty where the goal is to give our Google Cloud customers full visibility and control over the provider operations, right? So if there are any updates on hardware, software stack, any operations, there is full transparency, full visibility. And then the third pillar is around software sovereignty, where the customer wants to ensure that they can run their workloads without dependency on the provider's software. So they have sometimes is often referred as survivability that you can actually survive if you are untethered to the cloud and that you can use open source. Now, let's take a deep dive on data sovereignty, which by the way is one of my favorite topics. And we typically focus on saying, hey, we need to care about data residency. We care where the data resides because where the data is at rest or in processing need to typically abides to the jurisdiction, the regulations of the jurisdiction where the data resides. And others say, hey, let's focus on data protection, we want to ensure the confidentiality, and integrity, and availability of the data, which confidential computing is at the heart of that data protection. But it is yet another element that people typically don't talk about when talking about data sovereignty, which is the element of user control. And here Dave, is about what happens to the data when I give you access to my data, and this reminds me of security two decades ago, even a decade ago, where we started the security movement by putting firewall protections and logging accesses. But once you were in, you were able to do everything you wanted with the data. An insider had access to all the infrastructure, the data, and the code. And that's similar because with data sovereignty, we care about whether it resides, who is operating on the data, but the moment that the data is being processed, I need to trust that the processing of the data we abide by user's control, by the policies that I put in place of how my data is going to be used. And if you look at a lot of the regulation today and a lot of the initiatives around the International Data Space Association, IDSA and Gaia-X, there is a movement of saying the two parties, the provider of the data and the receiver of the data going to agree on a contract that describes what my data can be used for. The challenge is to ensure that once the data crosses boundaries, that the data will be used for the purposes that it was intended and specified in the contract. And if you actually bring together, and this is the exciting part, confidential computing together with policy enforcement. Now, the policy enforcement can guarantee that the data is only processed within the confines of a confidential computing environment, that the workload is in cryptographically verified that there is the workload that was meant to process the data and that the data will be only used when abiding to the confidentiality and integrity safety of the confidential computing environment. And that's why we believe confidential computing is one necessary and essential technology that will allow us to ensure data sovereignty, especially when it comes to user's control. >> Thank you for that. I mean it was a deep dive, I mean brief, but really detailed. So I appreciate that, especially the verification of the enforcement. Last question, I met you two because as part of my year-end prediction post, you guys sent in some predictions and I wasn't able to get to them in the predictions post, so I'm thrilled that you were able to make the time to come on the program. How widespread do you think the adoption of confidential computing will be in '23 and what's the maturity curve look like this decade in your opinion? Maybe each of you could give us a brief answer. >> So my prediction in five, seven years as I started, it will become utility, it will become TLS. As of freakin' 10 years ago, we couldn't believe that websites will have certificates and we will support encrypted traffic. Now we do, and it's become ubiquity. It's exactly where our confidential computing is heeding and heading, I don't know we deserve yet. It'll take a few years of maturity for us, but we'll do that. >> Thank you. And Patricia, what's your prediction? >> I would double that and say, hey, in the very near future, you will not be able to afford not having it. I believe as digital sovereignty becomes ever more top of mind with sovereign states and also for multinational organizations, and for organizations that want to collaborate with each other, confidential computing will become the norm, it will become the default, if I say mode of operation. I like to compare that today is inconceivable if we talk to the young technologists, it's inconceivable to think that at some point in history and I happen to be alive, that we had data at rest that was non-encrypted, data in transit that was not encrypted. And I think that we'll be inconceivable at some point in the near future that to have unencrypted data while we use. >> You know, and plus I think the beauty of the this industry is because there's so much competition, this essentially comes for free. I want to thank you both for spending some time on Breaking Analysis, there's so much more we could cover. I hope you'll come back to share the progress that you're making in this area and we can double click on some of these topics. Really appreciate your time. >> Anytime. >> Thank you so much, yeah. >> In summary, while confidential computing is being touted by the cloud players as a promising technology for enhancing data privacy and security, there are also those as we said, who remain skeptical. The truth probably lies somewhere in between and it will depend on the specific implementation and the use case as to how effective confidential computing will be. Look as with any new tech, it's important to carefully evaluate the potential benefits, the drawbacks, and make informed decisions based on the specific requirements in the situation and the constraints of each individual customer. But the bottom line is silicon manufacturers are working with cloud providers and other system companies to include confidential computing into their architectures. Competition in our view will moderate price hikes and at the end of the day, this is under-the-covers technology that essentially will come for free, so we'll take it. I want to thank our guests today, Nelly and Patricia from Google. And thanks to Alex Myerson who's on production and manages the podcast. Ken Schiffman as well out of our Boston studio. Kristin Martin and Cheryl Knight help get the word out on social media and in our newsletters, and Rob Hoof is our editor-in-chief over at siliconangle.com, does some great editing for us. Thank you all. Remember all these episodes are available as podcasts. Wherever you listen, just search Breaking Analysis podcast. I publish each week on wikibon.com and siliconangle.com where you can get all the news. If you want to get in touch, you can email me at david.vellante@siliconangle.com or DM me at D Vellante, and you can also comment on my LinkedIn post. Definitely you want to check out etr.ai for the best survey data in the enterprise tech business. I know we didn't hit on a lot today, but there's some amazing data and it's always being updated, so check that out. This is Dave Vellante for theCUBE Insights powered by ETR. Thanks for watching and we'll see you next time on Breaking Analysis. (subtle music)

Published Date : Feb 10 2023

SUMMARY :

bringing you data-driven and at the end of the day, and then Patricia, you can weigh in. contribute to get with my team Okay, Patricia? Director in the Office of the CTO, for that both of you. in the data to cloud into the architecture a bit, and privacy of the data. that are scared of the cloud. and eliminate some of the we could stay with you and they fed to memory controller. to you is in addition to, and integrity of the data and of the code. that the applications is transparent, and ideas of our partners to this role One of the frequent examples and a lot of the initiatives of the enforcement. and we will support encrypted traffic. And Patricia, and I happen to be alive, the beauty of the this industry and at the end of the day,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
NellyPERSON

0.99+

PatriciaPERSON

0.99+

Alex MyersonPERSON

0.99+

AWSORGANIZATION

0.99+

International Data Space AssociationORGANIZATION

0.99+

DavePERSON

0.99+

AWS'ORGANIZATION

0.99+

MicrosoftORGANIZATION

0.99+

Dave VellantePERSON

0.99+

Rob HoofPERSON

0.99+

Cheryl KnightPERSON

0.99+

Nelly PorterPERSON

0.99+

GoogleORGANIZATION

0.99+

NvidiaORGANIZATION

0.99+

IDSAORGANIZATION

0.99+

Rodrigo BroncoPERSON

0.99+

2019DATE

0.99+

Ken SchiffmanPERSON

0.99+

IntelORGANIZATION

0.99+

AMDORGANIZATION

0.99+

2017DATE

0.99+

ARMORGANIZATION

0.99+

AemORGANIZATION

0.99+

NelliePERSON

0.99+

Kristin MartinPERSON

0.99+

Red HatORGANIZATION

0.99+

two partiesQUANTITY

0.99+

Palo AltoLOCATION

0.99+

last yearDATE

0.99+

Patricia FlorissiPERSON

0.99+

oneQUANTITY

0.99+

MetaORGANIZATION

0.99+

twoQUANTITY

0.99+

thirdQUANTITY

0.99+

Gaia-XORGANIZATION

0.99+

second pointQUANTITY

0.99+

two expertsQUANTITY

0.99+

david.vellante@siliconangle.comOTHER

0.99+

secondQUANTITY

0.99+

bothQUANTITY

0.99+

first questionQUANTITY

0.99+

fiveQUANTITY

0.99+

OneQUANTITY

0.99+

theCUBE StudiosORGANIZATION

0.99+

two decades agoDATE

0.99+

'23DATE

0.99+

eachQUANTITY

0.99+

a decade agoDATE

0.99+

threeQUANTITY

0.99+

zero daysQUANTITY

0.98+

fourQUANTITY

0.98+

OCTOORGANIZATION

0.98+

todayDATE

0.98+

Ken Durazzo, Dell Technologies and Matt Keesan, IonQ | Super Computing 2022


 

>>How do y'all and welcome back to the cube where we're live from Dallas at a Supercomputing 2022. My name is Savannah Peterson. Joined with L AED today, as well as some very exciting guests talking about one of my favorite and most complex topics out there, talking about quantum a bit today. Please welcome Ken and Matthew. Thank you so much for reading here. Matthew. Everyone's gonna be able to see your shirt. What's going on with hybrid quantum? I have >>To ask. Wait, what is hybrid quantum? Yeah, let's not pretend that. >>Let's not >>Pretend that everybody knows, Everyone already knows what quantum computing is if we goes straight to highway. Yeah. Okay. So with the brief tour detour took qu regular quantum computing. Yeah, >>No, no. Yeah. Let's start with quantum start before. >>So you know, like regular computers made of transistors gives us ones and zeros, right? Binary, like you were talking about just like half of the Cheerios, right? The joke, it turns out there's some problems that even if we could build a computer as big as the whole universe, which would be pretty expensive, >>That might not be a bad thing, but >>Yeah. Yeah. Good for Dell Got mill. >>Yeah. >>Yeah. We wouldn't be able to solve them cuz they scale exponentially. And it turns out some of those problems have efficient solutions in quantum computing where we take any two state quantum system, which I'll explain in a sec and turn it into what we call a quantum bit or qubit. And those qubits can actually solve some problems that are just infeasible on even these world's largest computers by offering exponential advantage. And it turns out that today's quantum computers are a little too small and a little too noisy to do that alone. So by pairing a quantum computer with a classical computer, hence the partnership between IQ and Dell, you allow each kind of compute to do what it's best at and thereby get answers you can't get with either one alone. >>Okay. So the concept of introducing hybridity, I love that word bridge. I dunno if I made it up, but it's it for it. Let's about it. Abri, ding ding. So does this include simulating the quantum world within the, what was the opposite? The classical quantum world? Classical. Classical, classical computer. Yeah. So does it include the concept of simulating quantum in classical compute? >>Absolutely. >>Okay. How, how, how do, how do you do that? >>So there's simulators and emulators that effectively are programmed in exactly the same way that a physical quantum machine is through circuits translated into chasm or quantum assembly language. And those are the exact same ways that you would program either a physical q p or a simulated >>Q p. So, so access to quantum computing today is scarce, right? I mean it's, it's, it's, it's limited. So having the ability to have the world at large or a greater segment of society be able to access this through simulation is probably a good idea. >>Fair. It's absolutely a wonderful one. And so I often talk to customers and I tell them about the journey, which is hands on keyboard, learning, experimentation, building proof of concepts, and then finally productization. And you could do much of that first two steps anyway very robustly with simulation. >>It's much like classical computing where if you imagine back in the fifties, if, if the cube was at some conference in 1955, you know, we wouldn't have possibly been able to predict what we'd be doing with computing 70 years later, right? Yeah. That teenagers be making apps on their phones that changed the world, right? And so by democratizing access this way, suddenly we can open up all sorts of new use cases. We sort of like to joke, there's only a couple hundred people in the world who really know how to program quantum computers today. And so how are we gonna make thousands, tens of thousands, millions of quantum programmers? The answer is access and simulators are an amazingly accessible way for everyone to start playing around with the >>Fields. Very powerful tool. >>Wow. Yeah. I'm just thinking about how many, there's, are there really only hundreds of people who can program quantum computing? >>I kind of generally throw it out there and I say, you know, if you looked at a matrix of a thousand operations with hundreds of qubits, there's probably, I don't know, 2000 people worldwide that could program that type of a circuit. I mean it's a fairly complex circuit at that point and >>I, I mean it's pretty phenomenal When you think about how early we are in adoption and, and the rollout of this technology as a whole, can you see quite a bit as, as you look across your customer portfolio, what are some of the other trends you're seeing? >>Well, non quantum related trends or just any type you give us >>Both. >>Yeah. So >>We're a thought leader. This is >>Your moment. Yeah, so we do quite a bit. We see quite a bit actually. There's a lot of work happening at the edge as you're probably well aware of. And we see a lot of autonomous mobile robots. I actually lead the, the research office. So I get to see all the cool stuff that's really kind of emerging before it really regrets >>What's coming next. >>Let's see, Oh, I can't tell you what's coming next, but we see edge applications. Yes, we see a lot of, of AI applications and artificial intelligence is morphing dramatically through the number of frameworks and through the, the types and places you would place ai, even places I, I personally never thought we would go like manufacturing environments. Some places that were traditionally not very early adopters. We're seeing AI move very quickly in some of those areas. One of the areas that I'm really excited about is digital twins and the ability to eventually do, let's come up on acceleration with quantum technologies on, on things like computational fluid dynamics. And I think it's gonna be a wonderful, wonderful area for us moving forward. >>So, So I can hear the people screaming at the screen right now. Wait a minute, You said it was hybrid, you're only talking the front half. That's, that's cat. What about the back half? That's dog. What about the quantum part of it? So I, on Q and, and I apologize. Ion Q >>Ion >>Q, Yeah Ion Q cuz you never know. You never never know. Yeah. Where does the actual quantum come in? >>That's a great >>Question. So you guys have one of these things. >>Yeah, we've built, we currently have the world's best quantum computer by, by sub measures I drop there. Yeah, no big deal. Give me some snaps for that. Yeah, Ken knows how to pick em. Yeah, so right. Our, our approach, which is actually based on technology that's 50 years old, so it's quite, quite has a long history. The way we build atomic clocks is the basis for trapped eye quantum computing. And in fact the first quantum logic gate ever made in 1995 was at NIST where they modified their atomic clock experiment to do quantum gates. And that launched really the hardware experimentalist quantum Peter Revolution. And that was by Chris Monroe, our co-founder. So you know that history has flown directly into us. So to simplify, we start with an ion trap. Imagine a gold block with a bunch of electrodes that allow you to make precisely shaped electromagnetic fields, sort of like a rotating saddle. >>Then take a source of atoms. Now obviously we're all sources of atoms. We have a highly purified source of metal atium. We heat it up, we get a nice hot plume of atoms, we ionize those atoms with an ionizing later laser. Now they're hot and heavy and charged. So we can trap them in one of these fields. And now our electromagnetic field that's spitting rapidly holds the, the ions like balls in a bowl if you can imagine them. And they line up in a nice straight line and we hold them in place with these fields and with cooling laser beams. And up to now, that's how an atomic clock works. Trap an item and shine it with a laser beam. Count the oscillations, that's your clock. Now if you got 32 of those and you can manipulate their energy states, in our case we use the hyper fine energy states of the atom. >>But you can basically think of your high school chemistry where you have like an unexcited electron, an excited electron. Take your unexcited state as a zero, your excited state as a one. And it turns out with commercially available lasers, you can drive anywhere between a zero, a one or a super position of zero and one. And so that is our quantum bit, the hyper fine energy state of the atrium atom. And we just line up a bunch of them and through there access the magical powers of supervision entanglement, as we were talking about before, they don't really make sense to us here in the regular world, but >>They do exist. But what you just described is one cubit. That's right. And the way that you do it isn't exactly the same way that others who are doing quantum computing do it. That's right. Is that okay? >>And there's a lot of advantages to the trapped iron approach. So for example, you can also build a super conducting qubit where you, where you basically cool a chip to 47 mil kelvin and coerce millions of atoms to work together as a single system. The problem is that's not naturally quantum. So it's inherently noisy and it wants to deco here does not want to be a quantum bit. Whereas an atom is very happy to be by itself a qubit because we don't have to do anything to it. It's naturally quantum, if that makes sense. And so atomic qubits, like we use feature a few things. One the longest coherence times in the industry, meaning you can run very deep circuits, the most accurate operations, very low noise operations. And we don't have any wires. Our atoms are connected by laser light. That means you can connect any pair. So with some other technologies, the qubits are connected by wires. That means you can only run operations between physically connected qubits. It's like programming. If you could only use, for example, bits that are adjacent with an i untrapped approach, you can connect any pair so that all to all connectivity means your compilation is much more efficient and you can do much wider and deeper circuits. >>So what's the, what is the closest thing to a practical application that we've been able to achieve at this point? Question. And when I say practical, it doesn't have to be super practical. I mean, what is the, what is the sort of demonstration, the least esoteric demonstration of this at this point? >>To tie into what Ken was saying earlier, I think there's at least two areas that are very exciting. One is chemistry. Chemistry. So for example, you know, we have water in our cup and we understand water pretty well, but there's lots of molecules that in order to study them, we actually have to make them in a lab and do lots of experiments. And to give you a sense of the order of magnitude, if you wanted to understand the ground state of the caffeine molecule, which we all know and has 200 electrons, you would need to build a computer bigger than the moon. So, which is, you know, again, would be good profit for Dell, but probably not gonna happen time soon. That's >>Kind of fun to think about though. Yeah, that's a great analogy. That >>Was, yeah. And in fact it'd be like 10 moons of compute. Okay. So build 10 moons of >>Computer. I >>Love the sci-fi issue. Exactly. And now you can calculate caffeine, it's crazy or it just fits in a quantum computer the size of this table. And so we're using hybrid quantum computing now to start proving out these algorithms not for molecules as complex as caffeine or what we want in the future. Like biologics, you know, new cancer medications, new materials and so forth. But we are able to show, for example, the ground state of smaller molecules and prove a path to where, you know, decision maker could see in a few years from now, Oh, we'll be able to actually simulate not molecules we already understand, but molecules we've never been able to study a prayer, if that makes sense. And then, >>Yeah, I think there's a key point underneath that, and I think goes back to the question that you asked earlier about the why hybrid applications inherently run on the classical infrastructure and algorithms are accelerated through qs, the quantum processing units. >>And so are you sort of time sharing in the sense that this environment that you set up starts with classical, with simulation and then you get to a point where you say, okay, we're ready, you pick up the bat phone and you say I wanna, >>I would say it's more like a partnership, really. Yeah, >>Yeah. And I think, I think it's kind of the, the way I normally describe it is, you know, we've taken a look at it it from a really kind of a software development life cycle type of perspective where again, if you follow that learn experiment, pro proof of concept, and then finally productize, we, we can cover and allow for a developer to start prototyping and proofing on simulators and when they're ready all they do is flip a switch and a manifest and they can automatically engage a qu a real quantum physical quantum system. And so we've made it super simple and very accessible in a democratizing access for developers. >>Yeah. Makes such big difference. Go ahead. >>A good analogy is to like GPUs, right? Where it's not really like, you know, you send it away, but rather the GPU accelerates certain operations. The q p. Yeah, because quantum mechanics, it turns out the universe runs on linear algebra. So one way to think about the q p is the most efficient way of doing linear algebra that exists. So lots of problems that can be expressed in that form. Combinatorial optimization problems in general, certain kinds of machine learning, et cetera, get an exponential speed up by running a section of the algorithm on the quantum computer. But of course you wouldn't like port Microsoft Word. Yeah, exactly. You know, you're not gonna do that in your product. It would be a waste of your quantum computer. >>Not just that you wanna know exactly how much money is in your bank account, not probabilistically how much might be ballpark. Yeah. Realm 10, moon ballpark, right? >>10 moon ballpark. Be using that for the rest of the show. Yeah. Oh, I love that. Ken, tell me a little bit about how you identify companies and like I n Q and and end up working with Matthew. What, what's that like, >>What's it like or how do you >>Find it's the process? Like, so, you know, let's say I've got the the >>We're not going there though. Yeah. We're not >>Personal relationship. >>Well, >>You can answer these questions however you want, you know. No, but, but what does that look like for Dell? How do you, how do you curate and figure out who you're gonna bring into this partnership nest? >>Yeah, you know, I, I think it was a, it's, it was a, a very long drawn out learning opportunity. We started actually our working quantum back in 2016. So we've been at it for a long time. And only >>In quantum would we say six years is a long time. I love >>That. Exactly. >>By the way, that was like, we've been doing this for age for a >>Long time. Yeah. Very long time before >>You were born. Yes. >>Feels like it actually, believe it or not. But, so we've been at it for a long time and you know, we went down some very specific learning paths. We took a lot of different time to, to learn about different types of qubits available, different companies, what their approaches were, et cetera. Yeah. And, and we ended up meeting up with, with I N Q and, and we also have other partners as well, like ibm, but I N q you know, we, there is a nice symbiotic relationship. We're actually doing some really cool technologies that are even much, much further ahead than the, you know, strict classical does this, quantum does that where there's significant amount of interplay between the simulation systems and between the real physical QS. And so it's, it's turning out to be a great relationship. They're, they're very easy to work with and, and a lot of fun too, as you could probably tell. Yeah. >>Clearly. So before we wrap, I've got it. Okay. Okay. So get it. Let's get, let's get, yeah, let's get deep. Let's get deep for a second or a little deeper than we've been. So our current, our current understanding of all this, of the universe, it's pretty limited. It's down to the point where we effectively have it assigned to witchcraft. It's all dark energy and dark matter. Right. What does that mean exactly? Nobody knows. But if you're in the quantum computing space and you're living this every day, do you believe that it represents the key to us understanding things that currently we just can't understand classical models, including classical computing, our brains as they're constructed aren't capable of understanding the real real that's out there. Yeah. If you're in the quantum computing space, do you possess that level of hubris? Do you think that you are gonna deliver the answers? >>I'm just like, I think the more you're in the space, the more mysterious and amazing it all seems. There's a, but there is a great quote by Richard Feinman that sort of kicked off the quantum exploration. So he gave a lecture in 1981, so, you know, long before any of this began, truly ages ago, right? Yeah. And in this lecture he said, you know, kind of wild at that time, right? We had to build these giant supercomputers to simulate just a couple atoms interacting, right? And it's kind of crazy that you need all this compute to simulate what nature does with just a handful >>Particles. Yeah. >>Really small. So, and, and famously he said, you know, nature just isn't classical. Damn it. And so you need to build a computer that works with nature to understand nature. I think, you know, the, the quantum revolution has only just begun. There's so many new things to learn, and I'm sure the quantum computers of 40 years from now are not gonna look like the, you know, the computers of today, just as the classical computers of 40 years ago look quite different to us now, >>And we're a bunch of apes. But you think we'll get there? >>I, yeah, I, I mean, I, I have, I think we have, I feel incredibly optimistic that this tool, quantum computing as a tool represents a sea change in what's possible for humans to compute. >>Yeah. I think it's that possibility. You know, I, when I tell people right now in the quantum era, we're in the inac stage of the quantum era, and so we have a long way to go, but the potential is absolutely enormous. In fact, incomprehensibly enormous, I >>Was just gonna say, I don't even think we could grasp >>In the, from the inac is they had no idea of computers inside of your hand, right? Yeah. >>They're calculating, you know, trajectories, right? Yeah. If you told them, like, we'd all be video chatting, you >>Know, >>Like, and kids would be doing synchronized dances, you know, you'd be like, What? >>I love that. Well, well, on that note, Ken Matthew, really great to have you both, everyone now will be pondering the scale and scope of the universe with their 10 moon computer, 10 moons. That's right. And, and you've given me my, my new favorite bumper sticker since we've been on a, on a roll here, David and I, which is just naturally quantum. Yeah, that's, that's, that's, that's one of my new favorite phrases from the show. Thank you both for being here. David, thank you for hanging out and thank all of you for tuning in to our cube footage live here in Dallas. We are at Supercomputing. This is our last show for the day, but we look forward to seeing you tomorrow morning. My name's Savannah Peterson. Y'all have a lovely night.

Published Date : Nov 16 2022

SUMMARY :

Thank you so much for reading here. Yeah, let's not pretend that. So with the brief tour detour took qu regular quantum computing. hence the partnership between IQ and Dell, you allow each kind of compute to do what it's So does it include the concept of simulating quantum in you would program either a physical q p or a simulated So having the ability to have the And you could do much of that first if, if the cube was at some conference in 1955, you know, we wouldn't have possibly been Very powerful tool. I kind of generally throw it out there and I say, you know, if you looked at a matrix of a thousand operations with We're a thought leader. And we see a lot of the types and places you would place ai, even places I, What about the quantum part of it? Q, Yeah Ion Q cuz you never know. So you guys have one of these things. So you know that history has flown directly into Now if you got 32 of those and you can manipulate their And it turns out with commercially available lasers, you can drive anywhere between a zero, And the way that you do it isn't for example, bits that are adjacent with an i untrapped approach, you can connect any pair so that all And when I say practical, it doesn't have to be super practical. And to give you a sense of the order of magnitude, Kind of fun to think about though. And in fact it'd be like 10 moons of compute. I And now you can calculate caffeine, it's crazy or it just fits in a quantum computer the size of Yeah, I think there's a key point underneath that, and I think goes back to the question that you asked earlier about the why hybrid Yeah, of a software development life cycle type of perspective where again, if you follow that learn experiment, Where it's not really like, you know, Not just that you wanna know exactly how much money is in your bank account, not probabilistically how tell me a little bit about how you identify companies and like I n Q and and end Yeah. You can answer these questions however you want, you know. Yeah, you know, I, I think it was a, it's, it was a, a very long drawn out learning opportunity. In quantum would we say six years is a long time. You were born. But, so we've been at it for a long time and you know, do you believe that it represents the key to us understanding And it's kind of crazy that you need all this compute to simulate what nature does Yeah. And so you need to build a computer that works with nature to understand nature. But you think we'll get there? I, yeah, I, I mean, I, I have, I think we have, I feel incredibly optimistic that this to go, but the potential is absolutely enormous. Yeah. They're calculating, you know, trajectories, right? but we look forward to seeing you tomorrow morning.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
KenPERSON

0.99+

Chris MonroePERSON

0.99+

MatthewPERSON

0.99+

DavidPERSON

0.99+

2016DATE

0.99+

Ken DurazzoPERSON

0.99+

Savannah PetersonPERSON

0.99+

Matt KeesanPERSON

0.99+

1995DATE

0.99+

10 moonsQUANTITY

0.99+

Ken MatthewPERSON

0.99+

Richard FeinmanPERSON

0.99+

DallasLOCATION

0.99+

1981DATE

0.99+

32QUANTITY

0.99+

six yearsQUANTITY

0.99+

tomorrow morningDATE

0.99+

200 electronsQUANTITY

0.99+

1955DATE

0.99+

DellORGANIZATION

0.99+

thousandsQUANTITY

0.99+

10 moonQUANTITY

0.99+

one cubitQUANTITY

0.99+

hundreds of qubitsQUANTITY

0.99+

BothQUANTITY

0.99+

firstQUANTITY

0.99+

millions of atomsQUANTITY

0.99+

two stateQUANTITY

0.99+

zeroQUANTITY

0.99+

2000 peopleQUANTITY

0.99+

tens of thousandsQUANTITY

0.98+

bothQUANTITY

0.98+

L AEDORGANIZATION

0.98+

OneQUANTITY

0.98+

todayDATE

0.98+

IQORGANIZATION

0.98+

70 years laterDATE

0.98+

first two stepsQUANTITY

0.98+

Dell TechnologiesORGANIZATION

0.98+

zerosQUANTITY

0.97+

oneQUANTITY

0.97+

47 mil kelvinQUANTITY

0.96+

40 yearsQUANTITY

0.95+

each kindQUANTITY

0.94+

40 years agoDATE

0.93+

50 years oldQUANTITY

0.93+

SupercomputingORGANIZATION

0.92+

single systemQUANTITY

0.92+

millions of quantum programmersQUANTITY

0.91+

NISTORGANIZATION

0.9+

AbriPERSON

0.89+

2022DATE

0.87+

ages agoDATE

0.86+

hundreds of peopleQUANTITY

0.86+

couple hundred peopleQUANTITY

0.84+

thousand operationsQUANTITY

0.84+

couple atomsQUANTITY

0.77+

a secondQUANTITY

0.77+

Supercomputing 2022EVENT

0.74+

onesQUANTITY

0.72+

IonQPERSON

0.71+

millORGANIZATION

0.71+

two areasQUANTITY

0.71+

one wayQUANTITY

0.7+

WordTITLE

0.69+

fieldsQUANTITY

0.67+

frontQUANTITY

0.66+

MicrosoftORGANIZATION

0.65+

SuperEVENT

0.58+

yearsDATE

0.52+

moonLOCATION

0.5+

Mike Palmer, Sigma Computing | Snowflake Summit 2022


 

>>Welcome back to Vegas guys, Lisa Martin and Dave Lanta here wrapping up our coverage of day two of snowflake summit. We have given you a lot of content in the last couple of days. We've had a lot of great conversations with snowflake folks with their customers and with partners. And we have an alumni back with us. Please. Welcome back to the queue. Mike Palmer, CEO of Sigma computing. Mike. It's great to see you. >>Thanks for having me. And I guess again >>Exactly. >>It's fantastic me. >>So talk to the audience about Sigma before we get into the snowflake partnership and what you guys are doing from a technical perspective, give us that overview of the vision and some of the differentiators. >>Sure. You know, you've over the last 12 years, companies have benefited from enormous investments and improvements in technology in particular, starting with cloud technologies, obviously going through companies like snowflake, but in terms of the normal user, the one that makes the business decision in the marketing department and the finance team, you know, in the works in the back room of the supply chain, doing inventory very little has changed for those people. And the time had come where the data availability, the ability to organize it, the ability to secure it was all there, but the ability to access it for those people was not. And so what Sigma's all about is taking great technology, finding the skillset they have, which happens to be spreadsheets. There are billion license spreadsheet users in the world and connecting that skillset with all of the power of the cloud. >>And how do you work with snowflake? What are some of the, the what's the joint value proposition? >>How are they as an investor? That's what I wanna know. Ah, >>Quiet, which is the way we like them. No, I'm just kidding. Snowflake is, well, first of all, investment is great, but partnership is even better. Right. You know, and I think snowflake themselves are going through some evolution, but let's start with the basics of technology where this all starts because you know, all of the rest doesn't matter if the product is not great, we work directly on snowflake. And what that means is as an end user, when I, when I sit on that marketing team and I want to understand and, and connect, how did I get a, a customer where I had a pay to add? And they showed up on my website and from my website, they went to a trial. And from there, they touched a piece of syndicated contents. All of that data sits in snowflake and I, as a marketer, understand what it means to me. >>So for the first time, I want to be able to see that data in one place. And I want to understand conversion rates. I want to understand how I can impact those conversion rates. I can make predictions. What that user is doing is going to, to Sigma accessing live data in snowflake, they're able to ask ad hoc questions, questions that were never asked questions, that they don't exist in a filter that were never prepped by a data engineer. So they could truly do something creative and novel in a very independent sort of way. And the connection with Snowflake's live data, the performance, the security and governance that we inherit. These are all facilitators to really expand that access across the enterprise. So at, at a product level, we were built by a team of people, frankly, that also were the original investors in snowflake by two amazing engineers and founders, Rob will and Jason France, they understood how snowflake worked and that shows up in the product for our end customers. >>So, but if I may just to follow up on that, I mean, you could do that without snowflake, but what, it would be harder, more expensive. Describe what you'd have to go through to accomplish that outcome. >>And I think snowflake does a good job of enabling the ecosystem at large. Right. But you know, you always appreciate seeing early access to understand what the architecture's going to look like. You know, some of the things that I will, you know, leaning forward that we've heard here that we're very excited about is snowflake going to attack the TP market, right? The transactional market, one of the transactional database market. I, yeah. Right. You know, one of the things that we see coming, and, and one of the bigger things that we'll be talking about in Sigma is not just that you can do analytics out of snowflake. I think that's something that we do exceptionally well on an ad hoc basis, but we're gonna be the first that allow you to write into snowflake and to do that with good performance. And to do that reliably, we go away from OAP, which is the terminology for data warehousing. >>And we go toward transactional databases. And in that world, understanding snowflake and working collaboratively with them creates again, a much better experience for the end customer. So they, they allow us into those programs, even coming to these conferences, we talk to folks that run the industry teams, trying to up level that message and not just talk database and, and analytics, but talk about inventory management. How do we cut down the gap that exists between POS systems and inventory ordering, right? So that we get fewer stockouts, but also that we don't overorder. So that's another benefit, >>Strong business use cases. >>That's correct. >>And you're enabling those business users to have access to that data. I presume in near real time or near real time, so that they can make decisions that drive marketing forward or finance forward or legal >>Forward. Exactly. We had a customer panel yesterday. An example of that go puff is hopefully most of the viewers are familiar with, as a delivery company. This is a complicated business to run. It's run on the fringes. When we think about how to make money at it, which means that the decisions need to be accurate. They need to be real time. You can't have a batch upload for delivery when they're people are on the street, and then there's an issue. They need to understand the exact order at that time, not in 10 minutes, not from five minutes ago, right. Then they need to understand, do I have inventory in the warehouse when the order comes in? If they don't, what's a replacement product. We had a Mike came in from go puff and walked us through all of the complexity of that and how they're using Sigma to really just shorten those decision cycles and make them more accurate. You know, that's where the business actually benefits and, >>And actually create a viable business model. Cuz you think back to the early, think back to the.com days and you had pets.com, right? They couldn't make any money. Yeah. Without chewy. Okay. They appears to be a viable business model. Right? Part of that is just the efficiencies. And it's sort of a, I dunno if those are customers that they may or may not be, but they should be if they're not >>Chewy is, but okay. You know, and that's another example, but I'll even pivot to the various REI and other retailers. What do they care about cohorts? I'm trying to understand who's buying my product. What can I sell to them next? That, that idea of again, I'm sitting in a department, that's not data engineering, that's not BI now working collaboratively where they can get addend engineer, putting data sets together. They have a BI person that can help in the analytics process. But now it's in a spreadsheet where I understand it as a marketer. So I can think about new hierarchies. I wanna know it by customer, by region, by product type. I wanna see it by all of those things. I want to be able to do that on the fly because then it creates new questions that sort of flow. If you' ever worked in development, we use the word flow constantly, right? And as people that flow is when we have a question, we get an answer that generates a question. We have, we just keep doing that iteratively. That that is where Sigma really shines for them. >>What does a company have to do to really take advantage of, of this? I, if they're kind of starting from a company that's somewhat immature, what are the sort of expectations, maybe even outta scope expectations so they can move faster, accelerate analytics, a lot of the themes that we've heard today, >>What does an immature company is actually even a question in, in and of itself? You know, I think a lot of companies consider themselves to be immature simply because for various constraint reasons, they haven't leveraged the data in the way that they thought possible. Good, >>Good, good definition. Okay. So not, not, >>Not, I use this definition for digital transformation. It very simple. It is. Do you make better decisions, faster McKenzie calls this corporate metabolism, right? Can you speed up the metabolism of, of an enterprise and for me and for the Sigma customer base, there's really not much you have to do once. You've adopted snowflake because for the first time the barriers and the silos that existed in terms of accessing data are gone. So I think the biggest barrier that customers have is curiosity. Because once you have curiosity and you have access, you can start building artifacts and assets and asking questions. Our customers are up and running in the product in hours. And I mean that literally in hours, we are a user in snowflake, that's a direct live connection. They are able to explore tables, raw. They can do joins themselves if they want to. They can obviously work with their data engineering team to, to create data sets. If that's the preferred method. And once they're there and they've ever built a pivot table, they can be working in Sigma. So our customers are getting insights in the first one to two days, you referenced some, those of us are old enough to remember pest.com. Also old enough to remember shelfware that we would buy. We are very good at showing customers that within hours they're getting value from their investment in Sigma. And that, that just creates momentum, right? Oh, >>Tremendous momentum and >>Trust and trust and expansion opportunities for Sigma. Because when you're in one of those departments, someone else says, well, you know, why do you get access to that data? But I don't, how are you doing this? Yeah. So we're, you know, I think that there's a big movement here. People, I often compare data to communication. If you go back a hundred years, our communication was not limited. As it turns out by our desire to communicate, it was limited by the infrastructure. We had the typewriter, a letter and the us postal service and a telephone that was wired. And now we have walk around here. We, everything is, is enabled for us. And we send, you know, hundreds and thousands of messages a day and probably could do more. You will find that is true. And we're seeing it in our product is true of data. If you give people access, they have 10 times as many questions as they thought they had. And that's the change that we're gonna see in business over the next few years, >>Frank Salman's first book, what he was was CEO of snowflake was rise of the data cloud. And he talked about network effects. Basically what he described was Metcalf's law. Again, go back to the.com days, right? And he, Bob Metcalf used the phone system. You know, if there's two people in the phone system, it's not that valuable, right. >>You know, exactly, >>You know, grow it. And that's where the value is. And that's what we're seeing now applied to data. >>And even more than that, I think that's a great analogy. In fact, the direct comparison to what Sigma is doing actually goes one step beyond everything that I've been talking about, which is great at the individual level, but now the finance team and the marketing team can collaborate in the platform. They can see data lineage. In fact, one of our, our big emphasis points here is to eliminate the sweet products. You know, the ones where, you know, you think you're buying something, but you really have a spreadsheet product here and a document product there and a slide product over there. And they, you know, you can do all of that in Sigma. You can write a narrative. You can real time live, edit on numbers. You, you know, if you want to, you could put a picture in it. But you know, at Sigma we present everything out of our product. Every meeting is live data. Every question is answered on the spot. And that's when, you know, you know, to your point about met cap's law. Now everybody's involved in the decision making. They're doing it real time. Your meetings are more productive. You have fewer of them because they're no action items, right. We're answering our questions there and we're, and we're moving forward. >>You know, view were meeting sounds good. Productivity is, is weird now with the, the pandemic. But you know, if you go back to the nineties here am I'm, I'm dating myself again, but that's okay. You know, you, you didn't see much productivity going on when the PC boom started in the eighties, but the nineties, it kicked in and pre pandemic, you know, productivity in the us and Europe anyway has been going down. But I feel like Mike, listen to what you just described. I, how many meetings have we been in where people are arguing about them numbers, what are the assumptions on the numbers wasting so much time? And then nothing gets done and they, then they, they bolt cut that away and you drive in productivity. So I feel like we're on a Renaissance of productivity and a lot of that's gonna be driven by, by data. Yeah. And obviously communications the whole 5g thing. We'll see how that builds out. But data is really the main spring of, I think, a new, new Renaissance in productivity. >>Well, first of all, if you could find an enterprise where you ask the question, would you rather use your data better? And they say, no, like, you know, show me, tell me that I'll short their stock immediately. But I do agree. And I, unfortunately I have a career history in that meeting that you just described where someone doesn't like, what you're showing them. And their first reaction is to say, where'd you get that data? You know, I don't trust it. You know? So they just undermined your entire argument with an invalid way of doing so. Right. When you walk into a meeting with Sigma where'd, where'd you get that data? I was like, that's the live data right now? What question do you want answer >>Lineage, right. Yeah. And you know, it's a Sen's book about, you know, gotta move faster. I mean, this is an example of just cutting through making decisions faster because you're right. Mike and the P the P and L manager in a meeting can, can kill the entire conversation, you know, throw FUD at it. Yeah. You know, protect his or her agenda. >>True. But now to be fair to the person, who's tended to do that. Part of the reason they've done that is that they haven't had access to that data before the meeting and they're getting blindsided. Right. So going back to the collaboration point. Yes. Right. The fact we're coming to this discussion more informed in and of itself takes care of some of that problem. Yeah. >>For sure. And if, and if everybody then agrees, we can move on and now talk about the really important stuff. Yeah. That's good. It >>Seems to me that Sigma is an enabler of that curiosity that you mentioned that that's been lacking. People need to be able to hire for that, but you've got a platform that's going here. You go ask >>Away. That's right in the we're very good. You know, we love being a SaaS platform. There's a lot of telemetry. We can watch what we call our mouse to Dows, you know, which is our monthly average users to our daily average users. We can see what level of user they are, what type of artifacts they build. Are they, you know, someone that creates things from scratch, are they people that tend to increment them, which by the way, is helpful to our customers because we can then advise them, Hey, here's, what's really going on. You might wanna work with this team over here. They could probably be a little better of us using the data, but look at this team over here, you know, they've originated five workbooks in the last, you know, six days they're really on it. There's, there's, you know, that ability to even train for the curiosity that you're referring to is now there, >>Where are your customer conversations? Are they at the lines of business? Are they with the chief data officer? What does that look like these days? >>Great question. So stepping back a bit, what, what is Sigma here to do? And, and our first phase is really to replace spreadsheets, right? And so one of the interesting things about the company is that there isn't a department where a spreadsheet isn't used. So Sigma has an enormous Tam, but also isn't necessarily associated with any particular department or any particular vertical. So when we tend to have conversations, it really depends on, you know, either what kind of investment are you making? A lot of mid-market companies are making best technology investments. They're on a public cloud, they're buying snowflake and they wanna understand what's, what's built to really make this work best over the next number of years. And those are very short sales for us because we, we prove that, you know, in, in minutes to hours, if you're working at a large enterprise and you have three or four other tools, you're asking a different question. >>And often you're asking a question of what I call exploration. We have a product that has dashboards and they've been working for us and we don't wanna replace the dashboard. But when we have a question about the data in the dashboard, we're stuck, how do we get to the raw data? How do we get to the example that we can actually manage? You can't manage a dashboard. You can't manage a trend line, but if you get into the data behind the trend line, you can make decisions to change business process, to change quality, accuracy, to change speed of execution. That is what we're trying to enable. Those conversations happen between the it team who runs technology and the business teams who are responsible for the decisions. So we are, you know, we have a cross departmental sale, but across every department, >>One of the things we're not talking about at this event, which is kind of interesting, cause it's all we've been talking about is the macro supply chain challenges, Ukraine, blah, blah, blah, and the stock market. But, but how are you thinking about that? Macro? The impacts you're seeing, you know, a lot of private companies being, you know, recapped, et cetera, you guys obviously very well funded. Yeah. But how do you think about, I mean, I asked Frank a similar question. He's like, look, it's a marathon. We don't worry about it. We, you know, they made the public market, they get 5 billion in cash. Yeah. Yeah. How are you thinking about it? >>You know, first of all, what's the expression, right? You never, never waste a good, you know, in this case recession, no, we don't have one yet, but the impetus is there, right. People are worried. And when they're worried, they're thinking about their bottom lines, they're thinking about where they're going to get efficiency and their costs. They're already dealing with the supply chain issues of inventory. We all have it in our personal lives. If you've ordered anything in the last six months, you're used to getting it in, you know, days to weeks. And now you're getting in months, you know, we had customers like us foods as a good example, like they're constantly trying to align inventory. They have with transportation that gets that inventory to their end customers, right? And they do that with better data accuracy at the end point, working with us on what we are launching. >>And I mentioned earlier, having more people be able to update that data creates more data, accuracy creates better decisions. We align that then with them and better collaboration with the folks that then coordinate the trucks with Prologis and the panel yesterday, they're the only commercial public company that reports their, their valuations on a quarterly basis. They work with Sigma to trim the amount of time it takes their finance team to produce that data that creates investor confidence that holds up your stock price. So I mean the, the importance of data relative to all the stakeholders in enterprise cannot be overstated. Supply chain is a great example. And yes, it's a marathon because a lot of the technology that drives supply chain is old, but you don't have to rip out those systems to put your data into snowflake, to get better access through Sigma, to enable the people in your environment to make better decisions. And that's the good news. So for me, while I agree, there's a marathon. I think that most of the, I dunno if I could continue this metaphor, but I think we could run quite far down that marathon without an awful lot of energy by just making those couple of changes. >>Awesome. Mike, this has been fantastic. Last question. I, I can tell, I know a lot of growth for Sigma. I can feel it in your energy alone. What are some of the key priorities that you're gonna be focusing on for the rest of the year? >>Our number one priority, our number two priority and number three priority are always build the best product on the market, right? We, we want customers to increase usage. We want them to be delighted. You know, we want them to be RA. Like we have customers at our booth that walk up and it's like, you're building a great company. We love your product. I, if you want to show up happy at work, have customers come up proactively and tell you how your products changed their life. And that is, that is the absolute, most important thing because the real marathon here is that enablement over the long term, right? It is being a great provider to a bunch of great companies under that. We are growing, you know, we've been tripling the company for the fast few years, every year, that takes a lot of hiring. So I would've alongside product is building a great culture with bringing the best people to the company that I guess have my energy level. >>You know, if you could get paid in energy, we would've more than tripled it, you know, but that's always gonna be number two, where we're focused on the segment side, you know, is really the large enterprise customer. At this point, we are doing a great job in the mid-market. We have customer, we have hundreds of customers in our free trial on a constant basis. I think that without wanting to seem over confident or arrogant, I think our technology speaks for itself and the product experience for those users, making a great ROI case to a large enterprise takes effort. It's a different motion. We're, we're very committed to building that motion. We're very committed to building out the partner ecosystem that has been doing that for years. And that is now coming around to the, the snowflake and all of the ecosystem changes around snowflake because they've learned these customers for decades and now have a new opportunity to bring to them. How do we enable them? That is where you're gonna see Sigma going over the next couple of years. >>Wow, fantastic. Good stuff. And a lot of momentum, Mike, thank you so much for joining Dave and me talking about Sigma, the momentum, the flywheel of what you're doing with snowflake and what you're enabling customers to achieve the massive business outcomes. Really cool stuff. >>Thank you. And thank you for continuing to give us a platform to do this and glad to be back in conferences, doing it face to face. It's fantastic. >>It it's the best. Awesome. Mike, thank you for Mike Palmer and Dave ante. I'm Lisa Martin. You've been watching the cube hopefully all day. We've been here since eight o'clock this morning, Pacific time giving you wall the wall coverage of snowflake summit 22 signing off for today. Dave and I will see you right bright and early tomorrow morning. I will take care guys.

Published Date : Jun 16 2022

SUMMARY :

And we have an alumni back with us. And I guess again So talk to the audience about Sigma before we get into the snowflake partnership and what you guys are doing from a technical the one that makes the business decision in the marketing department and the finance team, you know, in the works in How are they as an investor? know, all of the rest doesn't matter if the product is not great, we work directly on And the connection So, but if I may just to follow up on that, I mean, you could do that without some of the things that I will, you know, leaning forward that we've heard here that we're very excited about is And we go toward transactional databases. And you're enabling those business users to have access to that data. do I have inventory in the warehouse when the order comes in? Part of that is just the efficiencies. You know, and that's another example, but I'll even pivot to the various REI You know, I think a lot of companies consider Good, good definition. of an enterprise and for me and for the Sigma customer base, there's really not much you And that's the change that we're gonna see in business over the next few years, You know, if there's two people in the phone system, it's not that valuable, right. And that's what we're seeing now applied to data. You know, the ones where, you know, you think you're buying something, Mike, listen to what you just described. And their first reaction is to say, where'd you get that data? you know, throw FUD at it. So going back to the collaboration point. And if, and if everybody then agrees, we can move on and now talk about the really important stuff. Seems to me that Sigma is an enabler of that curiosity that you mentioned that that's been lacking. We can watch what we call our mouse to Dows, you know, which is our monthly average users to our daily we prove that, you know, in, in minutes to hours, if you're working at a large enterprise and you have three or four other So we are, you know, we have a cross departmental sale, but across every department, you know, a lot of private companies being, you know, recapped, et cetera, you guys obviously very You never, never waste a good, you know, in this case recession, And I mentioned earlier, having more people be able to update that data creates more data, What are some of the key priorities that you're gonna be focusing on for the We are growing, you know, we've been tripling the company for the fast few years, You know, if you could get paid in energy, we would've more than tripled it, you know, but that's always gonna And a lot of momentum, Mike, thank you so much for joining Dave and me talking about Sigma, And thank you for continuing to give us a platform to do this and glad to be back in conferences, Dave and I will see you right bright and early tomorrow morning.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Lisa MartinPERSON

0.99+

Dave LantaPERSON

0.99+

FrankPERSON

0.99+

MikePERSON

0.99+

Mike PalmerPERSON

0.99+

threeQUANTITY

0.99+

DavePERSON

0.99+

MetcalfPERSON

0.99+

10 timesQUANTITY

0.99+

Jason FrancePERSON

0.99+

Rob willPERSON

0.99+

5 billionQUANTITY

0.99+

Bob MetcalfPERSON

0.99+

Frank SalmanPERSON

0.99+

SigmaORGANIZATION

0.99+

two peopleQUANTITY

0.99+

hundredsQUANTITY

0.99+

yesterdayDATE

0.99+

PrologisORGANIZATION

0.99+

first bookQUANTITY

0.99+

first timeQUANTITY

0.99+

EuropeLOCATION

0.99+

two daysQUANTITY

0.99+

first phaseQUANTITY

0.99+

todayDATE

0.99+

VegasLOCATION

0.99+

five minutes agoDATE

0.99+

eightiesDATE

0.98+

six daysQUANTITY

0.98+

five workbooksQUANTITY

0.98+

one placeQUANTITY

0.98+

10 minutesQUANTITY

0.98+

oneQUANTITY

0.98+

eight o'clock this morningDATE

0.98+

first oneQUANTITY

0.98+

Snowflake Summit 2022EVENT

0.97+

SnowflakeORGANIZATION

0.97+

pets.comORGANIZATION

0.97+

billion license spreadsheetQUANTITY

0.96+

ninetiesDATE

0.96+

day twoQUANTITY

0.95+

OneQUANTITY

0.95+

four other toolsQUANTITY

0.94+

Sigma ComputingORGANIZATION

0.93+

two amazing engineersQUANTITY

0.92+

snowflake summit 22EVENT

0.92+

pandemicEVENT

0.9+

first reactionQUANTITY

0.89+

Sigma computingORGANIZATION

0.88+

UkraineLOCATION

0.86+

next couple of yearsDATE

0.85+

last six monthsDATE

0.85+

pest.comORGANIZATION

0.84+

early tomorrow morningDATE

0.82+

firstQUANTITY

0.82+

next few yearsDATE

0.77+

twoQUANTITY

0.76+

decadesQUANTITY

0.76+

Dr Eng Lim Goh, High Performance Computing & AI | HPE Discover 2021


 

>>Welcome back to HPD discovered 2021 the cubes virtual coverage, continuous coverage of H P. S H. P. S. Annual customer event. My name is Dave Volonte and we're going to dive into the intersection of high performance computing data and AI with DR Eng limb go who is the senior vice president and CTO for AI at Hewlett Packard enterprise Doctor go great to see you again. Welcome back to the cube. >>Hello Dave, Great to talk to you again. >>You might remember last year we talked a lot about swarm intelligence and how AI is evolving. Of course you hosted the day two keynotes here at discover you talked about thriving in the age of insights and how to craft a data centric strategy and you addressed you know some of the biggest problems I think organizations face with data that's You got a data is plentiful but insights they're harder to come by. And you really dug into some great examples in retail banking and medicine and health care and media. But stepping back a little bit with zoom out on discovered 21, what do you make of the events so far? And some of your big takeaways? >>Mm Well you started with the insightful question, right? Yeah. Data is everywhere then. But we like the insight. Right? That's also part of the reason why that's the main reason why you know Antonio on day one focused and talked about that. The fact that we are now in the age of insight. Right? Uh and and uh and and how to thrive thrive in that in this new age. What I then did on the day to kino following Antonio is to talk about the challenges that we need to overcome in order in order to thrive in this new age. >>So maybe we could talk a little bit about some of the things that you took away in terms I'm specifically interested in some of the barriers to achieving insights when you know customers are drowning in data. What do you hear from customers? What we take away from some of the ones you talked about today? >>Oh, very pertinent question. Dave you know the two challenges I spoke about right now that we need to overcome in order to thrive in this new age. The first one is is the current challenge and that current challenge is uh you know stated is you know, barriers to insight, you know when we are awash with data. So that's a statement right? How to overcome those barriers. What are the barriers of these two insight when we are awash in data? Um I in the data keynote I spoke about three main things. Three main areas that received from customers. The first one, the first barrier is in many with many of our customers. A data is siloed. All right. You know, like in a big corporation you've got data siloed by sales, finance, engineering, manufacturing, and so on, uh supply chain and so on. And uh, there's a major effort ongoing in many corporations to build a federation layer above all those silos so that when you build applications above they can be more intelligent. They can have access to all the different silos of data to get better intelligence and more intelligent applications built. So that was the that was the first barrier we spoke about barriers to incite when we are washed with data. The second barrier is uh, that we see amongst our customers is that uh data is raw and dispersed when they are stored and and uh and you know, it's tough to get tough to to get value out of them. Right? And I in that case I I used the example of uh you know the May 6 2010 event where the stock market dropped a trillion dollars in in tens of ministerial. We we all know those who are financially attuned with know about this uh incident But this is not the only incident. There are many of them out there and for for that particular May six event uh you know, it took a long time to get insight months. Yeah before we for months we had no insight as to what happened, why it happened, right. Um and and there were many other incidences like this. And the regulators were looking for that one rule that could, that could mitigate many of these incidences. Um one of our customers decided to take the hard road go with the tough data right? Because data is rolling dispersed. So they went into all the different feeds of financial transaction information. Uh took the took the tough uh took the tough road and analyze that data took a long time to assemble and they discovered that there was court stuffing right? That uh people were sending a lot of traits in and then cancelling them almost immediately. You have to manipulate the market. Um And why why why didn't we see it immediately? Well the reason is the process reports that everybody sees uh rule in there that says all trades. Less than 100 shares don't need to report in there. And so what people did was sending a lot of less than 103 100 100 shares trades uh to fly under the radar to do this manipulation. So here is here the second barrier right? Data could be raw and dispersed. Um Sometimes you just have to take the hard road and um and to get insight And this is 1 1 great example. And then the last barrier is uh is has to do with sometimes when you start a project to to get insight to get uh to get answers and insight. You you realize that all the datas around you but you don't you don't seem to find the right ones To get what you need. You don't you don't seem to get the right ones. Yeah. Um here we have three quick examples of customers. 111 was it was a great example right? Where uh they were trying to build a language translator, a machine language translator between two languages. Right? But not do that. They need to get hundreds of millions of word pairs, you know, of one language compared uh with the corresponding other hundreds of millions of them. They say we are going to get all these word pairs. Someone creative thought of a willing source and a huge, so it was a United Nations you see. So sometimes you think you don't have the right data with you, but there might be another source and a willing one that could give you that data right. The second one has to do with uh there was uh the uh sometimes you you may just have to generate that data, interesting one. We had an autonomous car customer that collects all these data from their cars, right, massive amounts of data, loss of senses, collect loss of data. And uh you know, but sometimes they don't have the data they need even after collection. For example, they may have collected the data with a car uh in in um in fine weather and collected the car driving on this highway in rain and also in stone, but never had the opportunity to collect the car in hale because that's a rare occurrence. So instead of waiting for a time where the car can dr inhale, they build a simulation you by having the car collector in snow and simulated him. So these are some of the examples where we have customers working to overcome barriers, right? You have barriers that is associated the fact that data is silo Federated, it various associated with data. That's tough to get that. They just took the hard road, right? And sometimes, thirdly, you just have to be creative to get the right data you need, >>wow, I tell you, I have about 100 questions based on what you just said. Uh, there's a great example, the flash crash. In fact, Michael Lewis wrote about this in his book, The Flash Boys and essentially right. It was high frequency traders trying to front run the market and sending in small block trades trying to get on the front end it. So that's and they, and they chalked it up to a glitch like you said, for months, nobody really knew what it was. So technology got us into this problem. I guess my question is, can technology help us get out of the problem? And that maybe is where AI fits in. >>Yes, yes. Uh, in fact, a lot of analytics, we went in, uh, to go back to the raw data that is highly dispersed from different sources, right, assemble them to see if you can find a material trend, right? You can see lots of trends right? Like, uh, you know, we, if if humans look at things right, we tend to see patterns in clouds, right? So sometimes you need to apply statistical analysis, um math to be sure that what the model is seeing is is real. Right? And and that required work. That's one area. The second area is uh you know, when um uh there are times when you you just need to to go through that uh that tough approach to to find the answer. Now, the issue comes to mind now is is that humans put in the rules to decide what goes into a report that everybody sees in this case uh before the change in the rules. Right? But by the way, after the discovery, the authorities change the rules and all all shares, all traits of different any sizes. It has to be reported. No. Yeah. Right. But the rule was applied uh you know, to say earlier that shares under 100 trades under 100 shares need not be reported. So sometimes you just have to understand that reports were decided by humans and and under for understandable reasons. I mean they probably didn't want that for various reasons not to put everything in there so that people could still read it uh in a reasonable amount of time. But uh we need to understand that rules were being put in by humans for the reports we read. And as such, there are times you just need to go back to the raw data. >>I want to ask, >>albeit that it's gonna be tough. >>Yeah. So I want to ask a question about AI is obviously it's in your title and it's something you know a lot about but and I want to make a statement, you tell me if it's on point or off point. So it seems that most of the Ai going on in the enterprise is modeling data science applied to troves of data >>but >>but there's also a lot of ai going on in consumer whether it's you know, fingerprint technology or facial recognition or natural language processing. Will a two part question will the consumer market has so often in the enterprise sort of inform us uh the first part and then will there be a shift from sort of modeling if you will to more you mentioned autonomous vehicles more ai influencing in real time. Especially with the edge. She can help us understand that better. >>Yeah, it's a great question. Right. Uh there are three stages to just simplify, I mean, you know, it's probably more sophisticated than that but let's simplify three stages. All right. To to building an Ai system that ultimately can predict, make a prediction right or to to assist you in decision making, have an outcome. So you start with the data massive amounts data that you have to decide what to feed the machine with. So you feed the machine with this massive chunk of data and the machine uh starts to evolve a model based on all the data is seeing. It starts to evolve right to the point that using a test set of data that you have separately campus site that you know the answer for. Then you test the model uh you know after you trained it with all that data to see whether it's prediction accuracy is high enough and once you are satisfied with it, you you then deploy the model to make the decision and that's the influence. Right? So a lot of times depend on what what we are focusing on. We we um in data science are we working hard on assembling the right data to feed the machine with, That's the data preparation organization work. And then after which you build your models, you have to pick the right models for the decisions and prediction you wanted to make. You pick the right models and then you start feeding the data with it. Sometimes you you pick one model and the prediction isn't that robust, it is good but then it is not consistent right now what you do is uh you try another model so sometimes it's just keep trying different models until you get the right kind. Yeah, that gives you a good robust decision making and prediction after which It is tested well Q eight. You would then take that model and deploy it at the edge. Yeah. And then at the edges is essentially just looking at new data, applying it to the model, you're you're trained and then that model will give you a prediction decision. Right? So uh it is these three stages. Yeah, but more and more uh you know, your question reminds me that more and more people are thinking as the edge become more and more powerful. Can you also do learning at the edge? Right. That's the reason why we spoke about swarm learning the last time, learning at the edge as a swamp, right? Because maybe individually they may not have enough power to do so. But as a swampy me, >>is that learning from the edge or learning at the edge? In other words? Yes. Yeah. Question Yeah. >>That's a great question. That's a great question. Right? So uh the quick answer is learning at the edge, right? Uh and also from the edge, but the main goal, right? The goal is to learn at the edge so that you don't have to move the data that the Edge sees first back to the cloud or the core to do the learning because that would be the reason. One of the main reasons why you want to learn at the edge, right? Uh So so that you don't need to have to send all that data back and assemble it back from all the different edge devices, assemble it back to the cloud side to to do the learning right? With swampland. You can learn it and keep the data at the edge and learn at that point. >>And then maybe only selectively send the autonomous vehicle example you gave us. Great because maybe there, you know, there may be only persisting, they're not persisting data that is inclement weather or when a deer runs across the front and then maybe they they do that and then they send that smaller data set back and maybe that's where it's modelling done. But the rest can be done at the edges. It's a new world that's coming down. Let me ask you a question, is there a limit to what data should be collected and how it should be collected? >>That's a great question again. You know uh wow today, full of these uh insightful questions that actually touches on the second challenge. Right? How do we uh in order to thrive in this new age of inside? The second challenge is are you know the is our future challenge, right? What do we do for our future? And and in there is uh the statement we make is we have to focus on collecting data strategically for the future of our enterprise. And within that I talk about what to collect right? When to organize it when you collect and then where will your data be, you know going forward that you are collecting from? So what, when and where for the what data for the what data to collect? That? That was the question you ask. Um it's it's a question that different industries have to ask themselves because it will vary, right? Um let me give you the you use the autonomous car example, let me use that. And you have this customer collecting massive amounts of data. You know, we're talking about 10 petabytes a day from the fleet of their cars. And these are not production autonomous cars, right? These are training autonomous cars collecting data so they can train and eventually deploy commercial cars, right? Um so this data collection cars they collect as a fleet of them collect temporal bikes a day. And when it came to us building a storage system to store all of that data, they realized they don't want to afford to store all of it. Now, here comes the dilemma, right? What should I after I spent so much effort building all these cars and sensors and collecting data, I've now decide what to delete. That's a dilemma right now in working with them on this process of trimming down what they collected. You know, I'm constantly reminded of the sixties and seventies, right? To remind myself 60 and seventies, we call a large part of our D. N. A junk DNA. Today. We realize that a large part of that what we call john has function as valuable function. They are not jeans, but they regulate the function of jeans, you know, So, so what's jump in the yesterday could be valuable today or what's junk today could be valuable tomorrow, Right? So, so there's this tension going on right between you decided not wanting to afford to store everything that you can get your hands on. But on the other hand, you you know, you worry you you you ignore the wrong ones, right? You can see this tension in our customers, right? And it depends on industry here, right? In health care, they say I have no choice. I I want it. All right. One very insightful point brought up by one health care provider that really touched me was, you know, we are not we don't only care. Of course we care a lot. We care a lot about the people we are caring for, right? But you also care for the people were not caring for. How do we find them? Mhm. Right. And that therefore, they did not just need to collect data. That is that they have with from their patients. They also need to reach out right to outside data so that they can figure out who they are not caring for, right? So they want it all. So I tell us them, so what do you do with funding if you want it all? They say they have no choice but to figure out a way to fund it and perhaps monetization of what they have now is the way to come around and find that. Of course they also come back to us rightfully that you know, we have to then work out a way to help them build that system, you know? So that's health care, right? And and if you go to other industries like banking, they say they can't afford to keep them off, but they are regulated, seems like healthcare, they are regulated as to uh privacy and such. Like so many examples different industries having different needs, but different approaches to how what they collect. But there is this constant tension between um you perhaps deciding not wanting to fund all of that uh all that you can store, right? But on the other hand, you know, if you if you kind of don't want to afford it and decide not to store some uh if he does some become highly valuable in the future, right? Yeah. >>We can make some assumptions about the future, can't we? I mean, we know there's gonna be a lot more data than than we've ever seen before. We know that we know well notwithstanding supply constraints on things like nand. We know the prices of storage is going to continue to decline. We also know, and not a lot of people are really talking about this but the processing power but he says moore's law is dead okay. It's waning. But the processing power when you combine the Cpus and NP US and GPUS and accelerators and and so forth actually is is increasing. And so when you think about these use cases at the edge, you're going to have much more processing power, you're gonna have cheaper storage and it's going to be less expensive processing And so as an ai practitioner, what can you do with that? >>Yeah, it's highly again, another insightful questions that we touched on our keynote and that that goes up to the why I do the where? Right, When will your data be? Right. We have one estimate that says that by next year there will be 55 billion connected devices out there. Right. 55 billion. Right. What's the population of the world? Of the other? Of 10 billion? But this thing is 55 billion. Right? Uh and many of them, most of them can collect data. So what do you what do you do? Right. Um So the amount of data that's gonna come in, it's gonna weigh exceed right? Our drop in storage costs are increasing computer power. Right? So what's the answer? Right. So, so the the answer must be knowing that we don't and and even the drop in price and increase in bandwidth, it will overwhelm the increased five G will overwhelm five G. Right? Given amount 55 billion of them collecting. Right? So, the answer must be that there might need to be a balance between you needing to bring all that data from the 55 billion devices of data back to a central as a bunch of central Cause because you may not be able to afford to do that firstly band with even with five G. M and and SD when you'll still be too expensive given the number of devices out there. Were you given storage cause dropping will still be too expensive to try and store them all. So the answer must be to start at least to mitigate the problem to some leave both a lot of the data out there. Right? And only send back the pertinent ones as you said before. But then if you did that, then how are we gonna do machine learning at the core and the cloud side? If you don't have all the data you want rich data to train with. Right? Some sometimes you want a mix of the uh positive type data and the negative type data so you can train the machine in a more balanced way. So the answer must be eventually right. As we move forward with these huge number of devices out of the edge to do machine learning at the edge. Today, we don't have enough power. Right? The edge typically is characterized by a lower uh, energy capability and therefore lower compute power. But soon, you know, even with lower energy, they can do more with compute power improving in energy efficiency, Right? Uh, so learning at the edge today, we do influence at the edge. So we data model deploy and you do influence at the age, that's what we do today. But more and more, I believe, given a massive amount of data at the edge, you you have to have to start doing machine learning at the edge. And and if when you don't have enough power, then you aggregate multiple devices, compute power into a swamp and learn as a swan, >>interesting. So now, of course, if I were sitting and fly on the wall in HP board meeting, I said, okay, HP is as a leading provider of compute, how do you take advantage of that? I mean, we're going, I know it's future, but you must be thinking about that and participating in those markets. I know today you are you have, you know, edge line and other products. But there's it seems to me that it's it's not the general purpose that we've known in the past. It's a new type of specialized computing. How are you thinking about participating in that >>opportunity for your customers? Uh the world will have to have a balance right? Where today the default, Well, the more common mode is to collect the data from the edge and train at uh at some centralized location or a number of centralized location um going forward. Given the proliferation of the edge devices, we'll need a balance. We need both. We need capability at the cloud side. Right. And it has to be hybrid. And then we need capability on the edge side. Yeah. That they want to build systems that that on one hand, uh is uh edge adapted, right? Meaning the environmentally adapted because the edge different they are on a lot of times on the outside. Uh They need to be packaging adapted and also power adapted, right? Because typically many of these devices are battery powered. Right? Um so you have to build systems that adapt to it, but at the same time they must not be custom. That's my belief. They must be using standard processes and standard operating system so that they can run rich a set of applications. So yes. Um that's that's also the insightful for that Antonio announced in 2018, Uh the next four years from 2018, right, $4 billion dollars invested to strengthen our edge portfolio, edge product lines, right Edge solutions. >>I get a doctor go. I could go on for hours with you. You're you're just such a great guest. Let's close what are you most excited about in the future of of of it? Certainly H. P. E. But the industry in general. >>Yeah I think the excitement is uh the customers right? The diversity of customers and and the diversity in a way they have approached their different problems with data strategy. So the excitement is around data strategy right? Just like you know uh you know the the statement made was was so was profound. Right? Um And Antonio said we are in the age of insight powered by data. That's the first line right? The line that comes after that is as such were becoming more and more data centric with data the currency. Now the next step is even more profound. That is um you know we are going as far as saying that you know um data should not be treated as cost anymore. No right. But instead as an investment in a new asset class called data with value on our balance sheet, this is a this is a step change right in thinking that is going to change the way we look at data the way we value it. So that's a statement that this is the exciting thing because because for for me a city of AI right uh machine is only as intelligent as the data you feed it with. Data is a source of the machine learning to be intelligent. So so that's that's why when when people start to value data right? And and and say that it is an investment when we collect it. It is very positive for ai because an Ai system gets intelligent, more intelligence because it has a huge amounts of data and the diversity of data. So it'd be great if the community values values data. Well >>you certainly see it in the valuations of many companies these days. Um and I think increasingly you see it on the income statement, you know data products and people monetizing data services and maybe eventually you'll see it in the in the balance. You know Doug Laney when he was a gardener group wrote a book about this and a lot of people are thinking about it. That's a big change isn't it? Dr >>yeah. Question is is the process and methods evaluation. Right. But uh I believe we'll get there, we need to get started then we'll get their belief >>doctor goes on and >>pleasure. And yeah and then the yeah I will will will will benefit greatly from it. >>Oh yeah, no doubt people will better understand how to align you know, some of these technology investments, Doctor goes great to see you again. Thanks so much for coming back in the cube. It's been a real pleasure. >>Yes. A system. It's only as smart as the data you feed it with. >>Excellent. We'll leave it there. Thank you for spending some time with us and keep it right there for more great interviews from HP discover 21. This is dave a lot for the cube. The leader in enterprise tech coverage right back.

Published Date : Jun 17 2021

SUMMARY :

at Hewlett Packard enterprise Doctor go great to see you again. the age of insights and how to craft a data centric strategy and you addressed you know That's also part of the reason why that's the main reason why you know Antonio on day one So maybe we could talk a little bit about some of the things that you The first one is is the current challenge and that current challenge is uh you know stated So that's and they, and they chalked it up to a glitch like you said, is is that humans put in the rules to decide what goes into So it seems that most of the Ai going on in the enterprise is modeling be a shift from sort of modeling if you will to more you mentioned autonomous It starts to evolve right to the point that using a test set of data that you have is that learning from the edge or learning at the edge? The goal is to learn at the edge so that you don't have to move the data that the And then maybe only selectively send the autonomous vehicle example you gave us. But on the other hand, you know, if you if you kind of don't want to afford it and But the processing power when you combine the Cpus and NP that there might need to be a balance between you needing to bring all that data from the I know today you are you have, you know, edge line and other products. Um so you have to build systems that adapt to it, but at the same time they must not Let's close what are you most excited about in the future of machine is only as intelligent as the data you feed it with. Um and I think increasingly you see it on the income statement, you know data products and Question is is the process and methods evaluation. And yeah and then the yeah I will will will will benefit greatly from it. Doctor goes great to see you again. It's only as smart as the data you feed it with. Thank you for spending some time with us and keep it right there for more great

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Michael LewisPERSON

0.99+

Dave VolontePERSON

0.99+

DavePERSON

0.99+

2018DATE

0.99+

HPORGANIZATION

0.99+

two languagesQUANTITY

0.99+

The Flash BoysTITLE

0.99+

55 billionQUANTITY

0.99+

10 billionQUANTITY

0.99+

second challengeQUANTITY

0.99+

Hewlett PackardORGANIZATION

0.99+

two challengesQUANTITY

0.99+

second areaQUANTITY

0.99+

one languageQUANTITY

0.99+

TodayDATE

0.99+

last yearDATE

0.99+

Doug LaneyPERSON

0.99+

tomorrowDATE

0.99+

next yearDATE

0.99+

bothQUANTITY

0.99+

OneQUANTITY

0.99+

todayDATE

0.99+

first lineQUANTITY

0.99+

first partQUANTITY

0.99+

May 6 2010DATE

0.99+

$4 billion dollarsQUANTITY

0.99+

two partQUANTITY

0.99+

Less than 100 sharesQUANTITY

0.99+

HPDORGANIZATION

0.99+

one modelQUANTITY

0.98+

one ruleQUANTITY

0.98+

one areaQUANTITY

0.98+

second barrierQUANTITY

0.98+

60QUANTITY

0.98+

55 billion devicesQUANTITY

0.98+

AntonioPERSON

0.98+

johnPERSON

0.98+

three stagesQUANTITY

0.98+

hundreds of millionsQUANTITY

0.97+

about 100 questionsQUANTITY

0.97+

Eng Lim GohPERSON

0.97+

HPEORGANIZATION

0.97+

first barrierQUANTITY

0.97+

first oneQUANTITY

0.97+

Three main areasQUANTITY

0.97+

yesterdayDATE

0.96+

tens of ministerialQUANTITY

0.96+

two insightQUANTITY

0.96+

Q eightOTHER

0.95+

2021DATE

0.94+

seventiesQUANTITY

0.94+

two keynotesQUANTITY

0.93+

a dayQUANTITY

0.93+

firstQUANTITY

0.92+

H P. S H. P. S. Annual customerEVENT

0.91+

United NationsORGANIZATION

0.91+

less than 103 100 100 sharesQUANTITY

0.91+

under 100 tradesQUANTITY

0.9+

under 100 sharesQUANTITY

0.9+

day oneQUANTITY

0.88+

about 10 petabytes a dayQUANTITY

0.88+

three quick examplesQUANTITY

0.85+

one health care providerQUANTITY

0.85+

one estimateQUANTITY

0.84+

three main thingsQUANTITY

0.83+

hundreds of millions of word pairsQUANTITY

0.82+

AntonioORGANIZATION

0.81+

sixtiesQUANTITY

0.78+

oneQUANTITY

0.77+

May sixDATE

0.75+

firstlyQUANTITY

0.74+

trillion dollarsQUANTITY

0.73+

second oneQUANTITY

0.71+

HP discover 21ORGANIZATION

0.69+

DR Eng limbPERSON

0.69+

one of our customersQUANTITY

0.66+

Dr Eng Lim Goh, Vice President, CTO, High Performance Computing & AI


 

(upbeat music) >> Welcome back to HPE Discover 2021, theCube's virtual coverage, continuous coverage of HPE's annual customer event. My name is Dave Vellante and we're going to dive into the intersection of high-performance computing, data and AI with Dr. Eng Lim Goh who's a Senior Vice President and CTO for AI at Hewlett Packard Enterprise. Dr. Goh, great to see you again. Welcome back to theCube. >> Hey, hello, Dave. Great to talk to you again. >> You might remember last year we talked a lot about swarm intelligence and how AI is evolving. Of course you hosted the Day 2 keynotes here at Discover. And you talked about thriving in the age of insights and how to craft a data-centric strategy and you addressed some of the biggest problems I think organizations face with data. And that's, you got to look, data is plentiful, but insights, they're harder to come by and you really dug into some great examples in retail, banking, and medicine and healthcare and media. But stepping back a little bit we'll zoom out on Discover '21, you know, what do you make of the events so far and some of your big takeaways? >> Hmm, well, you started with the insightful question. Data is everywhere then but we lack the insight. That's also part of the reason why that's a main reason why, Antonio on Day 1 focused and talked about that, the fact that we are in the now in the age of insight and how to thrive in this new age. What I then did on the Day 2 keynote following Antonio is to talk about the challenges that we need to overcome in order to thrive in this new age. >> So maybe we could talk a little bit about some of the things that you took away in terms of, I'm specifically interested in some of the barriers to achieving insights when you know customers are drowning in data. What do you hear from customers? What were your takeaway from some of the ones you talked about today? >> Very pertinent question, Dave. You know, the two challenges I spoke about how to, that we need to overcome in order to thrive in this new age, the first one is the current challenge. And that current challenge is, you know state of this, you know, barriers to insight, when we are awash with data. So that's a statement. How to overcome those barriers. One of the barriers to insight when we are awash in data, in the Day 2 keynote, I spoke about three main things, three main areas that receive from customers. The first one, the first barrier is with many of our customers, data is siloed. You know, like in a big corporation, you've got data siloed by sales, finance, engineering, manufacturing, and so on supply chain and so on. And there's a major effort ongoing in many corporations to build a Federation layer above all those silos so that when you build applications above they can be more intelligent. They can have access to all the different silos of data to get better intelligence and more intelligent applications built. So that was the first barrier we spoke about, you know, barriers to insight when we are awash with data. The second barrier is that we see amongst our customers is that data is raw and disperse when they are stored. And it's tough to get to value out of them. In that case I use the example of the May 6, 2010 event where the stock market dropped a trillion dollars in tens of minutes. We all know those who are financially attuned with, know about this incident. But that this is not the only incident. There are many of them out there. And for that particular May 6, event, you know it took a long time to get insight, months, yeah, before we, for months we had no insight as to what happened, why it happened. And there were many other incidences like this and the regulators were looking for that one rule that could mitigate many of these incidences. One of our customers decided to take the hard road to go with the tough data. Because data is raw and dispersed. So they went into all the different feeds of financial transaction information, took the tough, you know, took a tough road and analyze that data took a long time to assemble. And he discovered that there was quote stuffing. That people were sending a lot of trades in and then canceling them almost immediately. You have to manipulate the market. And why didn't we see it immediately? Well, the reason is the process reports that everybody sees had the rule in there that says all trades less than 100 shares don't need to report in there. And so what people did was sending a lot of less than 100 shares trades to fly under the radar to do this manipulation. So here is, here the second barrier. Data could be raw and disperse. Sometimes it's just have to take the hard road and to get insight. And this is one great example. And then the last barrier has to do with sometimes when you start a project to get insight, to get answers and insight, you realize that all the data's around you, but you don't seem to find the right ones to get what you need. You don't seem to get the right ones, yeah. Here we have three quick examples of customers. One was a great example where they were trying to build a language translator a machine language translator between two languages. But in order to do that they need to get hundreds of millions of word pairs of one language compare with the corresponding other hundreds of millions of them. They say, "Where I'm going to get all these word pairs?" Someone creative thought of a willing source and huge source, it was a United Nations. You see, so sometimes you think you don't have the right data with you, but there might be another source and a willing one that could give you that data. The second one has to do with, there was the, sometimes you may just have to generate that data. Interesting one. We had an autonomous car customer that collects all these data from their cars. Massive amounts of data, lots of sensors, collect lots of data. And, you know, but sometimes they don't have the data they need even after collection. For example, they may have collected the data with a car in fine weather and collected the car driving on this highway in rain and also in snow. But never had the opportunity to collect the car in hail because that's a rare occurrence. So instead of waiting for a time where the car can drive in hail, they build a simulation by having the car collected in snow and simulated hail. So these are some of the examples where we have customers working to overcome barriers. You have barriers that is associated with the fact, that data silo, if federated barriers associated with data that's tough to get at. They just took the hard road. And sometimes thirdly, you just have to be creative to get the right data you need. >> Wow, I tell you, I have about 100 questions based on what you just said. And as a great example, the flash crash in fact Michael Lewis wrote about this in his book, the "Flash Boys" and essentially. It was high frequency traders trying to front run the market and sending in small block trades trying to get sort of front ended. So that's, and they chalked it up to a glitch. Like you said, for months, nobody really knew what it was. So technology got us into this problem. Can I guess my question is can technology help us get get out of the problem? And that maybe is where AI fits in. >> Yes. Yes. In fact, a lot of analytics work went in to go back to the raw data that is highly dispersed from different sources, assemble them to see if you can find a material trend. You can see lots of trends. Like, no, we, if humans at things we tend to see patterns in clouds. So sometimes you need to apply statistical analysis, math to be sure that what the model is seeing is real. And that required work. That's one area. The second area is, you know, when this, there are times when you just need to go through that tough approach to find the answer. Now, the issue comes to mind now is that humans put in the rules to decide what goes into a report that everybody sees. And in this case before the change in the rules. By the way, after the discovery, the authorities changed the rules and all shares all trades of different, any sizes it has to be reported. Not, yeah. But the rule was applied to to say earlier that shares under 100, trades under 100 shares need not be reported. So sometimes you just have to understand that reports were decided by humans and for understandable reasons. I mean, they probably didn't, wanted for various reasons not to put everything in there so that people could still read it in a reasonable amount of time. But we need to understand that rules were being put in by humans for the reports we read. And as such there are times we just need to go back to the raw data. >> I want to ask you-- Or be it that it's going to be tough there. >> Yeah, so I want to ask you a question about AI as obviously it's in your title and it's something you know a lot about and I'm going to make a statement. You tell me if it's on point or off point. Seems that most of the AI going on in the enterprise is modeling data science applied to troves of data. But there's also a lot of AI going on in consumer, whether it's fingerprint technology or facial recognition or natural language processing. Will, to two-part question, will the consumer market, let's say as it has so often in the enterprise sort of inform us is sort of first part. And then will there be a shift from sort of modeling, if you will, to more, you mentioned autonomous vehicles more AI inferencing in real-time, especially with the Edge. I think you can help us understand that better. >> Yeah, this is a great question. There are three stages to just simplify, I mean, you know, it's probably more sophisticated than that, but let's just simplify there're three stages to building an AI system that ultimately can predict, make a prediction. Or to assist you in decision-making, have an outcome. So you start with the data, massive amounts of data that you have to decide what to feed the machine with. So you feed the machine with this massive chunk of data. And the machine starts to evolve a model based on all the data is seeing it starts to evolve. To a point that using a test set of data that you have separately kept a site that you know the answer for. Then you test the model, you know after you're trained it with all that data to see whether his prediction accuracy is high enough. And once you are satisfied with it, you then deploy the model to make the decision and that's the inference. So a lot of times depending on what we are focusing on. We in data science are we working hard on assembling the right data to feed the machine with? That's the data preparation organization work. And then after which you build your models you have to pick the right models for the decisions and prediction you wanted to make. You pick the right models and then you start feeding the data with it. Sometimes you pick one model and a prediction isn't that a robust, it is good, but then it is not consistent. Now what you do is you try another model. So sometimes you just keep trying different models until you get the right kind, yeah, that gives you a good robust decision-making and prediction. Now, after which, if it's tested well, Q8 you will then take that model and deploy it at the Edge, yeah. And then at the Edge is essentially just looking at new data applying it to the model that you have trained and then that model will give you a prediction or a decision. So it is these three stages, yeah. But more and more, your question reminds me that more and more people are thinking as the Edge become more and more powerful, can you also do learning at the Edge? That's the reason why we spoke about swarm learning the last time, learning at the Edge as a swarm. Because maybe individually they may not have enough power to do so, but as a swarm, they may. >> Is that learning from the Edge or learning at the Edge. In other words, is it-- >> Yes. >> Yeah, you don't understand my question, yeah. >> That's a great question. That's a great question. So answer is learning at the Edge, and also from the Edge, but the main goal, the goal is to learn at the Edge so that you don't have to move the data that Edge sees first back to the Cloud or the call to do the learning. Because that would be the reason, one of the main reasons why you want to learn at the Edge. So that you don't need to have to send all that data back and assemble it back from all the different Edge devices assemble it back to the Cloud side to do the learning. With swarm learning, you can learn it and keep the data at the Edge and learn at that point, yeah. >> And then maybe only selectively send the autonomous vehicle example you gave is great 'cause maybe they're, you know, there may be only persisting. They're not persisting data that is an inclement weather, or when a deer runs across the front and then maybe they do that and then they send that smaller data set back and maybe that's where it's modeling done but the rest can be done at the Edge. It's a new world that's coming to, let me ask you a question. Is there a limit to what data should be collected and how it should be collected? >> That's a great question again, yeah, well, today full of these insightful questions that actually touches on the second challenge. How do we, to in order to thrive in this new age of insight. The second challenge is our future challenge. What do we do for our future? And in there is the statement we make is we have to focus on collecting data strategically for the future of our enterprise. And within that, I talk about what to collect, and when to organize it when you collect, and then where will your data be going forward that you are collecting from? So what, when, and where. For the what data, for what data to collect that was the question you asked. It's a question that different industries have to ask themselves because it will vary. Let me give you the, you use the autonomous car example. Let me use that and you have this customer collecting massive amounts of data. You know, we talking about 10 petabytes a day from a fleet of their cars and these are not production autonomous cars. These are training autonomous cars, collecting data so they can train and eventually deploy a commercial cars. Also these data collection cars, they collect 10 as a fleet of them collect 10 petabytes a day. And then when it came to us, building a storage system to store all of that data they realize they don't want to afford to store all of it. Now here comes the dilemma. What should I, after I spent so much effort building all this cars and sensors and collecting data, I've now decide what to delete. That's a dilemma. Now in working with them on this process of trimming down what they collected. I'm constantly reminded of the 60s and 70s. To remind myself 60s and 70s, we call a large part of our DNA, junk DNA. Today we realized that a large part of that, what we call junk has function has valuable function. They are not genes but they regulate the function of genes. So what's junk in yesterday could be valuable today, or what's junk today could be valuable tomorrow. So there's this tension going on between you deciding not wanting to afford to store everything that you can get your hands on. But on the other hand, you know you worry, you ignore the wrong ones. You can see this tension in our customers. And then it depends on industry here. In healthcare they say, I have no choice. I want it all, why? One very insightful point brought up by one healthcare provider that really touched me was you know, we are not, we don't only care. Of course we care a lot. We care a lot about the people we are caring for. But we also care for the people we are not caring for. How do we find them? And therefore, they did not just need to collect data that they have with, from their patients they also need to reach out to outside data so that they can figure out who they are not caring for. So they want it all. So I asked them, "So what do you do with funding if you want it all?" They say they have no choice but they'll figure out a way to fund it and perhaps monetization of what they have now is the way to come around and fund that. Of course, they also come back to us, rightfully that you know, we have to then work out a way to to help them build a system. So that healthcare. And if you go to other industries like banking, they say they can afford to keep them all. But they are regulated same like healthcare. They are regulated as to privacy and such like. So many examples, different industries having different needs but different approaches to how, what they collect. But there is this constant tension between you perhaps deciding not wanting to fund all of that, all that you can store. But on the other hand you know, if you kind of don't want to afford it and decide not to store some, maybe those some become highly valuable in the future. You worry. >> Well, we can make some assumptions about the future, can't we? I mean we know there's going to be a lot more data than we've ever seen before, we know that. We know, well not withstanding supply constraints and things like NAND. We know the price of storage is going to continue to decline. We also know and not a lot of people are really talking about this but the processing power, everybody says, Moore's Law is dead. Okay, it's waning but the processing power when you combine the CPUs and NPUs, and GPUs and accelerators and so forth, actually is increasing. And so when you think about these use cases at the Edge you're going to have much more processing power. You're going to have cheaper storage and it's going to be less expensive processing. And so as an AI practitioner, what can you do with that? >> Yeah, it's a highly, again another insightful question that we touched on, on our keynote and that goes up to the why, I'll do the where. Where will your data be? We have one estimate that says that by next year, there will be 55 billion connected devices out there. 55 billion. What's the population of the world? Well, off the order of 10 billion, but this thing is 55 billion. And many of them, most of them can collect data. So what do you do? So the amount of data that's going to come in is going to way exceed our drop in storage costs our increasing compute power. So what's the answer? The answer must be knowing that we don't and even a drop in price and increase in bandwidth, it will overwhelm the 5G, it'll will overwhelm 5G, given the amount of 55 billion of them collecting. So the answer must be that there needs to be a balance between you needing to bring all that data from the 55 billion devices of the data back out to a central, as a bunch of central cost because you may not be able to afford to do that. Firstly bandwidth, even with 5G and as the, when you still be too expensive given the number of devices out there. You know given storage costs dropping it'll still be too expensive to try and install them all. So the answer must be to start at least to mitigate the problem to some leave most a lot of the data out there. And only send back the pertinent ones, as you said before. But then if you did that then, how are we going to do machine learning at the core and the Cloud side, if you don't have all the data you want rich data to train with. Sometimes you want to a mix of the positive type data, and the negative type data. So you can train the machine in a more balanced way. So the answer must be you eventually, as we move forward with these huge number of devices are at the Edge to do machine learning at the Edge. Today we don't even have power. The Edge typically is characterized by a lower energy capability and therefore, lower compute power. But soon, you know, even with low energy, they can do more with compute power, improving in energy efficiency. So learning at the Edge today we do inference at the Edge. So we data, model, deploy and you do inference at age. That's what we do today. But more and more, I believe given a massive amount of data at the Edge you have to have to start doing machine learning at the Edge. And if when you don't have enough power then you aggregate multiple devices' compute power into a swarm and learn as a swarm. >> Oh, interesting, so now of course, if I were sitting in a flyer flying the wall on HPE Board meeting I said, "Okay, HPE is a leading provider of compute." How do you take advantage that? I mean, we're going, I know it's future but you must be thinking about that and participating in those markets. I know today you are, you have, you know, Edge line and other products, but there's, it seems to me that it's not the general purpose that we've known in the past. It's a new type of specialized computing. How are you thinking about participating in that opportunity for your customers? >> The wall will have to have a balance. Where today the default, well, the more common mode is to collect the data from the Edge and train at some centralized location or number of centralized location. Going forward, given the proliferation of the Edge devices, we'll need a balance, we need both. We need capability at the Cloud side. And it has to be hybrid. And then we need capability on the Edge side. Yeah that we need to build systems that on one hand is Edge-adapted. Meaning they environmentally-adapted because the Edge differently are on it. A lot of times on the outside, they need to be packaging-adapted and also power-adapted. Because typically many of these devices are battery-powered. So you have to build systems that adapts to it. But at the same time, they must not be custom. That's my belief. They must be using standard processes and standard operating system so that they can run a rich set of applications. So yes, that's also the insightful for that. Antonio announced in 2018 for the next four years from 2018, $4 billion invested to strengthen our Edge portfolio our Edge product lines, Edge solutions. >> Dr. Goh, I could go on for hours with you. You're just such a great guest. Let's close. What are you most excited about in the future of certainly HPE, but the industry in general? >> Yeah, I think the excitement is the customers. The diversity of customers and the diversity in the way they have approached their different problems with data strategy. So the excitement is around data strategy. Just like, you know, the statement made for us was so, was profound. And Antonio said we are in the age of insight powered by data. That's the first line. The line that comes after that is as such we are becoming more and more data-centric with data the currency. Now the next step is even more profound. That is, you know, we are going as far as saying that data should not be treated as cost anymore, no. But instead, as an investment in a new asset class called data with value on our balance sheet. This is a step change in thinking that is going to change the way we look at data, the way we value it. So that's a statement. So this is the exciting thing, because for me a CTO of AI, a machine is only as intelligent as the data you feed it with. Data is a source of the machine learning to be intelligent. So that's why when the people start to value data and say that it is an investment when we collect it it is very positive for AI because an AI system gets intelligent, get more intelligence because it has huge amounts of data and a diversity of data. So it'd be great if the community values data. >> Well, are you certainly see it in the valuations of many companies these days? And I think increasingly you see it on the income statement, you know data products and people monetizing data services, and yeah, maybe eventually you'll see it in the balance sheet, I know. Doug Laney when he was at Gartner Group wrote a book about this and a lot of people are thinking about it. That's a big change, isn't it? Dr. Goh. >> Yeah, yeah, yeah. Your question is the process and methods in valuation. But I believe we'll get there. We need to get started and then we'll get there, I believe, yeah. >> Dr. Goh it's always my pleasure. >> And then the AI will benefit greatly from it. >> Oh yeah, no doubt. People will better understand how to align some of these technology investments. Dr. Goh, great to see you again. Thanks so much for coming back in theCube. It's been a real pleasure. >> Yes, a system is only as smart as the data you feed it with. (both chuckling) >> Well, excellent, we'll leave it there. Thank you for spending some time with us so keep it right there for more great interviews from HPE Discover '21. This is Dave Vellante for theCube, the leader in enterprise tech coverage. We'll be right back (upbeat music)

Published Date : Jun 10 2021

SUMMARY :

Dr. Goh, great to see you again. Great to talk to you again. and you addressed some and how to thrive in this new age. of the ones you talked about today? One of the barriers to insight And as a great example, the flash crash is that humans put in the rules to decide that it's going to be tough there. and it's something you know a lot about And the machine starts to evolve a model Is that learning from the Yeah, you don't So that you don't need to have but the rest can be done at the Edge. But on the other hand you know, And so when you think about and the Cloud side, if you I know today you are, you So you have to build about in the future as the data you feed it with. And I think increasingly you Your question is the process And then the AI will Dr. Goh, great to see you again. as the data you feed it with. Thank you for spending some time with us

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Dave VellantePERSON

0.99+

Michael LewisPERSON

0.99+

Doug LaneyPERSON

0.99+

DavePERSON

0.99+

AntonioPERSON

0.99+

2018DATE

0.99+

10 billionQUANTITY

0.99+

$4 billionQUANTITY

0.99+

second challengeQUANTITY

0.99+

55 billionQUANTITY

0.99+

two languagesQUANTITY

0.99+

two challengesQUANTITY

0.99+

May 6DATE

0.99+

Flash BoysTITLE

0.99+

two-partQUANTITY

0.99+

55 billionQUANTITY

0.99+

tomorrowDATE

0.99+

Gartner GroupORGANIZATION

0.99+

second areaQUANTITY

0.99+

TodayDATE

0.99+

last yearDATE

0.99+

less than 100 sharesQUANTITY

0.99+

hundreds of millionsQUANTITY

0.99+

first lineQUANTITY

0.99+

OneQUANTITY

0.99+

HPEORGANIZATION

0.99+

todayDATE

0.99+

second barrierQUANTITY

0.99+

May 6, 2010DATE

0.99+

10QUANTITY

0.99+

first barrierQUANTITY

0.99+

bothQUANTITY

0.99+

less than 100 shareQUANTITY

0.99+

Dr.PERSON

0.99+

one modelQUANTITY

0.99+

tens of minutesQUANTITY

0.98+

one areaQUANTITY

0.98+

one languageQUANTITY

0.98+

EdgeORGANIZATION

0.98+

three stagesQUANTITY

0.98+

yesterdayDATE

0.98+

first partQUANTITY

0.98+

one ruleQUANTITY

0.98+

GohPERSON

0.98+

FirstlyQUANTITY

0.98+

first oneQUANTITY

0.97+

United NationsORGANIZATION

0.97+

firstQUANTITY

0.97+

oneQUANTITY

0.97+

first barrierQUANTITY

0.97+

Hewlett Packard EnterpriseORGANIZATION

0.96+

about 100 questionsQUANTITY

0.96+

10 petabytes a dayQUANTITY

0.95+

Day 2QUANTITY

0.94+

Eng Lim GohPERSON

0.94+

Day 1QUANTITY

0.93+

under 100QUANTITY

0.92+

DrPERSON

0.92+

one estimateQUANTITY

0.91+

Dr Eng Lim Goh, Vice President, CTO, High Performance Computing & AI


 

(upbeat music) >> Welcome back to HPE Discover 2021, theCUBE's virtual coverage, continuous coverage of HPE's Annual Customer Event. My name is Dave Vellante, and we're going to dive into the intersection of high-performance computing, data and AI with Doctor Eng Lim Goh, who's a Senior Vice President and CTO for AI at Hewlett Packard Enterprise. Doctor Goh, great to see you again. Welcome back to theCUBE. >> Hello, Dave, great to talk to you again. >> You might remember last year we talked a lot about Swarm intelligence and how AI is evolving. Of course, you hosted the Day 2 Keynotes here at Discover. And you talked about thriving in the age of insights, and how to craft a data-centric strategy. And you addressed some of the biggest problems, I think organizations face with data. That's, you've got a, data is plentiful, but insights, they're harder to come by. >> Yeah. >> And you really dug into some great examples in retail, banking, in medicine, healthcare and media. But stepping back a little bit we zoomed out on Discover '21. What do you make of the events so far and some of your big takeaways? >> Hmm, well, we started with the insightful question, right, yeah? Data is everywhere then, but we lack the insight. That's also part of the reason why, that's a main reason why Antonio on day one focused and talked about the fact that we are in the now in the age of insight, right? And how to try thrive in that age, in this new age? What I then did on a Day 2 Keynote following Antonio is to talk about the challenges that we need to overcome in order to thrive in this new age. >> So, maybe we could talk a little bit about some of the things that you took away in terms of, I'm specifically interested in some of the barriers to achieving insights. You know customers are drowning in data. What do you hear from customers? What were your takeaway from some of the ones you talked about today? >> Oh, very pertinent question, Dave. You know the two challenges I spoke about, that we need to overcome in order to thrive in this new age. The first one is the current challenge. And that current challenge is, you know, stated is now barriers to insight, when we are awash with data. So that's a statement on how do you overcome those barriers? What are the barriers to insight when we are awash in data? In the Day 2 Keynote, I spoke about three main things. Three main areas that we receive from customers. The first one, the first barrier is in many, with many of our customers, data is siloed, all right. You know, like in a big corporation, you've got data siloed by sales, finance, engineering, manufacturing and so on supply chain and so on. And there's a major effort ongoing in many corporations to build a federation layer above all those silos so that when you build applications above, they can be more intelligent. They can have access to all the different silos of data to get better intelligence and more intelligent applications built. So that was the first barrier we spoke about, you know? Barriers to insight when we are awash with data. The second barrier is that we see amongst our customers is that data is raw and disperse when they are stored. And you know, it's tough to get at, to tough to get a value out of them, right? And in that case, I use the example of, you know, the May 6, 2010 event where the stock market dropped a trillion dollars in terms of minutes. We all know those who are financially attuned with know about this incident but that this is not the only incident. There are many of them out there. And for that particular May 6 event, you know, it took a long time to get insight. Months, yeah, before we, for months we had no insight as to what happened. Why it happened? Right, and there were many other incidences like this and the regulators were looking for that one rule that could mitigate many of these incidences. One of our customers decided to take the hard road they go with the tough data, right? Because data is raw and dispersed. So they went into all the different feeds of financial transaction information, took the tough, you know, took a tough road. And analyze that data took a long time to assemble. And they discovered that there was caught stuffing, right? That people were sending a lot of trades in and then canceling them almost immediately. You have to manipulate the market. And why didn't we see it immediately? Well, the reason is the process reports that everybody sees, the rule in there that says, all trades less than a hundred shares don't need to report in there. And so what people did was sending a lot of less than a hundred shares trades to fly under the radar to do this manipulation. So here is the second barrier, right? Data could be raw and dispersed. Sometimes it's just have to take the hard road and to get insight. And this is one great example. And then the last barrier has to do with sometimes when you start a project to get insight, to get answers and insight, you realize that all the data's around you, but you don't seem to find the right ones to get what you need. You don't seem to get the right ones, yeah? Here we have three quick examples of customers. One was a great example, right? Where they were trying to build a language translator or machine language translator between two languages, right? By not do that, they need to get hundreds of millions of word pairs. You know of one language compare with the corresponding other. Hundreds of millions of them. They say, well, I'm going to get all these word pairs. Someone creative thought of a willing source and a huge, it was a United Nations. You see? So sometimes you think you don't have the right data with you, but there might be another source and a willing one that could give you that data, right? The second one has to do with, there was the sometimes you may just have to generate that data. Interesting one, we had an autonomous car customer that collects all these data from their their cars, right? Massive amounts of data, lots of sensors, collect lots of data. And, you know, but sometimes they don't have the data they need even after collection. For example, they may have collected the data with a car in fine weather and collected the car driving on this highway in rain and also in snow. But never had the opportunity to collect the car in hill because that's a rare occurrence. So instead of waiting for a time where the car can drive in hill, they build a simulation by having the car collected in snow and simulated him. So these are some of the examples where we have customers working to overcome barriers, right? You have barriers that is associated. In fact, that data silo, they federated it. Virus associated with data, that's tough to get at. They just took the hard road, right? And sometimes thirdly, you just have to be creative to get the right data you need. >> Wow! I tell you, I have about a hundred questions based on what you just said, you know? (Dave chuckles) And as a great example, the Flash Crash. In fact, Michael Lewis, wrote about this in his book, the Flash Boys. And essentially, right, it was high frequency traders trying to front run the market and sending into small block trades (Dave chuckles) trying to get sort of front ended. So that's, and they chalked it up to a glitch. Like you said, for months, nobody really knew what it was. So technology got us into this problem. (Dave chuckles) I guess my question is can technology help us get out of the problem? And that maybe is where AI fits in? >> Yes, yes. In fact, a lot of analytics work went in to go back to the raw data that is highly dispersed from different sources, right? Assembled them to see if you can find a material trend, right? You can see lots of trends, right? Like, no, we, if humans look at things that we tend to see patterns in Clouds, right? So sometimes you need to apply statistical analysis math to be sure that what the model is seeing is real, right? And that required, well, that's one area. The second area is you know, when this, there are times when you just need to go through that tough approach to find the answer. Now, the issue comes to mind now is that humans put in the rules to decide what goes into a report that everybody sees. Now, in this case, before the change in the rules, right? But by the way, after the discovery, the authorities changed the rules and all shares, all trades of different any sizes it has to be reported. >> Right. >> Right, yeah? But the rule was applied, you know, I say earlier that shares under a hundred, trades under a hundred shares need not be reported. So, sometimes you just have to understand that reports were decided by humans and for understandable reasons. I mean, they probably didn't wanted a various reasons not to put everything in there. So that people could still read it in a reasonable amount of time. But we need to understand that rules were being put in by humans for the reports we read. And as such, there are times we just need to go back to the raw data. >> I want to ask you... >> Oh, it could be, that it's going to be tough, yeah. >> Yeah, I want to ask you a question about AI as obviously it's in your title and it's something you know a lot about but. And I'm going to make a statement, you tell me if it's on point or off point. So seems that most of the AI going on in the enterprise is modeling data science applied to, you know, troves of data. But there's also a lot of AI going on in consumer. Whether it's, you know, fingerprint technology or facial recognition or natural language processing. Well, two part question will the consumer market, as it has so often in the enterprise sort of inform us is sort of first part. And then, there'll be a shift from sort of modeling if you will to more, you mentioned the autonomous vehicles, more AI inferencing in real time, especially with the Edge. Could you help us understand that better? >> Yeah, this is a great question, right? There are three stages to just simplify. I mean, you know, it's probably more sophisticated than that. But let's just simplify that three stages, right? To building an AI system that ultimately can predict, make a prediction, right? Or to assist you in decision-making. I have an outcome. So you start with the data, massive amounts of data that you have to decide what to feed the machine with. So you feed the machine with this massive chunk of data, and the machine starts to evolve a model based on all the data it's seeing. It starts to evolve, right? To a point that using a test set of data that you have separately kept aside that you know the answer for. Then you test the model, you know? After you've trained it with all that data to see whether its prediction accuracy is high enough. And once you are satisfied with it, you then deploy the model to make the decision. And that's the inference, right? So a lot of times, depending on what we are focusing on, we in data science are, are we working hard on assembling the right data to feed the machine with? That's the data preparation organization work. And then after which you build your models you have to pick the right models for the decisions and prediction you need to make. You pick the right models. And then you start feeding the data with it. Sometimes you pick one model and a prediction isn't that robust. It is good, but then it is not consistent, right? Now what you do is you try another model. So sometimes it gets keep trying different models until you get the right kind, yeah? That gives you a good robust decision-making and prediction. Now, after which, if it's tested well, QA, you will then take that model and deploy it at the Edge. Yeah, and then at the Edge is essentially just looking at new data, applying it to the model that you have trained. And then that model will give you a prediction or a decision, right? So it is these three stages, yeah. But more and more, your question reminds me that more and more people are thinking as the Edge become more and more powerful. Can you also do learning at the Edge? >> Right. >> That's the reason why we spoke about Swarm Learning the last time. Learning at the Edge as a Swarm, right? Because maybe individually, they may not have enough power to do so. But as a Swarm, they may. >> Is that learning from the Edge or learning at the Edge? In other words, is that... >> Yes. >> Yeah. You do understand my question. >> Yes. >> Yeah. (Dave chuckles) >> That's a great question. That's a great question, right? So the quick answer is learning at the Edge, right? And also from the Edge, but the main goal, right? The goal is to learn at the Edge so that you don't have to move the data that Edge sees first back to the Cloud or the Call to do the learning. Because that would be the reason, one of the main reasons why you want to learn at the Edge. Right? So that you don't need to have to send all that data back and assemble it back from all the different Edge devices. Assemble it back to the Cloud Site to do the learning, right? Some on you can learn it and keep the data at the Edge and learn at that point, yeah. >> And then maybe only selectively send. >> Yeah. >> The autonomous vehicle, example you gave is great. 'Cause maybe they're, you know, there may be only persisting. They're not persisting data that is an inclement weather, or when a deer runs across the front. And then maybe they do that and then they send that smaller data setback and maybe that's where it's modeling done but the rest can be done at the Edge. It's a new world that's coming through. Let me ask you a question. Is there a limit to what data should be collected and how it should be collected? >> That's a great question again, yeah. Well, today full of these insightful questions. (Dr. Eng chuckles) That actually touches on the the second challenge, right? How do we, in order to thrive in this new age of insight? The second challenge is our future challenge, right? What do we do for our future? And in there is the statement we make is we have to focus on collecting data strategically for the future of our enterprise. And within that, I talked about what to collect, right? When to organize it when you collect? And then where will your data be going forward that you are collecting from? So what, when, and where? For what data to collect? That was the question you asked, it's a question that different industries have to ask themselves because it will vary, right? Let me give you the, you use the autonomous car example. Let me use that. And we do have this customer collecting massive amounts of data. You know, we're talking about 10 petabytes a day from a fleet of their cars. And these are not production autonomous cars, right? These are training autonomous cars, collecting data so they can train and eventually deploy commercial cars, right? Also this data collection cars, they collect 10, as a fleet of them collect 10 petabytes a day. And then when they came to us, building a storage system you know, to store all of that data, they realized they don't want to afford to store all of it. Now here comes the dilemma, right? What should I, after I spent so much effort building all this cars and sensors and collecting data, I've now decide what to delete. That's a dilemma, right? Now in working with them on this process of trimming down what they collected, you know, I'm constantly reminded of the 60s and 70s, right? To remind myself 60s and 70s, we called a large part of our DNA, junk DNA. >> Yeah. (Dave chuckles) >> Ah! Today, we realized that a large part of that what we call junk has function as valuable function. They are not genes but they regulate the function of genes. You know? So what's junk in yesterday could be valuable today. Or what's junk today could be valuable tomorrow, right? So, there's this tension going on, right? Between you deciding not wanting to afford to store everything that you can get your hands on. But on the other hand, you worry, you ignore the wrong ones, right? You can see this tension in our customers, right? And then it depends on industry here, right? In healthcare they say, I have no choice. I want it all, right? Oh, one very insightful point brought up by one healthcare provider that really touched me was you know, we don't only care. Of course we care a lot. We care a lot about the people we are caring for, right? But who also care for the people we are not caring for? How do we find them? >> Uh-huh. >> Right, and that definitely, they did not just need to collect data that they have with from their patients. They also need to reach out, right? To outside data so that they can figure out who they are not caring for, right? So they want it all. So I asked them, so what do you do with funding if you want it all? They say they have no choice but to figure out a way to fund it and perhaps monetization of what they have now is the way to come around and fund that. Of course, they also come back to us rightfully, that you know we have to then work out a way to help them build a system, you know? So that's healthcare, right? And if you go to other industries like banking, they say they can afford to keep them all. >> Yeah. >> But they are regulated, seemed like healthcare, they are regulated as to privacy and such like. So many examples different industries having different needs but different approaches to what they collect. But there is this constant tension between you perhaps deciding not wanting to fund all of that, all that you can install, right? But on the other hand, you know if you kind of don't want to afford it and decide not to start some. Maybe those some become highly valuable in the future, right? (Dr. Eng chuckles) You worry. >> Well, we can make some assumptions about the future. Can't we? I mean, we know there's going to be a lot more data than we've ever seen before. We know that. We know, well, not withstanding supply constraints and things like NAND. We know the prices of storage is going to continue to decline. We also know and not a lot of people are really talking about this, but the processing power, but the says, Moore's law is dead. Okay, it's waning, but the processing power when you combine the CPUs and NPUs, and GPUs and accelerators and so forth actually is increasing. And so when you think about these use cases at the Edge you're going to have much more processing power. You're going to have cheaper storage and it's going to be less expensive processing. And so as an AI practitioner, what can you do with that? >> Yeah, it's a highly, again, another insightful question that we touched on our Keynote. And that goes up to the why, uh, to the where? Where will your data be? Right? We have one estimate that says that by next year there will be 55 billion connected devices out there, right? 55 billion, right? What's the population of the world? Well, of the other 10 billion? But this thing is 55 billion. (Dave chuckles) Right? And many of them, most of them can collect data. So what do you do? Right? So the amount of data that's going to come in, it's going to way exceed, right? Drop in storage costs are increasing compute power. >> Right. >> Right. So what's the answer, right? So the answer must be knowing that we don't, and even a drop in price and increase in bandwidth, it will overwhelm the, 5G, it will overwhelm 5G, right? Given the amount of 55 billion of them collecting. So the answer must be that there needs to be a balance between you needing to bring all of that data from the 55 billion devices of the data back to a central, as a bunch of central cost. Because you may not be able to afford to do that. Firstly bandwidth, even with 5G and as the, when you'll still be too expensive given the number of devices out there. You know given storage costs dropping is still be too expensive to try and install them all. So the answer must be to start, at least to mitigate from to, some leave most a lot of the data out there, right? And only send back the pertinent ones, as you said before. But then if you did that then how are we going to do machine learning at the Core and the Cloud Site, if you don't have all the data? You want rich data to train with, right? Sometimes you want to mix up the positive type data and the negative type data. So you can train the machine in a more balanced way. So the answer must be eventually, right? As we move forward with these huge number of devices all at the Edge to do machine learning at the Edge. Today we don't even have power, right? The Edge typically is characterized by a lower energy capability and therefore lower compute power. But soon, you know? Even with low energy, they can do more with compute power improving in energy efficiency, right? So learning at the Edge, today we do inference at the Edge. So we data, model, deploy and you do inference there is. That's what we do today. But more and more, I believe given a massive amount of data at the Edge, you have to start doing machine learning at the Edge. And when you don't have enough power then you aggregate multiple devices, compute power into a Swarm and learn as a Swarm, yeah. >> Oh, interesting. So now of course, if I were sitting and fly on the wall and the HPE board meeting I said, okay, HPE is a leading provider of compute. How do you take advantage of that? I mean, we're going, I know it's future but you must be thinking about that and participating in those markets. I know today you are, you have, you know, Edge line and other products. But there's, it seems to me that it's not the general purpose that we've known in the past. It's a new type of specialized computing. How are you thinking about participating in that opportunity for the customers? >> Hmm, the wall will have to have a balance, right? Where today the default, well, the more common mode is to collect the data from the Edge and train at some centralized location or number of centralized location. Going forward, given the proliferation of the Edge devices, we'll need a balance, we need both. We need capability at the Cloud Site, right? And it has to be hybrid. And then we need capability on the Edge side that we need to build systems that on one hand is an Edge adapter, right? Meaning they environmentally adapted because the Edge differently are on it, a lot of times on the outside. They need to be packaging adapted and also power adapted, right? Because typically many of these devices are battery powered. Right? So you have to build systems that adapts to it. But at the same time, they must not be custom. That's my belief. It must be using standard processes and standard operating system so that they can run a rich set of applications. So yes, that's also the insight for that Antonio announced in 2018. For the next four years from 2018, right? $4 billion invested to strengthen our Edge portfolio. >> Uh-huh. >> Edge product lines. >> Right. >> Uh-huh, Edge solutions. >> I could, Doctor Goh, I could go on for hours with you. You're just such a great guest. Let's close. What are you most excited about in the future of, certainly HPE, but the industry in general? >> Yeah, I think the excitement is the customers, right? The diversity of customers and the diversity in the way they have approached different problems of data strategy. So the excitement is around data strategy, right? Just like, you know, the statement made for us was so was profound, right? And Antonio said, we are in the age of insight powered by data. That's the first line, right? The line that comes after that is as such we are becoming more and more data centric with data that currency. Now the next step is even more profound. That is, you know, we are going as far as saying that, you know, data should not be treated as cost anymore. No, right? But instead as an investment in a new asset class called data with value on our balance sheet. This is a step change, right? Right, in thinking that is going to change the way we look at data, the way we value it. So that's a statement. (Dr. Eng chuckles) This is the exciting thing, because for me a CTO of AI, right? A machine is only as intelligent as the data you feed it with. Data is a source of the machine learning to be intelligent. Right? (Dr. Eng chuckles) So, that's why when the people start to value data, right? And say that it is an investment when we collect it it is very positive for AI. Because an AI system gets intelligent, get more intelligence because it has huge amounts of data and a diversity of data. >> Yeah. >> So it'd be great, if the community values data. >> Well, you certainly see it in the valuations of many companies these days. And I think increasingly you see it on the income statement. You know data products and people monetizing data services. And yeah, maybe eventually you'll see it in the balance sheet. I know Doug Laney, when he was at Gartner Group, wrote a book about this and a lot of people are thinking about it. That's a big change, isn't it? >> Yeah, yeah. >> Dr. Goh... (Dave chuckles) >> The question is the process and methods in valuation. Right? >> Yeah, right. >> But I believe we will get there. We need to get started. And then we'll get there. I believe, yeah. >> Doctor Goh, it's always my pleasure. >> And then the AI will benefit greatly from it. >> Oh, yeah, no doubt. People will better understand how to align, you know some of these technology investments. Dr. Goh, great to see you again. Thanks so much for coming back in theCUBE. It's been a real pleasure. >> Yes, a system is only as smart as the data you feed it with. (Dave chuckles) (Dr. Eng laughs) >> Excellent. We'll leave it there. Thank you for spending some time with us and keep it right there for more great interviews from HPE Discover 21. This is Dave Vellante for theCUBE, the leader in Enterprise Tech Coverage. We'll be right back. (upbeat music)

Published Date : Jun 8 2021

SUMMARY :

Doctor Goh, great to see you again. great to talk to you again. And you talked about thriving And you really dug in the age of insight, right? of the ones you talked about today? to get what you need. And as a great example, the Flash Crash. is that humans put in the rules to decide But the rule was applied, you know, that it's going to be tough, yeah. So seems that most of the AI and the machine starts to evolve a model they may not have enough power to do so. Is that learning from the Edge You do understand my question. or the Call to do the learning. but the rest can be done at the Edge. When to organize it when you collect? But on the other hand, to help them build a system, you know? all that you can install, right? And so when you think about So what do you do? of the data back to a central, in that opportunity for the customers? And it has to be hybrid. about in the future of, as the data you feed it with. if the community values data. And I think increasingly you The question is the process We need to get started. And then the AI will Dr. Goh, great to see you again. as smart as the data Thank you for spending some time with us

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Dave VellantePERSON

0.99+

Michael LewisPERSON

0.99+

Doug LaneyPERSON

0.99+

DavePERSON

0.99+

2018DATE

0.99+

$4 billionQUANTITY

0.99+

AntonioPERSON

0.99+

two languagesQUANTITY

0.99+

10 billionQUANTITY

0.99+

55 billionQUANTITY

0.99+

two challengesQUANTITY

0.99+

second challengeQUANTITY

0.99+

55 billionQUANTITY

0.99+

HPEORGANIZATION

0.99+

last yearDATE

0.99+

Gartner GroupORGANIZATION

0.99+

first lineQUANTITY

0.99+

10QUANTITY

0.99+

second areaQUANTITY

0.99+

bothQUANTITY

0.99+

tomorrowDATE

0.99+

Hundreds of millionsQUANTITY

0.99+

TodayDATE

0.99+

todayDATE

0.99+

second barrierQUANTITY

0.99+

two partQUANTITY

0.99+

May 6, 2010DATE

0.99+

OneQUANTITY

0.99+

EdgeORGANIZATION

0.99+

first barrierQUANTITY

0.99+

less than a hundred sharesQUANTITY

0.99+

next yearDATE

0.98+

EngPERSON

0.98+

yesterdayDATE

0.98+

first partQUANTITY

0.98+

May 6DATE

0.98+

United NationsORGANIZATION

0.98+

theCUBEORGANIZATION

0.98+

one areaQUANTITY

0.98+

one modelQUANTITY

0.98+

first oneQUANTITY

0.98+

Hewlett Packard EnterpriseORGANIZATION

0.98+

Dr.PERSON

0.97+

less than a hundred sharesQUANTITY

0.97+

three stagesQUANTITY

0.97+

one ruleQUANTITY

0.97+

Three main areasQUANTITY

0.97+

Flash BoysTITLE

0.97+

one languageQUANTITY

0.97+

oneQUANTITY

0.96+

10 petabytes a dayQUANTITY

0.96+

Flash CrashTITLE

0.95+

under a hundredQUANTITY

0.95+

FirstlyQUANTITY

0.95+

Fernando Brandao, AWS & Richard Moulds, AWS Quantum Computing | AWS re:Invent 2020


 

>>From around the globe. It's the cube with digital coverage of AWS reinvent 2020, sponsored by Intel and AWS. >>Welcome back to the queue. It's virtual coverage of Avis reinvent 2020 I'm John furry, your host. Um, this is a cute virtual we're here. Not in, in remote. We're not in person this year, so we're doing the remote interviews. And then this segment is going to build on the quantum conversation we had last year, Richard moles, general manager of Amazon bracket and aid was quantum computing and Fernando Brandao head of quantum algorithms at AWS and Brent professor of theoretical physics at Caltech. Fernando, thanks for coming on, Richard. Thanks for joining us. >>You're welcome to be here. >>So, Fernando, first of all, love your title, quantum algorithms. That's the coolest title I've heard so far and you're pretty smart because you're a theoretical professor of physics at Caltech. So, um, which I'd never be able to get into, but I wish I could get into there someday, but, uh, thanks for coming on. Um, quantum has been quite the rage and you know, there's a lot of people talking about it. Um, it's not ready for prime time. Some say it's moving faster than others, but where are we on quantum right now? What are, what are you, what are you seeing Fernanda where the quantum, where are peg us in the evolution of, of, uh, where we are? >>Um, yeah, what quantum, uh, it's an emerging and rapidly developing fields. Uh, but we are see where are you on, uh, both in terms of, uh, hardware development and in terms of identifying the most impactful use cases of one company. Uh, so, so it's, it's, it's early days for everyone and, and we have like, uh, different players and different technologies that are being sport. And I think it's, it's, it's early, but it's exciting time to be doing quantum computing. And, uh, and it's very interesting to see the interest in industry growing and, and customers. Uh, for example, Casa from AWS, uh, being, uh, being willing to take part in this journey with us in developmental technology. >>Awesome. Richard, last year we talked to bill Vass about this and he was, you know, he set expectations really well, I thought, but it was pretty much in classic Amazonian way. You know, it makes the announcement a lot of progress then makes me give us the update on your end. You guys now are shipping brackets available. What's the update on your end and Verner mentioned in his keynote this week >> as well. Yeah, it was a, it was great until I was really looking at your interview with bill. It was, uh, that was when we launched the launch the service a year ago, almost exactly a year ago this week. And we've come a long way. So as you mentioned, we've, uh, we've, uh, we've gone to general availability with the service now that that happened in August. So now a customer can kind of look into the, uh, to the bracket console and, uh, installed programming concept computers. You know, there's, uh, there's tremendous excitement obviously, as, as you mentioned, and Fernando mentioned, you know, quantum computers, uh, we think >>Have the potential to solve problems that are currently, uh, uh, unsolvable. Um, the goal of bracket is to fundamentally give customers the ability to, uh, to go test, uh, some of those notions to explore the technology and to just start planning for the future. You know, our goal was always to try and solve some of the problems that customers have had for, you know, gee, a decade or so now, you know, they tell us from a variety of different industries, whether it's drug discovery or financial services, whether it's energy or there's chemical engineering, machine learning, you know, th the potential for quantum computer impacts may industries could potentially be disruptive to those industries. And, uh, it's, it's essential that customers can can plan for the future, you know, build their own internal resources, become experts, hire the right staff, figure out where it might impact their business and, uh, and potentially disrupt. >>So, uh, you know, in the past they're finding it hard to, to get involved. You know, these machines are very different, different technologies building in different ways of different characteristics. Uh, the tooling is very disparate, very fragmented. Historically, it's hard for companies to get access to the machines. These tend to be, you know, owned by startups or in, you know, physics labs or universities, very difficult to get access to these things, very different commercial models. Um, and, uh, as you, as you suggested, a lot of interests, a lot of hype, a lot of claims in the industry, customers want to cut through all that. They want to understand what's real, uh, what they can do today, uh, how they can experiment and, uh, and get started. So, you know, we see bracket as a catalyst for innovation. We want to bring together end-users, um, consultants, uh, software developers, um, providers that want to host services on top of bracket, try and get the industry, you know, rubbing along them. You spoke to lots of Amazonians. I'm sure you've heard the phrase innovation flywheel, plenty of times. Um, we see the same approach that we've used successfully in IOT and robotics and machine learning and apply that same approach to content, machine learning software, to quantum computing, and to learn, to bring it together. And, uh, if we get the tooling right, and we make it easy, um, then we don't see any reason why we can't, uh, you know, rapidly try and move this industry forward. And >>It was fun areas where there's a lot of, you know, intellectual computer science, um, technology science involved in super exciting. And Amazon's supposed to some of that undifferentiated heavy. >>That's what I am, you know, it's like, >>There's a Maslow hierarchy of needs in the tech industry. You know, people say, Oh, why five people freak out when there's no wifi? You know, you can't get enough compute. Right. So, you know, um, compute is one of those things with machine learning is seeing the benefits and quantum there's so much benefits there. Um, and you guys made some announcements at, at re-invent, uh, around BRACA. Can you share just quickly share some of those updates, Richard? >>Sure. I mean, it's the way we innovate at AWS. You know, we, we start simple and we, and we build up features. We listen to customers and we learn as we go along, we try and move as quickly as possible. So since going public in, uh, in, in August, we've actually had a string of releases, uh, pretty consistent, um, delivering new features. So we try to tie not the integration with the platform. Customers have told us really very early on that they, they don't just want to play with the technology. They want to figure out how to, how to envisage a production quantum computing service, how it might look, you know, in the context of a broad cloud platform with AWS. So we've, uh, we launched some integration with, uh, other AWS capabilities around security, managing limits, quotas, tagging resources, that type of thing, things that are familiar to, uh, to, to, to current AWS users. >>Uh, we launched some new hardware. Uh, all of our partners D-Wave launched some, uh, uh, you know, a 5,000 cubit machine, uh, just in September. Uh, so we made that available on bracket the same day that they launched that hardware, which was very cool. Um, you know, we've made it, uh, we've, we've made it easier for researchers. We've been, you know, impressed how many academics and researchers have used the service, not just large corporations. Um, they want to have really deep access to these machines. They want to program these things at a low level. So we launched some features, uh, to enable them to do their research, but reinvent, we were really focused on two things, um, simulators and making it much easier to use, uh, hybrid systems systems that, uh, incorporate classical compute, traditional digital computing with quantum machinery, um, in the vein that follow some of the liens that we've seen, uh, in machine learning. >>So, uh, simulators are important. They're a very important part of, uh, learning how to use concepts, computers. They're always available 24, seven they're super convenient to use. And of course they're critical in verifying the accuracy of the results that we get from quantum hardware. When we launched the service behind free simulator for customers to help debug their circuits and experiments quickly, um, but simulating large experiments and large systems is a real challenge on classical computers. You know, it, wasn't hard on classical. Uh, then you wouldn't need a quantum computer. That's the whole point. So running large simulations, you know, is expensive in terms of resources. It's complicated. Uh, we launched a pretty powerful simulator, uh, back in August, which we thought at the time was always powerful managed. Quantum stimulates circuit handled 34 cubits, and it reinvented last week, we launched a new simulator, which actually the first managed simulator to use tensor network technology. >>And it can run up to 50 cubits. So we think is, we think is probably the most powerful, uh, managed quantum simulator on the market today. And customers can flip easily between either using real quantum hardware or either of our, uh, stimulators just by changing a line of code. Um, the other thing we launched was the ability to run these hybrid systems. You know, quantum computers will get more, no don't get onto in a moment is, uh, today's computers are very imperfect, you know, lots of errors. Um, we working, obviously the industry towards fault-tolerant machines and Fernando can talk about some research papers that were published in that area, but right now the machines are far from perfect. And, uh, and the way that we can try to squeeze as much value out of these devices today is to run them in tandem with classical systems. >>We think of the notion of a self-learning quantum algorithm, where you use a classical optimization techniques, such as we see machine learning to tweak and tune the parameters of a quantum algorithm to try and iterate and converge on the best answer and try and overcome some of these issues surrounding errors. That's a lot of moving parts to orchestrate for customers, a lot of different systems, a lot of different programming techniques. And we wanted to make that much easier. We've been impressed with a, a, an open projects, been around for a couple of years, uh, called penny lane after the Beatles song. And, um, so we wanted to double down on that. We were getting a lot of positive feedback from customers about the penny lane talk it, so we decided to, uh, uh, make it a first class citizen on bracket, make it available as a native feature, uh, in our, uh, in our Jupiter notebooks and our tutorials learning examples, um, that open source project has very similar, um, guiding principles that we do, you know, it's open, it's cross platform, it's technology agnostic, and we thought he was a great fit to the service. >>So we, uh, we announced that and made it available to customers and, uh, and, and, uh, already getting great feedback. So, uh, you know, finishing the finishing the year strongly, I think, um, looking forward to 2021, you know, looking forward to some really cool technology it's on the horizon, uh, from a hardware point of view, making it easy to use, um, you know, and always, obviously trying to work back from customer problems. And so congratulations on the success. I'm sure it's not hard to hire people interested, at least finding qualified people it'd be different, but, you know, sign me up. I love quantum great people, Fernando real quick, understanding the relationship with Caltech unique to Amazon. Um, tell us how that fits into the, into this, >>Uh, right. John S no, as I was saying, it's it's early days, uh, for, for quantum computing, uh, and to make progress, uh, in abreast, uh, put together a team of experts, right. To work both on, on find new use cases of quantum computing and also, uh, building more powerful, uh, quantum hardware. Uh, so the AWS center for quantum computing is based at Caltech. Uh, and, and this comes from the belief of AWS that, uh, in quantum computing is key to, uh, to keep close, to stay close of like fresh ideas and to the latest scientific developments. Right. And Caltech is if you're near one computing. So what's the ideal place for doing that? Uh, so in the center, we, we put together researchers and engineers, uh, from computer science, physics, and other subjects, uh, from Amazon, but also from all the academic institutions, uh, of course some context, but we also have Stanford and university of Chicago, uh, among others. So we broke wrongs, uh, in the beauty for AWS and for quantum computer in the summer, uh, and under construction right now. Uh, but, uh, as we speak, John, the team is busy, uh, uh, you know, getting stuff in, in temporary lab space that we have at cottage. >>Awesome. Great. And real quick, I know we've got some time pressure here, but you published some new research, give a quick a plug for the new research. Tell us about that. >>Um, right. So, so, you know, as part of the effort or the integration for one company, uh, we are developing a new cubix, uh, which we choose a combination of acoustic and electric components. So this kind of hybrid Aquacel execute, it has the promise for a much smaller footprint, think about like a few microliters and much longer storage times, like up to settlements, uh, which, which is a big improvement over the scale of the arts sort of writing all export based cubits, but that's not the whole story, right? On six, if you have a good security should make good use of it. Uh, so what we did in this paper, they were just put out, uh, is, is a proposal for an architecture of how to build a scalable quantum computer using these cubits. So we found from our analysis that we can get more than a 10 X overheads in the resources required from URI, a universal thought around quantum computer. >>Uh, so what are these resources? This is like a smaller number of physical cubits. Uh, this is a smaller footprint is, uh, fewer control lines in like a smaller approach and a consistent, right. And, and these are all like, uh, I think this is a solid contribution. Uh, no, it's a theoretical analysis, right? So, so the, uh, the experimental development has to come, but I think this is a solid contribution in the big challenge of scaling up this quantum systems. Uh, so, so, so John, as we speak like, uh, data blessed in the, for quantum computing is, uh, working on the experimental development of this, uh, a highly adequacy architecture, but we also keep exploring other promising ways of doing scalable quantum computers and eventually, uh, to bring a more powerful computer resources to AWS customers. >>It's kind of like machine learning and data science, the smartest people work on it. Then you democratize that. I can see where this is going. Um, Richard real quick, um, for people who want to get involved and participate or consume, what do they do? Give us the playbook real quick. Uh, so simple, just go to the AWS console and kind of log onto the, to the bracket, uh, bracket console, jump in, you know, uh, create, um, create a Jupiter notebook, pull down some of our sample, uh, applications run through the notebook and program a quantum computer. It's literally that simple. There's plenty of tutorials. It's easy to get started, you know, classic cloud style right now from commitment. Jump in, start simple, get going. We want you to go quantum. You can't go back, go quantum. You can't go back to regular computing. I think people will be running concert classical systems in parallel for quite some time. So yeah, this is the, this is definitely not a one way door. You know, you go explore quantum computing and see how it fits into, uh, >>You know, into the, into solving some of the problems that you wanted to solve in the future. But definitely this is not a replacement technology. This is a complimentary technology. >>It's great. It's a great innovation. It's kind of intoxicating technically to get, think about the benefits Fernando, Richard, thanks for coming on. It's really exciting. I'm looking forward to keeping up keeping track of the progress. Thanks for coming on the cube coverage of reinvent, quantum computing going the next level coexisting building on top of the shoulders of other giant technologies. This is where the computing wave is going. It's different. It's impacting people's lives. This is the cube coverage of re-invent. Thanks for watching.

Published Date : Dec 16 2020

SUMMARY :

It's the cube with digital coverage of AWS And then this segment is going to build on the quantum conversation we had last Um, quantum has been quite the rage and you know, Uh, but we are see where are you on, uh, both in terms of, uh, hardware development and Richard, last year we talked to bill Vass about this and he was, you know, he set expectations really well, there's, uh, there's tremendous excitement obviously, as, as you mentioned, and Fernando mentioned, Have the potential to solve problems that are currently, uh, uh, unsolvable. So, uh, you know, in the past they're finding it hard to, to get involved. It was fun areas where there's a lot of, you know, intellectual computer science, So, you know, um, compute is one of those things how it might look, you know, in the context of a broad cloud platform with AWS. uh, uh, you know, a 5,000 cubit machine, uh, just in September. So running large simulations, you know, is expensive in terms of resources. And, uh, and the way that we can try to you know, it's open, it's cross platform, it's technology agnostic, and we thought he was a great fit to So, uh, you know, finishing the finishing the year strongly, but also from all the academic institutions, uh, of course some context, but we also have Stanford And real quick, I know we've got some time pressure here, but you published some new research, uh, we are developing a new cubix, uh, which we choose a combination of acoustic So, so the, uh, the experimental development has to come, to the bracket, uh, bracket console, jump in, you know, uh, create, You know, into the, into solving some of the problems that you wanted to solve in the future. It's kind of intoxicating technically to get, think about the benefits Fernando,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Fernando BrandaoPERSON

0.99+

AWSORGANIZATION

0.99+

RichardPERSON

0.99+

CaltechORGANIZATION

0.99+

AmazonORGANIZATION

0.99+

Richard MouldsPERSON

0.99+

SeptemberDATE

0.99+

John SPERSON

0.99+

JohnPERSON

0.99+

FernandoPERSON

0.99+

BrentPERSON

0.99+

AugustDATE

0.99+

last weekDATE

0.99+

VernerPERSON

0.99+

2021DATE

0.99+

StanfordORGANIZATION

0.99+

sixQUANTITY

0.99+

last yearDATE

0.99+

last yearDATE

0.99+

34 cubitsQUANTITY

0.99+

a year agoDATE

0.99+

firstQUANTITY

0.99+

five peopleQUANTITY

0.99+

IntelORGANIZATION

0.99+

FernandaPERSON

0.98+

5,000 cubitQUANTITY

0.98+

todayDATE

0.98+

two thingsQUANTITY

0.98+

bothQUANTITY

0.97+

oneQUANTITY

0.97+

this weekDATE

0.96+

sevenQUANTITY

0.96+

D-WaveORGANIZATION

0.95+

Richard molesPERSON

0.95+

this yearDATE

0.95+

bill VassPERSON

0.94+

up to 50 cubitsQUANTITY

0.94+

24QUANTITY

0.93+

one wayQUANTITY

0.93+

a year ago this weekDATE

0.89+

AquacelORGANIZATION

0.89+

Avis reinvent 2020TITLE

0.88+

one companyQUANTITY

0.87+

BeatlesORGANIZATION

0.86+

AWS Quantum ComputingORGANIZATION

0.8+

BRACALOCATION

0.76+

a decadeQUANTITY

0.76+

computingEVENT

0.75+

couple of yearsQUANTITY

0.75+

10 XQUANTITY

0.74+

more thanQUANTITY

0.73+

re:Invent 2020TITLE

0.62+

playbookCOMMERCIAL_ITEM

0.62+

JupiterORGANIZATION

0.6+

waveEVENT

0.55+

ChicagoLOCATION

0.54+

MaslowORGANIZATION

0.52+

pennyTITLE

0.49+

Towards Understanding the Fundamental Limits of Analog, Continuous Time Computing


 

>> Hello everyone. My name is Zoltan Toroczkai. I am from University of Notre Dame, Physics department, and I'd like to thank the organizers for their kind invitation to participate in this very interesting and promising workshop. Also like to say that I look forward to collaborations with the Redefine Lab and Yoshian collaborators on the topics of this work. So today I'll briefly talk about, our attempt to understand, the fundamental limits of analog, continuous-time computing at least from the point of view of Boolean Satisfiability problem-solving using ordinary differential equations. But I think the issues that we raise during this occasion actually apply to other approaches, analog approaches as well, until to other problems as well. I think everyone here, knows what Boolean Satisfiability problems are. You have N Boolean variables, you have M clauses. Each a disjunction of K literals. Literal is a variable or it's negation. And the goal is to find an assignment to the variable such that all the clauses are true. This is a decision type problem from the NP class, which means you can check in polynomial time for satisfiability of any assignment. And the 3-SAT is NP-complete with K, 3 or larger, which means an efficient 3-SAT solver, (clears throat) implies an efficient solver for all the problems in the NP clause because all the problems in the NP clause can be reduced in polynomial time to 3-SAT. As a matter of fact you can, reduce the NP-complete problems into each other. You can go from 3-SAT to Set Packing or to Maximum Independent Set which is the set packing in graph theoretic notions or terms, to the ising graph SAT problem decision version. This is useful when you are comparing different approaches or working on different kinds of problems. When not all the clauses can be satisfied, you're looking at the optimization version of SAT, called Max-SAT and the goal here is to find the assignment that satisfies the maximum number of clauses, and this is from the NP-hard class. In terms of applications, if we had an efficient SAT solver, or NP-complete problem solver, it would literally, positively influence thousands of problems in applications in industry and science. I'm not going to read this. But this of course gives us some motivation, to work on these kind of problems. Now, our approach to SAT solving, involves embedding the problem in a continuous space, and you use all these to do that. So instead of working zeros and ones, we work with minus one and plus ones, and if we allow the corresponding variables, to change continuously between the two bounds, we formulate the problem with the help of a Clause Matrix. If, if a clause does not contain a variable or its negation, the corresponding matrix element is zero. If it contains the variable in positive form it's one. If it contains the variable in negated form, it's negative one. And now we use this to formulate these products, called clause violation functions, one for every clause, which rarely continues between zero and one and beyond zero if and only if the clause itself is true. Then we form... We define, also define the dynamics, search dynamics in this and the M-dimensional hypercube, where the search happens and if there exists solutions they're sitting in some of the corners of this hypercube. So we define this energy, potential or landscape function as shown here in a way that it, this is zero if and only if all the clauses, all the Kms are zero. All the clauses are satisfied, keeping these auxiliary variables, Ams always positive. And therefore what we do here is a dynamics that is essentially a gradient descent on this potential energy landscape. If you are to keep all the Ams constant then it would get stuck in some local minimum. However what do you do here is, we couple it with the dynamics. We couple it with the clause violation functions as shown here. And if you didn't have these Am here, just had just the Kms, for example, you have essentially, both case you have a positive feedback. You have a decreasing variable, but in that case you'll still get stuck, would still behave... We'll still find solutions better than the constant version or still would get stuck. Only when we put here this Am, which makes them dynamics in this variable exponential like, only then it keeps searching until it finds a solution. And there's a reason for that, that I'm not going to talk about here, but essentially boils down to performing a gradient descent on a globally time-varying landscape. And, and, and this is what works. Now, I'm going to talk about the good or bad, and maybe the ugly. This is, this is... What's good is that it's a hyperbolic dynamical system, which means that if you take any domain in the search space that doesn't have a solution in it or any solution in it, then the number of trajectories in it, the case exponentially quickly and the decay rate is a characteristic, invariant characteristic of the dynamics itself with the dynamical systems called the escape rate. The inverse of that is the timescale in which you find solutions by this dynamical system. And you can see here some trajectories, they are curved because it's, it's not linear but it's transiently curved to give, if there are solutions of course, we could see eventually, it does lead to the solutions. Now, in terms of performance, here what you show, for a bunch of, constraint densities, defined by, M over N, the ratio between clauses to variables, for random SAT problems, is random 3-SAT problems. And they, they, as, as function of N, and we look at, monitor the wall time, the wall clock time, and it, it behaves quite well, it behaves as a, as a polynomialy, until you actually hit, or reach the set on set transition, where the hardest problems are found. But what's more interesting is if you monitor the continuous-time t, the performance in terms of the analog continuous-time t, because that seems to be a polynomial. And the way we show that, is we can see the random K-SAT or random 3-SAT for a fixed constraint density. And we here, what you show here is at the, right at the threshold where it's really hard. And, (clears throat) we monitor the fraction of problems that we have not been able to solve it. We select thousands of problems at that cost rate ratio and we solve them with our algorithm, and we monitor the fraction of problems that have not yet been solved by continuous-time t. And these, as you see these decays exponentially in different decay rates for different system sizes and in this spot shows that this decay rate behaves polynomialy. or actually as a power law. So if you combine these two, you find that the time needed to solve all problems, except maybe appeared fraction of them, scales polynomialy with problem size. So you have polynomial continuous-time complexity. And this is also true, for other types of very hard constraints of the SAT problem such as exact color, because you can always transform them into 3-SAT as we discussed before, Ramsay coloring and, and on these problems, even algorithms like a survey propagation wheel will fail. But this doesn't mean that P equals NP because what you have, first of all, if you were to implement these equations in a device, whose behavior is described by these ODEs, then of course, t the continuous-time variable, becomes a physical wall clock time. And that would be polynomialy scaling but you have other variables, auxiliary variables, which fluctuate in an exponential manner. So if they represent currents or voltages in your realization and it would be an exponential cost algorithm. But this is some kind of trade between time and energy while I know how to generate energy or I don't know how to generate time but I know how to generate energy so it could be useful. But there's other issues as well, especially if you're trying to do this on a digital machine, but also happens, problems happen, appear, other problems appear on in physical devices as well as we discuss later. So if you implement these in GPU, you can, then you can get an order of two magnitude speedup, and you can also modify this, to solve Max-SAT problems quite efficiently, we are competitive with the best, heuristics solvers, this is all the problems in 2016, Max-SAT competition. So, so this, this, this is definitely, this is like a good approach, but there's of course, interesting limitations, I would say interesting, because it kind of makes you think about what it needs and how you can explore this, these observations in understanding better analog continuous-time complexity. If you monitor the discrete number, the number of discrete steps, done by the Runge Kutta integrator, and you solve this on a digital machine. You're using some kind of integrator, and, you know, using the same approach, but now you measure the number of problems you haven't solved, by a given number of discrete steps taken by the integrator. You find out, you have exponential discrete-time complexity. And of course, this is a problem. And if you look closely, what happens, even though the analog mathematical trajectory, that's the red curve here, if you monitor what happens in discrete time, the integrator fluctuates very little. So this is like you know, third or four digits precision, but fluctuates like crazy. So it really is like the integration freezes out, and this is because of the phenomenon of stiffness that I'll talk a little bit, more about a little bit later. So you know, it may look like an integration issue on your digital machines that you could improve and you could definitely improve, but actually the issue is bigger than that. It's, it's deeper than that because on a digital machine there is no time energy conversion. So the auxiliary variables are efficiently represented in a digital machine, so there's no exponential fluctuating current or voltage in your computer when you do this. So if e is not equal NP, then the exponential time complexity or exponential cost complexity has to hit you somewhere. And this is how. But you know one would be tempted to think maybe, this wouldn't be an issue in a analog device, and to some extent is true. Analog devices can be orders of magnitude faster, but they also suffer from their own problems because P not equal NP affects that clause of followers as well. So, indeed if you look at other systems, like Coherent Ising Machine with Measurement-Feedback, or Polariton Condensate Graphs or Oscillator Networks, they all hinge on some kind of, our ability to control real variables with arbitrarily high precision, and Oscillator Networks, you want to read out arbitrarily close frequencies. In case of CIMs, we require identical analog amplitudes which is hard to keep and they kind of fluctuate away from one another, shift away from one another, And, and if you control that, of course, then you can control the performance. So, actually one can ask if whether or not this is a universal bottleneck, and it seems so, as I will argue next. We can recall a fundamental result by A. Schönhage, Graham Schönhage from 1978 who says that, it's a purely computer science proof, that, "If you are able to compute, "the addition, multiplication, division "of real variables with infinite precision then, "you could solve NP-complete problems in polynomial time." He doesn't actually propose a solid work, he just shows mathematically that this will be the case. Now, of course, in real world, you have loss of precision. So the next question is, "How does that affect the computation of our problems?" This is what we are after. Loss of precision means information loss or entropy production. So what we are really looking at, the relationship between hardness and cost of computing of a problem. (clears throat) And according to Sean Harget, there is this left branch, which in principle could be polynomial time, but the question, whether or not this is achievable, that is not achievable, but something more achievable that's on the right-hand side. You know, there's always going to be some information loss, some entropy generation that could keep you away from, possibly from polynomial time. So this is what we'd like to understand. And this information loss, the source of this is not just noise, as, as I will argue in any physical system, but it's also of algorithmic nature. So that is a questionable area or, or approach, but Schönhage's result is purely theoretical, no actual solver is proposed. So we can ask, you know, just theoretically, out of curiosity, "Would in principle be such solvers?" Because he's not proposing a solver. In such properties in principle, if you were to look mathematically, precisely what that solver does, would have the right properties. And I argue, yes, I don't have a mathematical proof but I have some arguments that this would be the case. And this is the case for actually our sitdia solver, that if you could calculate, it's subjectivity in a loss this way, then it would be... Would solve NP-complete problems in polynomial continuous-time. Now, as a matter of fact, this is a bit more difficult question because time in all these can be re-scaled however you want. So what Bournez says, that you actually have to measure the length of the trajectory which is an invariant of the dynamical system or the property of the dynamical system, not of it's parametrization. And we did that. So Shubha Kharel my student did that, by first improving on the stiffness of the problem of the integrations using the implicit solvers and some smart tricks, such that you actually are closer to the actual trajectory and using the same approach to know, what fraction of problems you can solve. We did not give a length of the trajectory, you find that it is polynomialy scaling with the problem size. So we have polynomial scale complexity. That means that our solver is both poly-length, and as it is defined, it's also poly-time analog solver. But if you look at as a discrete algorithm, which will measure the discrete steps on a digital machine, it is an exponential solver, and the reason is because of all this stiffness. So every integrator has to truncate, digitize and truncate the equations. And what it has to do is to keep the integration within this so-called Stimpy TD gen for, for that scheme. And you have to keep this product within Eigenvalues of the Jacobian and the step size within this region, if you use explicit methods, you want to stay within this region. But what happens, that some of the eigenvalues grow fast for stiff problems, and then you're, you're forced to reduce that t, so the product stays in this bounded domain, which means that now you have to, we are forced to take smaller and smaller time steps, so you're, you're freezing out the integration and what I will show you, that's the case. Now you can move to implicit solvers, which is a new trick, in this case, your stability domain is actually on the outside, but what happens in this case, is some of the eigenvalues of the Jacobian, also for this instant start to move to zero, as they are moving to zero, they are going to enter this instability region. So your solver is going to try to keep it out, so it's going to increase the delta t, but if you increase that t, you increase the truncation errors, so you get randomized in the large search space. So it's, it's really not, not willing to work out. Now, one can sort of, introduce a theory or a language to discuss computational, analog computational complexity, using the language from dynamical systems theory. But basically I don't have time to go into this but you have for hard problems, the chaotic object the chaotic saddle in the middle of the search space somewhere, and that dictates how the dynamics happens and invariant properties of the dynamics, of course, of that saddle is what determines performance and many things. So an important measure that we find that, is also helpful in describing, this analog complexity is the so-called Kolmogorov or metric entropy. And basically what this does in an intuitive way, is to describe the rate at which the uncertainty, containing the insignificant digits of a trajectory in the back, they flow towards the significant ones, as you lose information because of errors being, grown or, or or, or developed into larger errors in an exponential, at an exponential rate because you have positive Lyapunov exponents. But this is an invariant property. It's the property of the set of all these, not how you compute them. And it's really the intrinsic rate of accuracy loss of a dynamical system. As I said that you have in such a high dimensional dynamical system, you have positive and negative Lyapunov exponents, as many as the total is the dimension of the space and user dimension, the number of unstable manufactured dimensions and assets now more stable many forms dimensions. And there's an interesting and I think important Pesin equality, equality called the Pesin equality, that connects the information theoretic, as per the rate of information loss with the geometric data each trajectory separate minus cut part which is the escape rate that I already talked about. Now, one can actually prove a simple theorem strike back of the calculation. The idea here is that, you know the rate at which the largest rate at which the closely started trajectory, separate from one another. So now you can say that, that is fine, as long as my trajectory finds the solution, before the trajectory separate too quickly. In that case, I can have the hope, that if I start from some region of the face space, several closely started trajectories, they kind of go into the same solution over time and that's, that's, that's this upper bound of this limit. And it is really showing that it has to be... It's an exponentially smaller number, but it depends on the N, dependence of the exponent right here, which combines information loss rate and the solution time performance. So these, if these exponent here or there, has a large independence, so even a linear independence, then you really have to start trajectories, exponentially closer to one another, in order to end up in the same order. So this is sort of like the, the direction that you are going into, and this formulation is applicable to, to all dynamical systems, deterministic dynamical systems. (clears throat) And I think we can expand this further because the, there is a way of getting the expression for the escape rates in terms of N the number of variables from cycle expansions, that I don't have time to talk about, but it's kind of like a program that you can try to pursue. And this is it. So uh, uh... The conclusions, I think are self-explanatory. I think there is a lot of future in, in analog continuous-time computing. They can be efficient by orders of magnitude than digital ones in solving NP-hard problems, because first of all, many of the systems lack of von Neumann bottleneck, there's parallelism involved and you can also have a larger spectrum of continuous-time dynamical algorithms than discrete ones. And, and, you know, but we also have to be mindful of what are the possibilities, what are the limits? And one, one open question, if any important open question is you know, "What are these limits? "Is there some kind of no-go theorem that tells you that, "you can never perform better than this limit "or, or that limit?" And I think that's, that's the exciting part to, to derive these, these limits and to get to an understanding about what's possible in this, in this area. Thank you.

Published Date : Sep 21 2020

SUMMARY :

in some of the corners of this hypercube.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Zoltan ToroczkaiPERSON

0.99+

Sean HargetPERSON

0.99+

1978DATE

0.99+

2016DATE

0.99+

A. SchönhagePERSON

0.99+

Graham SchönhagePERSON

0.99+

Shubha KharelPERSON

0.99+

BournezPERSON

0.99+

University of Notre DameORGANIZATION

0.99+

SchönhagePERSON

0.99+

Redefine LabORGANIZATION

0.99+

twoQUANTITY

0.99+

todayDATE

0.99+

thirdQUANTITY

0.99+

thousandsQUANTITY

0.99+

two boundsQUANTITY

0.99+

RamsayPERSON

0.98+

zeroQUANTITY

0.98+

oneQUANTITY

0.98+

each trajectoryQUANTITY

0.98+

bothQUANTITY

0.98+

four digitsQUANTITY

0.96+

firstQUANTITY

0.96+

von NeumannPERSON

0.96+

JacobianOTHER

0.95+

EachQUANTITY

0.9+

YoshianORGANIZATION

0.9+

LyapunovOTHER

0.9+

zerosQUANTITY

0.87+

thousands of problemsQUANTITY

0.86+

two magnitudeQUANTITY

0.85+

both caseQUANTITY

0.81+

minus oneQUANTITY

0.78+

3-SATOTHER

0.74+

one open questionQUANTITY

0.73+

Runge KuttaPERSON

0.73+

plusQUANTITY

0.66+

3OTHER

0.65+

everyQUANTITY

0.65+

ClauseOTHER

0.56+

onesQUANTITY

0.55+

N BooleanOTHER

0.55+

SATOTHER

0.46+

problemsQUANTITY

0.44+

Ruth Marinshaw, Research Computing | WiDS 2018


 

>> Narrator: Live from Stanford University in Palo Alto, California, it's theCube, covering Women in Data Science conference 2018. Brought to you by Stanford. >> Welcome back to theCube. I'm Lisa Martin and we're live at Stanford University, the third annual Women in Data Science conference, WiDS. This is a great one day technical event with keynote speakers, with technical vision tracks, career panel and some very inspiring leaders. It's also expected to reach over 100,000 people today, which is incredible. So we're very fortunate to be joined by our next guest, Ruth Marinshaw, the CTO for Research Computing at Stanford University. Welcome to theCube, Ruth. >> Thank you. It's an honor to be here. >> It's great to have you here. You've been in this role as CTO for Research Computing at Stanford for nearly six years. >> That's correct. I came here after about 25 years at the University of North Carolina Chapel Hill. >> So tell us a little bit about what you do in terms of the services that you support to the Institute for Computational Mathematics and Engineering. >> So our team and we're about 17 now supports systems, file systems storage, databases, software across the university to support computational and data intensive science. So ICME, being really the home of computational science education at Stanford from a degree perspective, is a close partner with us. We help them with training opportunities. We try to do some collaborative planning, event promotion, sharing of ideas. We have joint office hours where we can provide system support. Margot's graduate students and data scientists can provide algorithmic support to some thousands of users across the campus, about 500 faculty. >> Wow. So this is the third year for WiDS, your third year here. >> Ruth: It is. >> When you spoke with Margot Gerritsen, who's going to be joining us later today, about the idea for WiDS, what were some of your thoughts about that? Did you expect it to make as big of >> Ruth: No. >> an impact? >> No, no people have been talking about this data tsunami and the rise of big data, literally for 10 years, but actually it arrived. This is the world we live in, data everywhere, that data deluge that had been foreseen or promised or feared was really there. And so when Margot had the idea to start WiDS, I actually thought what a nice campus event. There are women all over Stanford, across this disciplines who are engaged in data science and more who should. Stanford, if anything, is known for its interdisciplinary research and data science is one of those fields that really crosses the schools and the disciplines. So I thought, what a great way to bring women together at Stanford. I clearly did not expect that it would turn into this global phenomenon. >> That is exactly. I love that word, it is a phenomenon. It's a movement. They're expecting, there's, I said over a 100,000 participants today, at more than 150 regional events. I think that number will go up. >> Ruth: Yes. >> During the day. And more than 50 countries. >> Ruth: Yes. >> But it shows, even in three years, not only is there a need for this, there's a demand for it. That last year, I think it was upwards of 75,000 people. To make that massive of a jump in one year and global impact, is huge. But it also speaks to some of the things that Margot and her team have said. It may have been comfortable as one of or the only woman at a boardroom table, but maybe there are others that aren't comfortable and how do we help them >> Ruth: Exactly. >> and inspire them and inspire the next generation. >> Exactly. I think it's a really very powerful statement and demonstration of the importance of community and building technical teams in making, as you said, people comfortable and feeling like they're not alone. We see what 100,000 women maybe joining in internationally over this week for these events. That's such a small fraction compared to what the need probably is to what the hunger probably is. And as Margot said, we're a room full of women here today, but we're still such a minority in the industry, in the field. >> Yes. So you mentioned, you've been here at Stanford for over five years, but you were at Chapel Hill before. >> Ruth: Yes. >> Tell me a little bit about your career path in the STEM field. What was your inspiration all those years ago to study this? >> My background is actually computational social sciences. >> Lisa: Oh interesting. >> And so from an undergraduate and graduate perspective and this was the dawn of western civilization, long ago, not quite that long (Lisa laughs) but long ago and even then, I was drawn to programming and data analysis and data sort of discovery. I as a graduate student and then for a career worked at a demographic research center at UNC Chapel Hill, where firsthand you did data science, you did original data collection and data analysis, data manipulation, interpretation. And then parlayed that into more of a technical role, learning more programming languages, computer hardware, software systems and the like. And went on to find that this was really my love, was technology. And it's so exciting to be here at Stanford from that perspective because this is the birthplace of many technologies and again, referencing the interdisciplinary nature of work here, we have some of the best data scientists in the world. We have some of the best statisticians and algorithm developers and social scientists, humanists, who together can really make a difference in solving, using big data, data science, to solve some of the pressing problems. >> The social impact that data science and computer science alone can make with ideally a diverse set of eyes and perspectives looking at it, is infinite. >> Absolutely. And that's one reason I'm super excited today, this third WiDS for one of the keynote speakers, Latanya from Harvard. She's going to be talking, she's from government and sort of political science, but she's going to be talking about data science from the policy perspective and also the privacy perspective. >> Lisa: Oh yes. >> I think that this data science provides such great opportunity, not just to have the traditional STEM fields participating but really to leverage the ethicists and the humanists and the social sciences so we have that diversity of opinions shaping decision making. >> Exactly. And as much as big data and those technologies open up a lot of opportunities for new business models for corporations, I think so does it also in parallel open up new opportunities for career paths and for women in the field all over the world to make a big, big difference. >> Exactly. I think that's another value add for WiDS over it's three years is to expose young women to the range of career paths in which data science can have an impact. It's not just about coding, although that's an important part. As we heard this morning, investment banking, go figure. Right now SAP is talking about the impact on precision medicine and precision healthcare. Last year, we had the National Security Agency here, talking about use of data. We've had geographers. So I think it helps broaden the perspective about where you can take your skills in data science. And also expose you to the full range of skills that's needed to make a good data science team. >> Right. The hard skills, right, the data and statistical analyses, the computational skills, but also the softer skills. >> Ruth: Exactly. >> How do you see that in your career as those two sides, the hard skills, the soft skills coming together to formulate the things that you're doing today? >> Well we have to have a diverse team, so I think the soft skills come into play not just from having women on your team but a diversity of opinions. In all that we do in managing our systems and making decisions about what to do, we do look at data. They may not be data at scale that we see in healthcare or mobile devices or you know, our mobile health, our Fitbit data. But we try to base our decisions on an analysis of data. And purely running an algorithm or applying a formula to something will give you one perspective, but it's only part of the answer. So working as a team to evaluate other alternative methods. There never is just one right way to model something, right. And I think that, having the diversity across the team and pulling in external decision makers as well to help us evaluate the data. We look at the hard science and then we ask about, is this the right thing to do, is this really what the data are telling us. >> So with WiDS being aimed at inspiring and educating data scientists worldwide, we kind of talked a little bit already about inspiring the younger generation who are maybe as Maria Callaway said that the ideal time to inspire young females is first semester of college. But there's also sort of a flip side to that and I think that's reinvigorating. >> Yes. >> That the women who've been in the STEM field or in technology for awhile. What are some of the things that you have found invigorating in your own career about WiDS and the collaboration with other females in the industry? >> I think hearing inspirational speakers like Maria, last here and this year, Diane Greene from Google last year, talk about just the point you made that there's always opportunity, there's always time to learn new things, to start a new career. We don't have to be first year freshmen in college in order to start a career. We're all lifelong learners and to hear women present and to see and meet with people at the breakout sessions and the lunch, whose careers have been shaped by and some cases remade by the opportunity to learn new things and apply those skills in new areas. It's just exciting. Today for this conference, I brought along four or five of my colleagues from IT at Stanford, who are not data scientists. They would not call themselves data scientists, but there are data elements to all of their careers. And watching them in there this morning as they see what people are doing and hear about the possibilities, it's just exciting. It's exciting and it's empowering as well. Again back to that idea of community, you're not in it alone. >> Lisa: Right. >> And to be connected to all of these women across a generation is really, it's just invigorating. >> I love that. It's empowering, it is invigorating. Did you have mentors when you were in your undergraduate >> Ruth: I did. >> days? Were they males, females, both? >> I'd say in undergraduate and graduate school, actually they were more males from an academic perspective. But as a graduate student, I worked in a programming unit and my mentors there were all females and one in particular became then my boss. And she was a lifelong mentor to me. And I found that really important. She believed in women. She believed that programming was not a male field. She did not believe that technology was the domain only of men. And she really was supportive throughout. And I think it's important for young women as well as mid-career women to continue to have mentors to help bounce ideas off of and to help encourage inquiries. >> Definitely, definitely. I'm always surprised every now and then when I'm interviewing females in tech, they'll say I didn't have a mentor. >> Lisa: Oh. >> So I had to become one. But I think you know we think maybe think of mentors in an earlier stage of our careers, but at a later stage we talked about that reinvigoration. Are you finding WiDS as a source of maybe not only for you to have the opportunity to mentor more women but also are you finding more mentors of different generations >> Oh sure. >> as being part of WiDS? >> Absolutely, think of Karen Mathis, not just Margot but Karen, getting to know her. And we go for sort of walks around the campus and bounce ideas of each other. I think it is a community for yes, for all of us. It's not just for the young women and we want to remain engaged in this. The fact that it's global now, I think a new challenge is how do we leverage this international community now. So our opportunities for mentorship and partnership aren't limited to our local WiDS. They're an important group. But how do we connect across those different communities? >> Lisa: Exactly. >> They're international now. >> Exactly. I think I was on Twitter last night and there was the WiDS New Zealand about to go live. >> Yeah, yeah. >> And I just thought, wow it's this great community. But you make a good point that it's reached such scale so quickly. Now it's about how can we learn from women in different industries in other parts of the world. How can they learn from us? To really grow this foundation of collaboration and to a word you said earlier, community. >> It really is amazing though that in three years WiDS has become what it has because if you think about other organizations, special interest groups and the like, often they really are, they're not parochial. But they tend to be local and if they're national, they're not at this scale. >> Right. >> And so again back to it's the right time, it's the right set of organizers. I mean Margot, anything that she touches, she puts it herself completely into it and it's almost always successful. The right people, the right time. And finding ways to harness and encourage enthusiasm in really productive ways. I think it's just been fabulous. >> I agree. Last question for you. Looking back at your career, what advice would you have given young Ruth? >> Oh gosh. That's a really great question. I think to try to connect as much as you can outside your comfort zone. Back to that idea of mentorship. You think when you're an undergraduate, you explore curricula, you take crazy classes, Chinese or, not that that's crazy, but you know if you're a math major and you go take art or something. To really explore not just your academic breadth but also career opportunities and career understanding earlier on that really, oh I want to be a doctor, actually what do you know about being a doctor. I don't want to be a statistician, well why not? So I think to encourage more curiosity outside the classroom in terms of thinking about what is the world about and how can you make a difference. >> I love that, getting out of the comfort zone. One of my mentors says get comfortably uncomfortable and I love that. >> Ruth: That's great, yeah. >> I love that. Well Ruth, thank you so much for joining us on theCube today. It's our pleasure to have you here and we hope you have a great time at the event. We look forward to talking with you next time. >> We'll see you next year. >> Lisa: Excellent. >> Thank you. Buh-bye. >> I'm Lisa Martin. You're watching theCube live from Stanford University at the third annual Women in Data Science conference. #WiDS2018, join the conversation. After this short break, I'll be right back with my next guest. Stick around. (techno music)

Published Date : Mar 5 2018

SUMMARY :

Brought to you by Stanford. It's also expected to reach over 100,000 people today, It's an honor to be here. It's great to have you here. at the University of North Carolina Chapel Hill. in terms of the services that you support So ICME, being really the home So this is the third year for WiDS, and the rise of big data, literally for 10 years, I love that word, it is a phenomenon. During the day. But it also speaks to some of the things that Margot and inspire the next generation. and demonstration of the importance of community So you mentioned, you've been here at Stanford in the STEM field. And it's so exciting to be here at Stanford The social impact that data science and computer science and also the privacy perspective. and the social sciences so we have that diversity and for women in the field all over the world And also expose you to the full range of skills The hard skills, right, the data and statistical analyses, to something will give you one perspective, But there's also sort of a flip side to that and the collaboration with other females in the industry? and to hear women present and to see and meet with people And to be connected to all of these women Did you have mentors when you were in your undergraduate and to help encourage inquiries. I'm always surprised every now and then But I think you know we think maybe think of mentors It's not just for the young women and there was the WiDS New Zealand about to go live. and to a word you said earlier, community. But they tend to be local and if they're national, And so again back to it's the right time, what advice would you have given young Ruth? I think to try to connect as much as you can I love that, getting out of the comfort zone. We look forward to talking with you next time. Thank you. at the third annual Women in Data Science conference.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
MargotPERSON

0.99+

KarenPERSON

0.99+

RuthPERSON

0.99+

Diane GreenePERSON

0.99+

Ruth MarinshawPERSON

0.99+

Lisa MartinPERSON

0.99+

Karen MathisPERSON

0.99+

Maria CallawayPERSON

0.99+

LisaPERSON

0.99+

National Security AgencyORGANIZATION

0.99+

Margot GerritsenPERSON

0.99+

Institute for Computational Mathematics and EngineeringORGANIZATION

0.99+

Last yearDATE

0.99+

one yearQUANTITY

0.99+

last yearDATE

0.99+

fiveQUANTITY

0.99+

third yearQUANTITY

0.99+

three yearsQUANTITY

0.99+

two sidesQUANTITY

0.99+

LatanyaPERSON

0.99+

100,000 womenQUANTITY

0.99+

TodayDATE

0.99+

fourQUANTITY

0.99+

10 yearsQUANTITY

0.99+

Palo Alto, CaliforniaLOCATION

0.99+

one dayQUANTITY

0.99+

next yearDATE

0.99+

bothQUANTITY

0.99+

last nightDATE

0.99+

OneQUANTITY

0.99+

more than 50 countriesQUANTITY

0.99+

MariaPERSON

0.99+

oneQUANTITY

0.99+

over five yearsQUANTITY

0.99+

more than 150 regional eventsQUANTITY

0.98+

this yearDATE

0.98+

over 100,000 peopleQUANTITY

0.98+

FitbitORGANIZATION

0.98+

Stanford UniversityORGANIZATION

0.98+

thousandsQUANTITY

0.98+

first yearQUANTITY

0.98+

about 500 facultyQUANTITY

0.98+

StanfordORGANIZATION

0.98+

todayDATE

0.98+

first semesterQUANTITY

0.98+

75,000 peopleQUANTITY

0.98+

WiDSEVENT

0.98+

one reasonQUANTITY

0.98+

University of North CarolinaORGANIZATION

0.97+

GoogleORGANIZATION

0.97+

about 25 yearsQUANTITY

0.97+

Chapel HillLOCATION

0.96+

SAPORGANIZATION

0.96+

one perspectiveQUANTITY

0.95+

UNC Chapel HillORGANIZATION

0.94+

#WiDS2018EVENT

0.94+

WiDSORGANIZATION

0.94+

New ZealandLOCATION

0.94+

this weekDATE

0.93+

thirdQUANTITY

0.92+

WiDS 2018EVENT

0.92+

Kelly Mungary, Lions Gate & Bob Muglia, Snowflake Computing | AWS re:Invent 2017


 

>> Narrator: Live from Las Vegas, it's The Cube, covering AWS re:Invent 2017. Presented by AWS, Intel, and our ecosystem of partners. >> Bob: It's actually a little quieter here. >> Hey, welcome back to AWS re:Invent 2017. I am Lisa Martin. We're all very chatty. You can hear a lot of chatty folks behind us. This is day two of our continuing coverage. 42,000 people here, amazing. I'm Lisa Martin with my co-host Keith Townsend, and we're very excited to be joined by a Cube alumni Bob Muglia, CEO and President of Snowflake. >> Thank you. >> Lisa: Welcome back. >> Thank you, good to be back. >> And Kelly Mungary, the Director of Enterprise Data and Analytics from Lionsgate. A great use case from Snowflake. Thanks so much guys for joining us. So one of the hot things going on today at the event is your announcement Bob with AWS and Snowpipe. What is Snowpipe? How do customers get started with it? >> Great, well thanks. We're excited about Snowpipe. Snowpipe is a way of ingesting data into Snowflake in a streaming, continuous way. You simply can drop new data that's coming in into S3 and we'll ingest it for you automatically. Makes that super, super simple. Brings the data in continuously into your data warehouse, ensuring that you're always up to date and your analysts are getting the latest insights and the latest data. >> So, when you guys were founded, about five years ago, as the marketing says on your website, a complete data warehouse built for the Cloud. What was the opportunity back then? What did you see that was missing, and how has Snowflake evolved to really be a leader in this space? >> So you know, if you go back five years this was a time frame where no SQL was the big rage, and everybody was talking about how SQL was passe and it's something that you're not see in the future. Our founders had a different view, they had been working on true relational databases for almost 20 years, and they recognized the power of SQL and relational technology but they also saw that customers were experiencing significant limits with existing technology, and those limits really restricted what people could do. They saw in the Cloud and what Amazon had done the ability to build a all new database that takes advantage of the full elasticity and power of the Cloud to deliver whatever set of analytics capabilities that the business requires. However much data you want, however many queries simultaneously. Snowflake takes what you love about a relational database and removes all the limits, and allows you to operate in a very different way. And our founders had that vision five years ago, and really successfully executed on it. The product has worked beyond our dreams, and our customers, our response from our customers is what we get so excited about. >> So, the saying is "Data is the new oil". However, just as oil is really hard to drill for and find, finding the data to service up, to even put in a data lake to analyze has been a challenge. How did you guys go about identifying what data should even be streamed to Snowpipe? >> Well, yeah, that's a great question. I mean, in entertainment today, we're experiencing probably like in pretty much every type of business. A data explosion. We have, you know, streaming is big now. We have subscription data coming in, billing data, social media data, and on and on. And the thing is, it's not coming in a normal, regular format. It's coming in what we call a semi-structured, structured, json, xml. So, up until Snowflake came onto the scene with a truly Cloud based SAAS solution for data warehousing pretty much everyone was struggling to wrangle in all these data sets. Snowpipe is a great example of one of the avenues of bringing in these multiple data sets, merging them real time, and getting the analytics out to your business in an agile way that has never been seen before. >> So, can you talk a little bit about that experience? Kinda that day one up, you were taking these separate data sources, whether it's ERP solution, data from original content, merging that together and then being able to analyze that. What was that day one experience like? >> Well, you know, I gotta tell you, it evolves around a word, that word is "Yes", okay? And data architects and executives and leaders within pretty much every company are used to saying, "We'll get to that" and "We'll put it on the road map", "We could do that six months out", "Three months out". So what happened when I implemented Snowflake was I was just walking into meetings and going, "Yes". "You got it". "No worries, let's do it". >> Lisa: It liberated. >> Well, it's changes, it's not only liberating, it changes the individual's opportunities, the team's opportunities, the company's opportunities, and ultimately, revenue. So, I think it's just an amazing new way of approaching data warehousing. >> So Bob, can you talk a little bit about the partnership with AWS, and the power to bring that type of capability to customers? Data lakes are really hard to do that type of thing run a query against to get instant answers. Talk about the partnership with AWS to bring that type of capability. >> Well Amazon's been a fantastic partner of ours, and we really enjoy working with Amazon. We wind up working together with them to solve customer problems. Which is what I think is so fantastic. And with Snowflake, on top of Amazon, you can do what Kelly's saying. You can say yes, because all of a sudden you can now bring all of your data together in one place. Technology has limited, it's technology that has caused data to be in disparate silos. People don't want their data all scattered all over the place. It's all in these different places because limits to technology force people to do that. With the Cloud, and with what Amazon has done and with a product like Snowflake, you can bring all of that data together, and the thing that's interesting, where Kelly is going, is it can change the culture of a company, and the way people work. All of a sudden, data is not power. Data is available to everyone, and it's democratizing. Every person can work with data and help to bring the business forward. And it can really change the dynamics about the way people work. >> And Kelly, you just spoke at the multi-city Cloud Analytics Tour that Snowflake just did. You spoke in Santa Monica, one of my favorite places. You talked about a data driven culture. And we hear data driven in so many different conversations, but how did you actually go about facilitating a data driven culture. Who are some of the early adopters, and what business problems have you been able to solve by saying yes? >> Well, I can speak entertainment in general. I think that it's all about technology it's about talent, and it's about teaching. And with technology being the core of that. If we go back five years, six years, seven years, it was really hard to walk into a room, have an idea, a concept, around social media, around streaming data, around billing, around accounting. And to have an agile approach that you could bring together within a week or so forth. So what's happening is, now that we've implemented Snowflake on AWS and some of the other what I call dream tools on top of that. The dream stack, which includes Snowflake. It's more about integrating with the business. Now we can speak the same language with them. Now we can walk into a room and they're glad to see me now. And at the end of the day, it's new, it's all new. So, this is something that I say sometimes, in kidding, but it's actually true. It's as if Snowflake had a time traveler on staff that went forward in the future ten years to determine how things should be done in the big data space, and then came back and developed it. And that's how futuristic they are, but proven at the same time. And that allows us to cultivate that data driven culture within entertainment, because we have tools and we have the agile approach that the business is looking for. >> So, Kelly, I'm really interested, and I love the concept of making data available to everyone. That's been a theme of this conference from the keynote this morning, which is putting tools in builder's hands, and allowing builders to do what they do. >> Kelly: That's right. >> And we're always surprised at what users come back with. What's one of the biggest surprises from the use cases, now that you've enabled your users. >> Well, I'm gonna give you one that's based on AWS and Snowflake. A catch phrase you hear a lot of is "Data center of excellence", and a lot of us are trying to build out these data centers of excellence, but it's a little bit of an oxymoron to the fact that a data center of excellence is really about enabling your business and finding champions within marketing, within sales, within accounting, and giving them the ability to have self-service business intelligence, self-service data warehousing. The kinds of things that, again, we go back five, six years ago, you couldn't even have that conversation. I'll tell you today, I can walk into a room, and say, "Okay, who here is interested in learning "about data warehousing?". And there'll be somebody, "Okay, great". Within an hour, I'll have you being dangerous in terms of setting up, standing up, configuring and loading a data warehouse. That's unheard of, and it's all due to Snowflake and their new technology. >> I'd love to understand Bob, from your perspective. First of all, it sounds like you have a crystal ball according to Kelly, which is awesome. But second of all, collaboration, we talked about that earlier. Andy Jassy is very well known and very vocal about visiting customers every week. And I love their bottom, their backwards approach to, before building a product, to try to say, "What problem can we solve?". They're actually working with customers first. What are their requirements? Tell me a little bit Bob about the collaboration that Snowflake has with Lionsgate, or other customers. How are they helping to influence your crystal ball? >> You know what, this is where I think what Amazon has done, and Andy has done a fantastic job. There's so much to learn from them, and the customer centricity that Amazon has always had is something that we have really focused to bring into Snowflake, and really build deeply into our culture. I've sort of said many, many times, Snowflake is a value space company. Our values are important to us, they're prominent in our website. Our first value is we put our customer's first. What I'm most proud of is, every customer who has focused on deploying Snowflake, has successfully deployed Snowflake, and we learn from them. We engage with them. We partner with them. All of our customers are our partners. Kelly and Lionsgate are examples of customers that we learn from every day, and it's such a rewarding thing to hear what they want to do. You look at Snowpipe and what Snowpipe is, that came from customers, we learned that from customers. You look at so many features, so many details. It's iterative learning with customers. And what's interesting about that, it's listening to customers, but it's also understanding what they do. One of the things that's interesting about Snowflake is is that as a company we run Snowflake on Snowflake. All of our data is in Snowflake. All of our sales data, our financial data, our marketing data, our product support data, our engineering data. Every time a user runs a query, that query is logged in Snowflake and intrinsics about it are logged. So what's interesting is because it's all in one place, and it's all accessible, we can answer essentially any question, about what's been done. And then, driving the culture to do that is an important thing. One of the things I do find interesting is, even at Snowflake, even at this data centered company, even where everything is all centralized, I still find sometimes people don't reference it. And I'm constantly reinforcing that your intuition, you know, you're really smart, you're really intuitive, but you could be wrong. And if you can answer the question based on what's happened, what your customers are doing, because it's in the data, and you can get that answer quickly, it's a totally different world. And that's what you can do when you have a tool with the power of what Snowflake can deliver, is you could answer effectively any business question in just a matter of minutes, and that's transformative, it's transformative to the way people work, and that, to me, that's about what it means to build a data driven culture. Is to reinforce that the answer is inside what customers are doing. And so often, that is encapsulated in the data. >> Wow, your energy is incredible. We thank you so much Bob and Kelly for coming on and sharing your story. And I think a lot of our viewers are gonna learn some great lessons from both of you on collaboration on transformations. So thanks so much for stopping by. >> Yeah. >> Thank you so much, we really enjoyed it. Thanks a lot. >> Likewise, great to meet you. >> Thanks Kelly. >> Thank you. >> For my co-host Keith Townsend, and for Kelly and Bob, I am Lisa Martin. You've been watching The Cube, live on day two, continuing coverage at AWS re:Invent 2017. Stick around, we have great more guests coming up. (upbeat music)

Published Date : Nov 29 2017

SUMMARY :

it's The Cube, covering AWS re:Invent 2017. Bob Muglia, CEO and President of Snowflake. And Kelly Mungary, the Director and the latest data. as the marketing says on your website, and power of the Cloud to deliver finding the data to service up, Snowpipe is a great example of one of the avenues Kinda that day one up, you were taking these separate Well, you know, I gotta tell you, it changes the individual's opportunities, the partnership with AWS, and the power and the thing that's interesting, And Kelly, you just spoke And at the end of the day, it's new, it's all new. and I love the concept of making data available to everyone. from the use cases, now that you've enabled your users. and a lot of us are trying to build out How are they helping to influence your crystal ball? and that, to me, that's about what it means are gonna learn some great lessons from both of you Thank you so much, we really enjoyed it. and for Kelly and Bob, I am Lisa Martin.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
KellyPERSON

0.99+

AWSORGANIZATION

0.99+

Keith TownsendPERSON

0.99+

Kelly MungaryPERSON

0.99+

AndyPERSON

0.99+

AmazonORGANIZATION

0.99+

Lisa MartinPERSON

0.99+

Bob MugliaPERSON

0.99+

BobPERSON

0.99+

Andy JassyPERSON

0.99+

Santa MonicaLOCATION

0.99+

LionsgateORGANIZATION

0.99+

LisaPERSON

0.99+

SnowflakeORGANIZATION

0.99+

The CubeTITLE

0.99+

todayDATE

0.99+

seven yearsQUANTITY

0.99+

six yearsQUANTITY

0.99+

42,000 peopleQUANTITY

0.99+

ten yearsQUANTITY

0.99+

five years agoDATE

0.99+

SQLTITLE

0.99+

Three monthsQUANTITY

0.98+

SnowflakeTITLE

0.98+

six monthsQUANTITY

0.98+

five yearsQUANTITY

0.98+

IntelORGANIZATION

0.98+

bothQUANTITY

0.97+

firstQUANTITY

0.97+

Lions GateORGANIZATION

0.97+

OneQUANTITY

0.97+

one placeQUANTITY

0.96+

CubeORGANIZATION

0.96+

almost 20 yearsQUANTITY

0.96+

oneQUANTITY

0.95+

FirstQUANTITY

0.95+

day twoQUANTITY

0.95+

first valueQUANTITY

0.95+

SnowpipeORGANIZATION

0.93+

Las VegasLOCATION

0.92+

SnowpipeTITLE

0.92+

Snowflake ComputingORGANIZATION

0.91+

six years agoDATE

0.9+

Armughan Ahmad, Dell EMC | Super Computing 2017


 

>> Announcer: From Denver, Colorado, it's theCUBE, covering Super Computing 17. Brought to you by Intel. (soft electronic music) Hey, welcome back, everybody. Jeff Frick here with theCUBE. We're gettin' towards the end of the day here at Super Computing 2017 in Denver, Colorado. 12,000 people talkin' really about the outer limits of what you can do with compute power and lookin' out into the universe and black holes and all kinds of exciting stuff. We're kind of bringin' it back, right? We're all about democratization of technology for people to solve real problems. We're really excited to have our last guest of the day, bringin' the energy, Armughan Ahmad. He's SVP and GM, Hybrid Cloud and Ready Solutions for Dell EMC, and a many-time CUBE alumni. Armughan, great to see you. >> Yeah, good to see you, Jeff. So, first off, just impressions of the show. 12,000 people, we had no idea. We've never been to this show before. This is great. >> This is a show that has been around. If you know the history of the show, this was an IEEE engineering show, that actually turned into high-performance computing around research-based analytics and other things that came out of it. But, it's just grown. We're seeing now, yesterday the super computing top petaflops were released here. So, it's fascinating. You have some of the brightest minds in the world that actually come to this event. 12,000 of them. >> Yeah, and Dell EMC is here in force, so a lot of announcements, a lot of excitement. What are you guys excited about participating in this type of show? >> Yeah, Jeff, so when we come to an event like this, HBC-- We know that HBC is also evolved from your traditional HBC, which was around modeling and simulation, and how it started from engineering to then clusters. It's now evolving more towards machine learning, deep learning, and artificial intelligence. So, what we announced here-- Yesterday, our press release went out. It was really related to how our strategy of advancing HBC, but also democratizing HBC's working. So, on the advancing, on the HBC side, the top 500 super computing list came out. We're powering some of the top 500 of those. One big one is TAC, which is Texas Institute out of UT, University of Texas. They now have, I believe, the number 12 spot in the top 500 super computers in the world, running an 8.2 petaflops off computing. >> So, a lot of zeros. I have no idea what a petaflop is. >> It's very, very big. It's very big. It's available for machine learning, but also eventually going to be available for deep learning. But, more importantly, we're also moving towards democratizing HBC because we feel that democratizing is also very important, where HBC should not only be for the research and the academia, but it should also be focused towards the manufacturing customers, the financial customers, our commercial customers, so that they can actually take the complexity of HBC out, and that's where our-- We call it our HBC 2.0 strategy, off learning from the advancements that we continue to drive, to then also democratizing it for our customers. >> It's interesting, I think, back to the old days of Intel microprocessors getting better and better and better, and you had Spark and you had Silicon Graphics, and these things that were way better. This huge differentiation. But, the Intel I32 just kept pluggin' along and it really begs the question, where is the distinction now? You have huge clusters of computers you can put together with virtualization. Where is the difference between just a really big cluster and HBC and super computing? >> So, I think, if you look at HBC, HBC is also evolving, so let's look at the customer view, right? So, the other part of our announcement here was artificial intelligence, which is really, what is artificial intelligence? It's, if you look at a customer retailer, a retailer has-- They start with data, for example. You buy beer and chips at J's Retailer, for example. You come in and do that, you usually used to run a SEQUEL database or you used to run a RDBMS database, and then that would basically tell you, these are the people who can purchase from me. You know their purchase history. But, then you evolved into BI, and then if that data got really, very large, you then had an HBC cluster, would which basically analyze a lot of that data for you, and show you trends and things. That would then tell you, you know what, these are my customers, this is how many times they are frequent. But, now it's moving more towards machine learning and deep learning as well. So, as the data gets larger and larger, we're seeing datas becoming larger, not just by social media, but your traditional computational frameworks, your traditional applications and others. We're finding that data is also growing at the edge, so by 2020, about 20 billion devices are going to wake up at the edge and start generating data. So, now, Internet data is going to look very small over the next three, four years, as the edge data comes up. So, you actually need to now start thinking of machine learning and deep learning a lot more. So, you asked the question, how do you see that evolving? So, you see an RDBMS traditional SQL evolving to BI. BI then evolves into either an HBC or hadoop. Then, from HBC and hadoop, what do you do next? What you do next is you start to now feed predictive analytics into machine learning kind of solutions, and then once those predictive analytics are there, then you really, truly start thinking about the full deep learning frameworks. >> Right, well and clearly like the data in motion. I think it's funny, we used to make decisions on a sample of data in the past. Now, we have the opportunity to take all the data in real time and make those decisions with Kafka and Spark and Flink and all these crazy systems that are comin' to play. Makes Hadoop look ancient, tired, and yesterday, right? But, it's still valid, right? >> A lot of customers are still paying. Customers are using it, and that's where we feel we need to simplify the complex for our customers. That's why we announced our Machine Learning Ready Bundle and our Deep Learning Ready Bundle. We announced it with Intel and Nvidia together, because we feel like our customers either go to the GPU route, which is your accelerator's route. We announced-- You were talking to Ravi, from our server team, earlier, where he talked about the C4140, which has the quad GPU power, and it's perfect for deep learning. But, with Intel, we've also worked on the same, where we worked on the AI software with Intel. Why are we doing all of this? We're saying that if you thought that RDBMS was difficult, and if you thought that building a hadoop cluster or HBC was a little challenging and time consuming, as the customers move to machine learning and deep learning, you now have to think about the whole stack. So, let me explain the stack to you. You think of a compute storage and network stack, then you think of-- The whole eternity. Yeah, that's right, the whole eternity of our data center. Then you talk about our-- These frameworks, like Theano, Caffe, TensorFlow, right? These are new frameworks. They are machine learning and deep learning frameworks. They're open source and others. Then you go to libraries. Then you go to accelerators, which accelerators you choose, then you go to your operating systems. Now, you haven't even talked about your use case. Retail use case or genomic sequencing use case. All you're trying to do is now figure out TensorFlow works with this accelerator or does not work with this accelerator. Or, does Caffe and Theano work with this operating system or not? And, that is a complexity that is way more complex. So, that's where we felt that we really needed to launch these new solutions, and we prelaunched them here at Super Computing, because we feel the evolution of HBC towards AI is happening. We're going to start shipping these Ready Bundles for machine learning and deep learning in first half of 2018. >> So, that's what the Ready Solutions are? You're basically putting the solution together for the client, then they can start-- You work together to build the application to fix whatever it is they're trying to do. >> That's exactly it. But, not just fix it. It's an outcome. So, I'm going to go back to the retailer. So, if you are the CEO of the biggest retailer and you are saying, hey, I just don't want to know who buys from me, I want to now do predictive analytics, which is who buys chips and beer, but who can I sell more things to, right? So, you now start thinking about demographic data. You start thinking about payroll data and other datas that surround-- You start feeding that data into it, so your machine now starts to learn a lot more of those frameworks, and then can actually give you predictive analytics. But, imagine a day where you actually-- The machine or the deep learning AI actually tells you that it's not just who you want to sell chips and beer to, it's who's going to buy the 4k TV? You're makin' a lot of presumptions. Well, there you go, and the 4k-- But, I'm glad you're doin' the 4k TV. So, that's important, right? That is where our customers need to understand how predictive analytics are going to move towards cognitive analytics. So, this is complex but we're trying to make that complex simple with these Ready Solutions from machine learning and deep learning. >> So, I want to just get your take on-- You've kind of talked about these three things a couple times, how you delineate between AI, machine learning, and deep learning. >> So, as I said, there is an evolution. I don't think a customer can achieve artificial intelligence unless they go through the whole crawl walk around space. There's no shortcuts there, right? What do you do? So, if you think about, Mastercard is a great customer of ours. They do an incredible amount of transactions per day, (laughs) as you can think, right? In millions. They want to do facial recognitions at kiosks, or they're looking at different policies based on your buying behavior-- That, hey, Jeff doesn't buy $20,000 Rolexes every year. Maybe once every week, you know, (laughs) it just depends how your mood is. I was in the Emirates. Exactly, you were in Dubai (laughs). Then, you think about his credit card is being used where? And, based on your behaviors that's important. Now, think about, even for Mastercard, they have traditional RDBMS databases. They went to BI. They have high-performance computing clusters. Then, they developed the hadoop cluster. So, what we did with them, we said okay. All that is good. That data that has been generated for you through customers and through internal IT organizations, those things are all very important. But, at the same time, now you need to start going through this data and start analyzing this data for predictive analytics. So, they had 1.2 million policies, for example, that they had to crunch. Now, think about 1.2 million policies that they had to say-- In which they had to take decisions on. That they had to take decisions on. One of the policies could be, hey, does Jeff go to Dubai to buy a Rolex or not? Or, does Jeff do these other patterns, or is Armughan taking his card and having a field day with it? So, those are policies that they feed into machine learning frameworks, and then machine learning actually gives you patterns that they can now see what your behavior is. Then, based on that, eventually deep learning is when they move to next. Deep learning now not only you actually talk about your behavior patterns on the credit card, but your entire other life data starts to-- Starts to also come into that. Then, now, you're actually talking about something before, that's for catching a fraud, you can actually be a lot more predictive about it and cognitive about it. So, that's where we feel that our Ready Solutions around machine learning and deep learning are really geared towards, so taking HBC to then democratizing it, advancing it, and then now helping our customers move towards machine learning and deep learning, 'cause these buzzwords of AIs are out there. If you're a financial institution and you're trying to figure out, who is that customer who's going to buy the next mortgage from you? Or, who are you going to lend to next? You want the machine and others to tell you this, not to take over your life, but to actually help you make these decisions so that your bottom line can go up along with your top line. Revenue and margins are important to every customer. >> It's amazing on the credit card example, because people get so pissed if there's a false positive. With the amount of effort that they've put into keep you from making fraudulent transactions, and if your credit card ever gets denied, people go bananas, right? The behavior just is amazing. But, I want to ask you-- We're comin' to the end of 2017, which is hard to believe. Things are rolling at Dell EMC. Michael Dell, ever since he took that thing private, you could see the sparkle in his eye. We got him on a CUBE interview a few years back. A year from now, 2018. What are we going to talk about? What are your top priorities for 2018? >> So, number one, Michael continues to talk about that our vision is advancing human progress through technology, right? That's our vision. We want to get there. But, at the same time we know that we have to drive IT transformation, we have to drive workforce transformation, we have to drive digital transformation, and we have to drive security transformation. All those things are important because lots of customers-- I mean, Jeff, do you know like 75% of the S&P 500 companies will not exist by 2027 because they're either not going to be able to make that shift from Blockbuster to Netflix, or Uber taxi-- It's happened to our friends at GE over the last little while. >> You can think about any customer-- That's what Michael did. Michael actually disrupted Dell with Dell technologies and the acquisition of EMC and Pivotal and VMWare. In a year from now, our strategy is really about edge to core to the cloud. We think the world is going to be all three, because the rise of 20 billion devices at the edge is going to require new computational frameworks. But, at the same time, people are going to bring them into the core, and then cloud will still exist. But, a lot of times-- Let me ask you, if you were driving an autonomous vehicle, do you want that data-- I'm an Edge guy. I know where you're going with this. It's not going to go, right? You want it at the edge, because data gravity is important. That's where we're going, so it's going to be huge. We feel data gravity is going to be big. We think core is going to be big. We think cloud's going to be big. And we really want to play in all three of those areas. >> That's when the speed of light is just too damn slow, in the car example. You don't want to send it to the data center and back. You don't want to send it to the data center, you want those decisions to be made at the edge. Your manufacturing floor needs to make the decision at the edge as well. You don't want a lot of that data going back to the cloud. All right, Armughan, thanks for bringing the energy to wrap up our day, and it's great to see you as always. Always good to see you guys, thank you. >> All right, this is Armughan, I'm Jeff Frick. You're watching theCUBE from Super Computing Summit 2017. Thanks for watching. We'll see you next time. (soft electronic music)

Published Date : Nov 16 2017

SUMMARY :

Brought to you by Intel. So, first off, just impressions of the show. You have some of the brightest minds in the world What are you guys excited about So, on the advancing, on the HBC side, So, a lot of zeros. the complexity of HBC out, and that's where our-- You have huge clusters of computers you can and then if that data got really, very large, you then had and all these crazy systems that are comin' to play. So, let me explain the stack to you. for the client, then they can start-- The machine or the deep learning AI actually tells you So, I want to just get your take on-- But, at the same time, now you need to start you could see the sparkle in his eye. But, at the same time we know that we have to But, at the same time, people are going to bring them and it's great to see you as always. We'll see you next time.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
MichaelPERSON

0.99+

Jeff FrickPERSON

0.99+

JeffPERSON

0.99+

DubaiLOCATION

0.99+

ArmughanPERSON

0.99+

$20,000QUANTITY

0.99+

Michael DellPERSON

0.99+

EMCORGANIZATION

0.99+

2018DATE

0.99+

TACORGANIZATION

0.99+

NvidiaORGANIZATION

0.99+

2027DATE

0.99+

Armughan AhmadPERSON

0.99+

DellORGANIZATION

0.99+

12,000QUANTITY

0.99+

EmiratesLOCATION

0.99+

75%QUANTITY

0.99+

MastercardORGANIZATION

0.99+

NetflixORGANIZATION

0.99+

2020DATE

0.99+

PivotalORGANIZATION

0.99+

8.2 petaflopsQUANTITY

0.99+

C4140COMMERCIAL_ITEM

0.99+

12,000 peopleQUANTITY

0.99+

Texas InstituteORGANIZATION

0.99+

GEORGANIZATION

0.99+

OneQUANTITY

0.99+

1.2 million policiesQUANTITY

0.99+

J's RetailerORGANIZATION

0.99+

Denver, ColoradoLOCATION

0.99+

YesterdayDATE

0.99+

500 super computersQUANTITY

0.99+

millionsQUANTITY

0.99+

20 billion devicesQUANTITY

0.99+

University of TexasORGANIZATION

0.99+

VMWareORGANIZATION

0.99+

CaffeORGANIZATION

0.98+

Super Computing Summit 2017EVENT

0.98+

yesterdayDATE

0.98+

Dell EMCORGANIZATION

0.98+

UberORGANIZATION

0.98+

IntelORGANIZATION

0.98+

HBCORGANIZATION

0.97+

RaviPERSON

0.97+

about 20 billion devicesQUANTITY

0.97+

end of 2017DATE

0.97+

I32COMMERCIAL_ITEM

0.97+

threeQUANTITY

0.96+

CUBEORGANIZATION

0.96+

first half of 2018DATE

0.96+

Super Computing 17EVENT

0.95+

Super Computing 2017EVENT

0.95+

Deep Learning Ready BundleCOMMERCIAL_ITEM

0.94+

GMORGANIZATION

0.94+

HadoopTITLE

0.93+

three thingsQUANTITY

0.91+

S&P 500ORGANIZATION

0.91+

SQLTITLE

0.9+

UTORGANIZATION

0.9+

about 1.2 million policiesQUANTITY

0.89+

firstQUANTITY

0.89+

RolexORGANIZATION

0.89+

Hybrid CloudORGANIZATION

0.88+

BlockbusterORGANIZATION

0.87+

TheanoORGANIZATION

0.86+

12QUANTITY

0.86+

IEEEORGANIZATION

0.85+

Bernie Spang, IBM & Wayne Glanfield, Red Bull Racing | Super Computing 2017


 

>> Announcer: From Denver, Colorado it's theCUBE. Covering Super Computing 17, brought to you by Intel. Welcome back everybody, Jeff Frick here with theCUBE. We're at Super Computing 2017 in Denver, Colorado talking about big big iron, we're talking about space and new frontiers, black holes, mapping the brain. That's all fine and dandy, but we're going to have a little bit more fun this next segment. We're excited to have our next guest Bernie Spang. He's a VP Software Defined Infrastructure for IBM. And his buddy and guest Wayne Glanfield HPC Manager for Red Bull Racing. And for those of you that don't know, that's not the pickup trucks, it's not the guy jumping out of space, this is the Formula One racing team. The fastest, most advanced race cars in the world. So gentlemen, first off welcome. Thank you. Thank you Jeff. So what is a race car company doing here for a super computing conference? Obviously we're very interested in high performance computing so traditionally we've used a wind tunnel to do our external aerodynamics. HPC allows us to do many many more iterations, design iterations of the car. So we can actually kind of get more iterations of the designs out there and make the car go faster very quicker. So that's great, you're not limited to how many times you can get it in the wind tunnel. The time you have in the wind tunnel. I'm sure there's all types of restrictions, cost and otherwise. There's lots of restrictions and both the wind tunnel and in HPC usage. So with HPC we're limited to 25 teraflops, which isn't many teraflops. 25 teraflops. >> Wayne: That's all. And Bernie, how did IBM get involved in Formula One racing? Well I mean our spectrum computing offerings are about virtualizing clusters to optimize efficiency, and the performance of the workloads. So our Spectrum LSF offering is used by manufacturers, designers to get ultimate efficiency out of the infrastructure. So with the Formula One restrictions on the teraflops you want to get as much work through that system as efficiently as you can. And that's where Spectrum computing comes in. That's great. And so again, back to the simulations. So not only can you just do simulations 'cause you got the capacity, but then you can customize it as you said I think before we turned on the cameras for specific tracks, specific race conditions. All types of variables that you couldn't do very easily in a traditional wind tunnel. Yes obviously it takes a lot longer to actually kind of develop, create, and rapid prototype the models and get them in the wind tunnel, and actually test them. And it's obviously much more expensive. So by having a HPC facility we can actually kind of do the design simulations in a virtual environment. So what's been kind of the ahah from that? Is it just simply more better faster data? Is there some other kind of transformational thing that you observed as a team when you started doing this type of simulation versus just physical simulation in a wind tunnel? We started using HPC and computational fluid dynamics about 12 years ago in anger. Traditionally it started out as a complementary tool to the wind tunnel. But now with the advances in HPC technology and software from IBM, it's actually beginning to overtake the wind tunnel. So it's actually kind of driving the way we design the car these days. That's great. So Bernie, working with super high end performance, right, where everything is really optimized to get that car to go a little bit faster, just a little bit faster. Right. Pretty exciting space to work in, you know, there's a lot of other great applications, aerospace, genomics, this and that. But this is kind of a fun thing you can actually put your hands on. Oh it's definitely fun, it's definitely fun being with the Red Bull Racing team, and with our clients when we brief them there. But we have commercial clients in automotive design, aeronautics, semiconductor manufacturing, where getting every bit of efficiency and performance out of their infrastructure is also important. Maybe they're not limited by rules, but they're limited by money, you know and the ability to investment. And their ability to get more out of the environment gives them a competitive advantage as well. And really what's interesting about racing, and a lot of sports is you get to witness the competition. We don't get to witness the competition between big companies day to day. You're not kind of watching it in those little micro instances. So the good thing is you get to learn a lot from such a focused, relatively small team as Red Bull Racing that you can apply to other things. So what are some of the learnings as you've got work with them that you've taken back? Well certainly they push the performance of the environment, and they push us, which is a great thing for us, and for our other clients who benefit. But one of the things I think that really stands out is the culture there of the entire team no matter what their role and function. From the driver on down to everybody else are focused on winning races and winning championships. And that team view of getting every bit of performance out of everything everybody does all the time really opened our thinking to being broader than just the scheduling of the IT infrastructure, it's also about making the design team more productive and taking steps out of the process, and anything we can do there. Inclusive of the storage management, and the data management over time. So it's not just the compute environment it's also the virtualized storage environment. Right, and just massive amounts of storage. You said not only are you running and generating, I'm just going to use boatloads 'cause I'm not sure which version of the flops you're going to use. But also you got historical data, and you have result data, and you have models that need to be tweaked, and continually upgraded so that you do better the following race. Exactly, I mean we're generating petabytes of data a year and I think one of the issues which is probably different from most industries is our workflows are incredibly complex. So we have up to 200 discrete job steps for each workflow to actually kind of produce a simulation. This is where the kind of IBM Spectrum product range actually helps us do that efficiently. If you imagine an aerospace engineer, or aerodynamics engineer trying to manually manage 200 individual job steps, it just wouldn't happen very efficiently. So this is where Spectrum scale actually kind of helps us do that. So you mentioned it briefly Bernie, but just a little bit more specifically. What are some of the other industries that you guys are showcasing that are leveraging the power of Spectrum to basically win their races. Yeah so and we talked about the infrastructure and manufacturing, but they're industrial clients. But also in financial services. So think in terms of risk analytics and financial models being an important area. Also healthcare life sciences. So molecular biology, finding new drugs. When you talk about the competition and who wins right. Genomics research and advances there. Again, you need a system and an infrastructure that can chew through vast amounts of data. Both the performance and the compute, as well as the longterm management with cost efficiency of huge volumes of data. And then you need that virtualized cluster so that you can run multiple workloads many times with an infrastructure that's running in 80%, 90% efficiency. You can't afford to have silos of clusters. Right we're seeing clients that have problems where they don't have this cluster virtualization software, have cluster creep, just like in the early days we had server sprawl, right? With a different app on a different server, and we needed to virtualize the servers. Well now we're seeing cluster creep. Right the Hadoop clusters and Spark clusters, and machine learning and deep learning clusters. As well as the traditional HPC workload. So what Spectrum computing does is virtualizes that shared cluster environment so that you can run all these different kind of workloads and drive up the efficiency of the environment. 'Cause efficiency is really the key right. You got to have efficiency that's what, that's really where cloud got its start, you know, kind of eating into the traditional space, right. There's a lot of inefficient stuff out there so you got to use your resources efficiently it's way too competitive. Correct well we're also seeing inefficiencies in the use of cloud, right. >> Jeff: Absolutely. So one of the features that we've added to the Spectrum computing recently is automated dynamic cloud bursting. So we have clients who say that they've got their scientists or their design engineers spinning up clusters in the cloud to run workloads, and then leaving the servers running, and they're paying the bill. So we built in automation where we push the workload and the data over the cloud, start the servers, run the workload. When the workload's done, spin down the servers and bring the data back to the user. And it's very cost effective that way. It's pretty fun everyone talks often about the spin up, but they forget to talk about the spin down. Well that's where the cost savings is, exactly. Alright so final words, Wayne, you know as you look forward, it's super a lot of technology in Formula One racing. You know kind of what's next, where do you guys go next in terms of trying to get another edge in Formula One racing for Red Bull specifically. I mean I'm hoping they reduce the restrictions on HPC so it can actually start using CFD and the software IBM provides in a serious manner. So it can actually start pushing the technologies way beyond where they are at the moment. It's really interesting that they, that as a restriction right, you think of like plates and size of the engine, and these types of things as the rule restrictions. But they're actually restricting based on data size, your use of high performance computing. They're trying to save money basically, but. It's crazy. So whether it's a rule or you know you're share holders, everybody's trying to save money. Alright so Bernie what are you looking at, sort of 2017 is coming to an end, it's hard for me to say that as you look forward to 2018 what are some of your priorities for 2018. Well the really important thing and we're hearing it at this conference, I'm talking with the analysts and with the clients. The next generation of HPC in analytics is what we're calling machine learning, deep learning, cognitive AI, whatever you want to call it. That's just the new generation of this workload. And our Spectrum conductor offering and our new deep learning impact capability to automate the training of deep learning models, so that you can more quickly get to an accurate model like in hours or minutes, not days or weeks. That's going to a huge break through. And based on our early client experience this year, I think 2018 is going to be a breakout year for putting that to work in commercial enterprise use cases. Alright well I look forward to the briefing a year from now at Super Computing 2018. Absolutely. Alright Bernie, Wayne, thanks for taking a few minutes out of your day, appreciate it. You're welcome, thank you. Alright he's Bernie, he's Wayne, I'm Jeff Frick we're talking Formula One Red Bull Racing here at Super Computing 2017. Thanks for watching.

Published Date : Nov 16 2017

SUMMARY :

and new frontiers, black holes, mapping the brain. So the good thing is you get to learn a lot and bring the data back to the user.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
WaynePERSON

0.99+

Jeff FrickPERSON

0.99+

JeffPERSON

0.99+

BerniePERSON

0.99+

Wayne GlanfieldPERSON

0.99+

IBMORGANIZATION

0.99+

90%QUANTITY

0.99+

80%QUANTITY

0.99+

Bernie SpangPERSON

0.99+

2018DATE

0.99+

25 teraflopsQUANTITY

0.99+

Red Bull RacingORGANIZATION

0.99+

2017DATE

0.99+

Denver, ColoradoLOCATION

0.99+

oneQUANTITY

0.98+

Super Computing 17EVENT

0.98+

Super Computing 2017EVENT

0.98+

IntelORGANIZATION

0.97+

each workflowQUANTITY

0.97+

Super Computing 2018EVENT

0.97+

Formula OneEVENT

0.96+

bothQUANTITY

0.96+

BothQUANTITY

0.95+

this yearDATE

0.92+

up to 200 discrete job stepsQUANTITY

0.92+

a yearQUANTITY

0.89+

Formula OneORGANIZATION

0.86+

about 12 years agoDATE

0.86+

firstQUANTITY

0.84+

200 individual job stepsQUANTITY

0.82+

SpectrumOTHER

0.79+

HPCORGANIZATION

0.79+

Red BullEVENT

0.79+

theCUBEORGANIZATION

0.73+

petabytesQUANTITY

0.65+

SparkTITLE

0.61+

HPCPERSON

0.6+

issuesQUANTITY

0.56+

featuresQUANTITY

0.52+

SpectrumCOMMERCIAL_ITEM

0.5+

SpectrumTITLE

0.49+

SpectrumORGANIZATION

0.44+

BernieLOCATION

0.39+

Ravi Pendekanti, Dell EMC | Super Computing 2017


 

>> Narrator: From Denver, Colorado, it's theCUBE. Covering Super Computing '17, brought to you by Intel. Hey welcome back everybody, Jeff Frick here with theCUBE. We're at Super Computing 2017, Denver, Colorado, 12,000 people talking about big iron, big questions, big challenges. It's really an interesting take on computing, really out on the edge. The key note was, literally, light years out in space, talking about predicting the future with quirks and all kinds of things, a little over my head for sure. But we're excited to kind of get back to the ground and we have Ravi Pendekanti. He's the Senior Vice President of Product Management and Marketing, Server Platforms, Dell EMC. It's a mouthful, Ravi great to see you. Great to see you too Jeff and thanks for having me here. Absolutely, so we were talking before we turned the cameras on. One of your big themes, which I love, is kind of democratizing this whole concept of high performance computing, so it's not just the academics answering the really, really, really big questions. You're absolutely right. I mean think about it Jeff, 20 years ago, even 10 years ago, when people talk about high performance computing, it was what I call as being in the back alleys of research and development. There were a few research scientists working on it, but we're at a time in our journey towards helping humanity in a bigger way. The HPC has found it's way into almost every single mainstream industry you can think of. Whether it is fraud detection, you see MasterCard is using it for ensuring that they can see and detect any of the fraud that can be committed earlier than the perpetrators come in and actually hack the system. Or if you get into life sciences, if you talk about genomics. I mean this is what might be good for our next set of generations, where they can probably go out and tweak some of the things in a genome sequence so that we don't have the same issues that we have had in the past. Right. Right? So, likewise, you can pick any favorite industry. I mean we are coming up to the holiday seasons soon. I know a lot of our customers are looking at how do they come up with the right schema to ensure that they can stock the right product and ensure that it is available for everyone at the right time? 'Cause timing is important. I don't think any kid wants to go with no toy and have the product ship later. So bottom line is, yes, we are looking at ensuring the HPC reaches every single industry you can think of. So how do you guys parse HPC verses a really big virtualized cluster? I mean there's so many ways that compute and store has evolved, right? So now, with cloud and virtual cloud and private cloud and virtualization, you know, I can pull quite a bit of horsepower together to attack a problem. So how do you kind of cut the line between Navigate, yeah. big, big compute, verses true HPC? HPC. It's interesting you ask. I'm actually glad you asked because people think that it's just feeding CPU or additional CPU will do the trick, it doesn't. The simple fact is, if you look at the amount of data that is being created. I'll give you a simple example. I mean, we are talking to one of the airlines right now, and they're interested in capturing all the data that comes through their flights. And one of the things they're doing is capturing all the data from their engines. 'Cause end of the day, you want to make sure that your engines are pristine as they're flying. And every hour that an engine flies out, I mean as an airplane flies out, it creates about 20 terabytes of data. So, if you have a dual engine, which is what most flights are. In one hour they create about 40 terabytes of data. And there are supposedly about 38,000 flights taking off at any given time around the world. I mean, it's one huge data collection problem. Right? I mean, I'm told it's like a real Godzilla number, so I'll let you do the computation. My point is if you really look at the data, data has no value, right? What really is important is getting information out of it. The CPU on the other side has gone to a time and a phase where it is hitting the, what I call as the threshold of the Moore's law. Moore's law was all about performance doubles every two years. But today, that performance is not sufficient. Which is where auxiliary technologies need to be brought in. This is where the GPUs, the FBGAs. Right, right. Right. So when you think about these, that's where the HPC world takes off, is you're augmenting your CPUs and your processors with additional auxiliary technology such as the GPUs and FBGAs to ensure that you have more juice to go do this kind of analytics and the massive amounts of data that you and I and the rest of the humanity is creating. It's funny that you talk about that. We were just at a Western Digital event a little while ago, talking about the next generation of drives and it was the same thing where now it's this energy assist method to change really the molecular way that it saves information to get more out of it. So that's kind of how you parse it. If you've got to juice the CPU, and kind of juice the traditional standard architecture, then you're moving into the realm of high performance computing. Absolutely, I mean this is why, Jeff, yesterday we launched a new PowerEdge C4140, right? The first of it's kind in terms of the fact that it's got two Intel Xeon processors, but beyond that, it also can support four Nvidia GPUs. So now you're looking at a server that's got both the CPUs, to your earlier comment on processors, but is augmented by four of the GPUs, that gives immense capacity to do this kind of high performance computing. But as you said, it's not just compute, it's store, it's networking, it's services, and then hopefully you package something together in a solution so I don't have to build the whole thing from scratch, you guys are making moves, right? Oh, this is a perfect lead in, perfect lead in. I know, my colleague, Armagon will be talking to you guys shortly. What his team does, is it takes all the building blocks we provide, such as the servers, obviously looks at the networking, the storage elements, and then puts them together to create what are called solutions. So if you've got solutions, which enable our customers to go back in and easily deploy a machine-learning or a deep-learning solution. Where now our customers don't have to do what I call as the heavy lift. In trying to make sure that they understand how the different pieces integrate together. So the goal behind what we are doing at Dell EMC is to remove the guess work out so that our customers and partners can go out and spend their time deploying the solution. Whether it is for machine learning, deep learning or pick your favorite industry, we can also verticalize it. So that's the beauty of what we are doing at Dell EMC. So the other thing we were talking about before we turned turned the cameras on is, I call them the itys from my old Intel days, reliability, sustainability, serviceability, and you had a different phrase for it. >> Ravi: Oh yes, I know you're talking about the RAS. The RAS, right. Which is the reliability, availability, and serviceability. >> Jeff: But you've got a new twist on it. Oh we do. Adding something very important, and we were just at a security show early this week, CyberConnect, and security now cuts through everything. Because it's no longer a walled garden, 'cause there are no walls. There are no walls. It's really got to be baked in every layer of the solution. Absolutely right. The reason is, if you really look at security, it's not about, you know till a few years ago, people used to think it's all about protecting yourself from external forces, but today we know that 40% of the hacks happen because of the internal, you know, system processes that we don't have in place. Or we could have a person with an intent to break in for whatever reason, so the integrated security becomes part and parcel of what we do. This is where, with in part of a 14G family, one of the things we said is we need to have integrated security built in. And along with that, we want to have the scalability because no two workloads are the same and we all know that the amount of data that's being created today is twice what it was the last year for each of us. Forget about everything else we are collecting. So when you think about it, we need integrated security. We need to have the scalability feature set, also we want to make sure there is automation built in. These three main tenets that we talked about feed into what we call internally, the monic of a user's PARIS. And that's what I think, Jeff, to our earlier conversation, PARIS is all about, P is for best price performance. Anybody can choose to get the right performance or the best performance, but you don't want to shell out a ton of dollars. Likewise, you don't want to pay minimal dollars and try and get the best performance, that's not going to happen. I think there's a healthy balance between price performance, that's important. Availability is important. Interoperability, as much as everybody thinks that they can act on their own, it's nearly impossible, or it's impossible that you can do it on your own. >> Jeff: These are big customers, they've got a lot of systems. You are. You need to have an ecosystem of partners and technologies that come together and then, end of the day, you have to go out and have availability and serviceability, or security, to your point, security is important. So PARIS is about price performance, availability, interoperability, reliability, availability and security. I like it. That's the way we design it. It's much sexier than that. We drop in, like an Eiffel Tower picture right now. There you go, you should. So Ravi, hard to believe we're at the end of 2017, if we get together a year from now at Super Computing 2018, what are some of your goals, what are your some objectives for 2018? What are we going to be talking about a year from today? Oh, well looking into a crystal ball, as much as I can look into that, I thin that-- >> Jeff: As much as you can disclose. And as much as we can disclose, a few things I think are going to happen. >> Jeff: Okay. Number one, I think you will see people talk about to where we started this conversation. HPC has become mainstream, we talked about it, but the adoption of high performance computing, in my personal belief, is not still at a level that it needs to be. So, if you go down next 12 to 18 months, lets say, I do think the adoption rates will be much higher than where we are. And we talk about security now, because it's a very topical subject, but as much as we are trying to emphasize to our partners and customers that you've got to think about security from ground zero. We still see a number of customers who are not ready. You know, some of the analysis show there are nearly 40% of the CIOs are not ready in helping and they truly understand, I should say, what it takes to have a secure system and a secure infrastructure. It's my humble belief that people will pay attention to it and move the needle on it. And we talked about, you know, four GPUs in our C4140, do anticipate that there will be a lot more of auxiliary technology packed into it. Sure, sure. So that's essentially what I can say without spilling the beans too much. Okay, all right, super. Ravi, thanks for taking a couple of minutes out of your day, appreciate it. = Thank you. All right, he's Ravi, I'm Jeff Frick, you're watching theCUBE from Super Computing 2017 in Denver, Colorado. Thanks for watching. (techno music)

Published Date : Nov 16 2017

SUMMARY :

and the massive amounts of data that you and I Which is the reliability, because of the internal, you know, and then, end of the day, you have to go out Jeff: As much as you can disclose. And we talked about, you know, four GPUs in our C4140,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
JeffPERSON

0.99+

Jeff FrickPERSON

0.99+

Ravi PendekantiPERSON

0.99+

40%QUANTITY

0.99+

RaviPERSON

0.99+

PARISORGANIZATION

0.99+

2018DATE

0.99+

one hourQUANTITY

0.99+

Dell EMCORGANIZATION

0.99+

12,000 peopleQUANTITY

0.99+

MasterCardORGANIZATION

0.99+

C4140COMMERCIAL_ITEM

0.99+

NvidiaORGANIZATION

0.99+

twiceQUANTITY

0.99+

eachQUANTITY

0.99+

Denver, ColoradoLOCATION

0.99+

bothQUANTITY

0.99+

ArmagonORGANIZATION

0.99+

last yearDATE

0.99+

todayDATE

0.99+

about 20 terabytesQUANTITY

0.99+

Denver,LOCATION

0.98+

oneQUANTITY

0.98+

IntelORGANIZATION

0.98+

yesterdayDATE

0.98+

about 38,000 flightsQUANTITY

0.98+

early this weekDATE

0.98+

PowerEdgeCOMMERCIAL_ITEM

0.97+

endDATE

0.97+

Eiffel TowerLOCATION

0.97+

10 years agoDATE

0.97+

nearly 40%QUANTITY

0.96+

twoQUANTITY

0.95+

20 years agoDATE

0.95+

18 monthsQUANTITY

0.95+

three main tenetsQUANTITY

0.94+

firstQUANTITY

0.94+

fourQUANTITY

0.93+

Super Computing '17EVENT

0.92+

OneQUANTITY

0.92+

every two yearsQUANTITY

0.92+

Super Computing 2017EVENT

0.91+

12QUANTITY

0.89+

2017DATE

0.88+

few years agoDATE

0.86+

MoorePERSON

0.86+

XeonCOMMERCIAL_ITEM

0.85+

Western DigitalORGANIZATION

0.84+

about 40 terabytes of dataQUANTITY

0.83+

Super Computing 2018EVENT

0.82+

two workloadsQUANTITY

0.81+

dualQUANTITY

0.76+

a yearQUANTITY

0.74+

a ton of dollarsQUANTITY

0.74+

14GORGANIZATION

0.7+

singleQUANTITY

0.66+

every hourQUANTITY

0.65+

ground zeroQUANTITY

0.64+

HPCORGANIZATION

0.6+

ColoradoLOCATION

0.56+

doublesQUANTITY

0.56+

PORGANIZATION

0.54+

CyberConnectORGANIZATION

0.49+

theCUBEORGANIZATION

0.49+

RASOTHER

0.34+

Susan Bobholz, Intel | Super Computing 2017


 

>> [Announcer] From Denver, Colorado, it's the Cube covering Super Computing 17, brought to you by Intel. (techno music) >> Welcome back, everybody, Jeff Frick with the Cube. We are at Super Computing 2017 here in Denver, Colorado. 12,000 people talking about big iron, heavy lifting, stars, future mapping the brain, all kinds of big applications. We're here, first time ever for the Cube, great to be here. We're excited for our next guest. She's Susan Bobholtz, she's the Fabric Alliance Manager for Omni-Path at Intel, Susan, welcome. >> Thank you. >> So what is Omni-Path, for those that don't know? >> Omni-Path is Intel's high performance fabric. What it does is it allows you to connect systems and make big huge supercomputers. >> Okay, so for the royal three-headed horsemen of compute, store, and networking, you're really into data center networking, connecting the compute and the store. >> Exactly, correct, yes. >> Okay. How long has this product been around? >> We started shipping 18 months ago. >> Oh, so pretty new? >> Very new. >> Great, okay and target market, I'm guessing has something to do with high performance computing. >> (laughing) Yes, our target market is high performance computing, but we're also seeing a lot of deployments in artificial intelligence now. >> Okay and so what's different? Why did Intel feel compelled that they needed to come out with a new connectivity solution? >> We were getting people telling us they were concerned that the existing solutions were becoming too expensive and weren't going to scale into the future, so they said Intel, can you do something about it, so we did. We made a couple of strategic acquisitions, we combined that with some of our own IP and came up with Omni-Path. Omni-Path is very much a proprietary protocol, but we use all the same software interfaces as InfiniBand, so your software applications just run. >> Okay, so to the machines it looks like InfiniBand? >> Yes. >> Just plug and play and run. >> Very much so, it's very similar. >> Okay what are some of the attributes that make it so special? >> The reason it's really going very well is that it's the price performance benefits, so we have equal to, or better, performance than InfiniBand today, but we also have our switch technology is 48 ports verses InfiniBand is 36 ports. So that means you can build denser clusters in less space and less cables, lower power, total cost of ownership goes down, and that's why people are buying it. >> Really fits into the data center strategy that Intel's executing very aggressively right now. >> Fits very nicely, absolutely, yes, very much so. >> Okay, awesome, so what are your thoughts here at the show? Any announcements, anything that you've seen that's of interest? >> Oh yeah, so, a couple things. We've had really had good luck on the Top 500 list. 60% of the servers that are running a 100 gigabyte fabrics in the Top 500 list are running connected via Omni-Path. >> What percentage again? >> 60% >> 60? >> Yes. >> You've only been at it for 18 months? >> Yes, exactly. >> Impressive. >> Very, very good. We've got systems in the Top 10 already. Some of the Top 10 systems in the world are using Omni-Path. >> Is it rip and replace, do you find, or these are new systems that people are putting in. >> Yeah, these are new systems. Usually when somebody's got a system they like and run, they don't want to touch it. >> Right. >> These are people saying I need a new system. I need more power, I need more oompf. They have the money, the budget, they want to put in something new, and that's when they look to Omni-Path. >> Okay, so what are you working on now, what's kind of next for Omni-Path? >> What's next for us is we are announcing a new higher, denser switch technology, so that will allow you to go for your director class switches, which is the really big ones, is now rather than having 768 ports, you can go to 1152, and that means, again, denser topologies, lower power, less cabling, it reduces your total cost of ownership. >> Right, I think you just answered my question, but I'm going to ask you anyway. >> (laughs) Okay. >> We talked a little bit before we turned the camera on about AI and some of the really unique challenges of AI, and that was part of the motivation behind this product. So what are some of the special attributes of AI that really require this type of connectivity? >> It's very much what you see even with high performance computing. You need low latency, you need high bandwidth. It's the same technologies, and in fact, in a lot of cases, it's the same systems, or sometimes they can be running software load that is HPC focused, and sometimes they're running a software load that is artificial intelligence focused. But they have the same exact needs. >> Okay. >> Do it fast, do it quick. >> Right, right, that's why I said you already answered the question. Higher density, more computing, more storing, faster. >> Exactly, right, exactly. >> And price performance. All right, good, so if we come back a year from now for Super Computing 2018, which I guess is in Dallas in November, they just announced. What are we going to be talking about, what are some of your priorities and the team's priorities as you look ahead to 2018? >> Oh we're continuing to advance the Omni-Path technology with software and additional capabilities moving forward, so we're hoping to have some really cool announcements next year. >> All right, well, we'll look forward to it, and we'll see you in Dallas in a year. >> Thanks, Cube. >> All right, she's Susan, and I'm Jeff. You're watching the Cube from Super Computing 2017. Thanks for watching, see ya next time. (techno music)

Published Date : Nov 15 2017

SUMMARY :

covering Super Computing 17, brought to you by Intel. She's Susan Bobholtz, she's the Fabric Alliance Manager What it does is it allows you to connect systems Okay, so for the royal three-headed horsemen Okay. has something to do with high performance computing. in artificial intelligence now. so they said Intel, can you do something So that means you can build denser clusters Really fits into the data center strategy in the Top 500 list are running connected via Omni-Path. Some of the Top 10 systems in the world are using Omni-Path. Is it rip and replace, do you find, and run, they don't want to touch it. They have the money, the budget, so that will allow you to go for your director class but I'm going to ask you anyway. about AI and some of the really unique challenges of AI, It's very much what you see you already answered the question. and the team's priorities as you look ahead to 2018? moving forward, so we're hoping to have and we'll see you in Dallas in a year. All right, she's Susan, and I'm Jeff.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Susan BobholtzPERSON

0.99+

Jeff FrickPERSON

0.99+

JeffPERSON

0.99+

Susan BobholzPERSON

0.99+

DallasLOCATION

0.99+

18 monthsQUANTITY

0.99+

NovemberDATE

0.99+

SusanPERSON

0.99+

2018DATE

0.99+

36 portsQUANTITY

0.99+

60%QUANTITY

0.99+

12,000 peopleQUANTITY

0.99+

CubePERSON

0.99+

next yearDATE

0.99+

100 gigabyteQUANTITY

0.99+

IntelORGANIZATION

0.99+

Denver, ColoradoLOCATION

0.99+

48 portsQUANTITY

0.99+

768 portsQUANTITY

0.99+

60QUANTITY

0.98+

first timeQUANTITY

0.97+

CubeCOMMERCIAL_ITEM

0.97+

18 months agoDATE

0.97+

Super Computing 2017EVENT

0.96+

todayDATE

0.92+

InfiniBandTITLE

0.91+

Top 10QUANTITY

0.91+

1152QUANTITY

0.91+

Super Computing 17EVENT

0.91+

Top 10 systemsQUANTITY

0.85+

a yearQUANTITY

0.82+

three-headedQUANTITY

0.8+

PathOTHER

0.79+

Super ComputingEVENT

0.76+

TopQUANTITY

0.72+

Omni-PathTITLE

0.72+

Omni-PathOTHER

0.72+

Omni-PathCOMMERCIAL_ITEM

0.71+

OmniTITLE

0.59+

OmniORGANIZATION

0.58+

Omni-PathORGANIZATION

0.57+

coupleQUANTITY

0.5+

-PathOTHER

0.49+

PathORGANIZATION

0.3+

500OTHER

0.29+

Jim Wu, Falcon Computing | Super Computing 2017


 

>> Announcer: From Denver, Colorado, it's theCUBE covering Super Computing '17. Brought to you by Intel. (upbeat techno music) Hey welcome back, everybody. Jeff Frick here with theCUBE. We're at Super Computing 2017 in Denver, Colorado. It's our first trip to the show, 12,000 people, a lot of exciting stuff going on, big iron, big lifting, heavy duty compute. We're excited to have our next guest on. He's Jim Wu, he's the Director of Customer Experience for Falcon Computing. Jim, welcome. Thank you. Good to see you. So, what does Falcon do for people that aren't familiar with the company? Yeah, Falcon Company is in our early stages startup, focus on AVA-based acceleration development. Our vision is to allow software engineers to develop a FPGA-based accelerators, accelerators without FPGA expertise. Right, you just said you closed your B round. So, congratulations on that. >> Jim: Thank you. Yeah, very exciting. So, it's a pretty interesting concept. To really bring the capability to traditional software engineers to program for hardware. That's kind of a new concept. What do you think? 'Cause it brings the power of a hardware system. but the flexibility of a software system. Yeah, so today, to develop FPGA accelerators is very challenging. So, today for the accelerations-based people use very low level language, like a Verilog and the VHDL to develop FPGA accelerators. Which was very time consuming, very labor-intensive. So, our goal is to liberate them to use, C/C++ space design flow to give them an environment that they are familiar with in C/C++. So now not only can they improve their productivity, we also do a lot of automatic organization under the hood, to give them the highest accelerator results. Right, so that really opens up the ecosystem well beyond the relatively small ecosystem that knows how to program their hardware. Definitely, that's what we are hoping to see. We want to the tool in the hands of all software programmers. They can use it in the Cloud. They can use it on premises. Okay. So what's the name of your product? And how does it fit within the stack? I know we've got the Intel microprocessor under the covers, we've got the accelerator, we've got the cards. There's a lot of pieces to the puzzle. >> Jim: Yeah. So where does Falcon fit? So our main product is a compiler, called the Merlin Compiler. >> Jeff: Okay. It's a pure C and the C++ flow that enables software programmers to design APGA-based accelerators without any knowledge of APGA. And it's highly integrated with Intel development tools. So users don't even need to learn anything about the Intel development environment. They can just use their C++ development environment. Then in the end, we give them the host code as well as APGA binaries so they can round on APGA to see a accelerated applications. Okay, and how long has Merlin been GA? Actually, we'll be GA early next year. Early next year. So finishing, doing the final polish here and there. Yes. So in this quarter, we are heavily investing a lot of ease-of-use features. Okay. We have most of the features we want to be in the tool, but we're still lacking a bit in terms of ease-of-use. >> Jeff: Okay. So we are enhancing our report capabilities, we are enhancing our profiling of capabilities. We want to really truly like a traditional C++-based development environment for software application engineers. Okay, that's fine. You want to get it done, right, before you ship it out the door? So you have some Alpha programs going on? Some Beta programs of some really early adopters? Yeah, exactly. So today we provide a 14 day free trial to any customers who are interested. We have it, you can set up your enterprise or you can set up on Cloud. Okay. We provide to where you want your work done. Okay. And so you'll support all the cloud service providers, the big public clouds, all the private clouds. All the traditional data servers as well. Right. So, we are twice already on Aduplas as well as Alibaba Cloud. So we are working on bringing the tool to other public cloud providers as well. Right. So what is some of the early feedback you're getting from some of the people you're talking to? As to where this is going to make the biggest impact. What type of application space has just been waiting for this solution? So our Merlin Compiler is a productivity tool, so any space that FPGA can traditionally play well that's where we want to be there. So like encryption, decryption, video codec, compression, decompression. Those kind of applications are very stable for APGA. Now traditionally they can only be developed by hardware engineers. Now with the Merlin Compiler, all of these software engineers can use the Merlin Compiler to do all of these applications. Okay. And when is the GA getting out, I know it's coming? When is it coming? Approximately So probably first quarter of 2018. Okay, that's just right around the corner. Exactly. Alright, super. And again, a little bit about the company, how many people are you? A little bit of the background on the founders. So we have about 30 employees, at the moment, so we have offices in Santa Clara which is our headquarters. We also have an office in Los Angeles. As well as a Beijing, China. Okay, great. Alright well Jim, thanks for taking a few minutes. We'll be looking for GA in a couple of months and wish you nothing but the best success. Okay, thank you so much, Jeff. Alright, he's Jim Lu I'm Jeff Frick. You're watching theCUBE from supering computing 2017. Thanks for watching. (upbeat techno music)

Published Date : Nov 14 2017

SUMMARY :

Brought to you by Intel. Verilog and the VHDL to develop FPGA accelerators. called the Merlin Compiler. We have most of the features we want to be in the tool, We provide to where you want your work done.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Jim WuPERSON

0.99+

JimPERSON

0.99+

JeffPERSON

0.99+

Jeff FrickPERSON

0.99+

Santa ClaraLOCATION

0.99+

BeijingLOCATION

0.99+

Los AngelesLOCATION

0.99+

14 dayQUANTITY

0.99+

todayDATE

0.99+

FalconORGANIZATION

0.99+

first quarter of 2018DATE

0.99+

12,000 peopleQUANTITY

0.99+

Denver, ColoradoLOCATION

0.99+

twiceQUANTITY

0.99+

first tripQUANTITY

0.99+

C++TITLE

0.99+

Early next yearDATE

0.98+

IntelORGANIZATION

0.98+

Super Computing '17EVENT

0.98+

early next yearDATE

0.98+

2017DATE

0.98+

GALOCATION

0.97+

Jim LuPERSON

0.97+

Falcon CompanyORGANIZATION

0.97+

about 30 employeesQUANTITY

0.97+

Super Computing 2017EVENT

0.97+

APGATITLE

0.94+

this quarterDATE

0.94+

theCUBEORGANIZATION

0.94+

CTITLE

0.92+

AduplasORGANIZATION

0.91+

C/C+TITLE

0.9+

C+TITLE

0.87+

Alibaba CloudORGANIZATION

0.84+

APGAORGANIZATION

0.82+

Falcon ComputingORGANIZATION

0.81+

ChinaLOCATION

0.76+

MerlinTITLE

0.71+

Merlin CompilerTITLE

0.65+

MerlinORGANIZATION

0.64+

FPGAORGANIZATION

0.62+

SuperEVENT

0.61+

GAORGANIZATION

0.61+

VerilogTITLE

0.54+

John Lockwood, Algo Logic Systems | Super Computing 2017


 

>> Narrator: From Denver, Colorado, it's theCUBE. Covering Super Computing '17, brought to you by Intel. (electronic music) >> Hey, welcome back everybody. Jeff Frick here with theCUBE. We're at Denver, Colorado at Super Computing 2017. 12,000 people, our first trip to the show. We've been trying to come for awhile, it's pretty amazing. A lot of heavy science in terms of the keynotes. All about space and looking into brain mapping and it's heavy lifting, academics all around. We're excited to have our next guest, who's an expert, all about speed and that's John Lockwood. He's the CEO of Algo-Logic. First off, John, great to see you. >> Yeah, thanks Jeff, glad to be here. >> Absolutely, so for folks that aren't familiar with the company, give them kind of the quick overview of Algo. >> Yes, Algo-Logic puts algorithms into logic. So our main focus is taking things are typically done in software and putting them into FPGAs and by doing that we make them go faster. >> So it's a pretty interesting phenomenon. We've heard a lot from some of the Intel execs about kind of the software overlay that now, kind of I guess, a broader ecosystem of programmers into hardware, but then still leveraging the speed that you get in hardware. So it's a pretty interesting combination to get those latencies down, down, down. >> Right, right, I mean Intel certainly made a shift to go on into heterogeneous compute. And so in this heterogeneous world, we've got software running on Xeons, Xeon Phis. And we've also got the need though, to use new compute in more than just the traditional microprocessor. And so with the acquisition of Altera, is that now Intel customers can use FPGAs in order to get the benefit in speed. And so Algo-Logic, we typically provide applications with software APIs, so it makes it really easy for end customers to deploy FPGAs into their data center, into their hosts, into their network and start using them right away. >> And you said one of your big customer sets is financial services and trading desk. So low latency there is critical as millions and millions and millions if not billions of dollars. >> Right, so Algo-Logic we have a whole product line of high-frequency trading systems. And so our Tick-To-Trade system is unique in the fact that it has a sub-microsecond trading latency and this means going from market data that comes in, for example on CME for options and futures trading, to time that we can place a fix order back out to the market. All of that happens in an FPGA. That happens in under a microsecond. So under a millionth of second and that beats every other software system that's being used. >> Right, which is a game change, right? Wins or losses can be made on those time frames. >> It's become a must have is that if you're trading on Wall Street or trading in Chicago and you're not trading with an FPGA, you're trading at a severe disadvantage. And so we make a product that enables all the trading firms to be playing on a fair, level playing field against the big firms. >> Right, so it's interesting because the adoption of Flash and some of these other kind of speed accelerator technologies that have been happening over the last several years, people are kind of getting accustomed to the fact that speed is better, but often it was kind of put aside in this kind of high-value applications like financial services and not really proliferating to a broader use of applications. I wonder if you're seeing that kind of change a little bit, where people are seeing the benefits of real time and speed beyond kind of the classic high-value applications? >> Well, I think the big change that's happened is that it's become machine-to-machine now. And so humans, for example in trading, are not part of the loop anymore and so it's not a matter of am I faster than another person? It's am I faster than the other person's machine? And so this notion of having compute that goes fast has become suddenly dramatically much more important because everything now is going to machine versus machine. And so if you're an ad tech advertiser, is that how quickly you can do an auction to place an ad matters and if you can get a higher value ad placed because you're able to do a couple rounds of an auction, that's worth a lot. And so, again, with Algo-Logic we make things go faster and that time benefit means, that all thing else being the same, you're the first to come to a decision. >> Right, right and then of course the machine-to-machine obviously brings up the hottest topic that everybody loves to talk about is autonomous vehicles and networked autonomous vehicles and just the whole IOT space with the compute moving out to the edge. So this machine-to-machine systems are only growing in importance and really percentage of the total compute consumption by far. >> That's right, yeah. So last year at Super Computing, we demonstrated a drone, bringing in realtime data from a drone. So doing realtime data collection and doing processing with our Key Value Store. So this year, we have a machine learning application, a Markov Decision Process where we show that we can scale-out a machine learning process and teach cars how to drive in a few minutes. >> Teach them how to drive in a few minutes? >> Right. >> So that's their learning. That's not somebody programming the commands. They're actually going through a process of learning? >> Right, well so the Key Value Store is just a part of this. We're just the part of the system that makes the scale-outs that runs well in a data center. And so we're still running the Markov Decision Process in simulations in software. So we have a couple Xeon servers that we brought with us to do the machine learning and a data center would scale-out to be dozens of racks, but even with a few machines though, for simple highway driving, what we can show is we start off with, the system's untrained and that in the Markov Decision Process, we reward the final state of not having accidents. And so at first, the cars drive and they're bouncing into each other. It's like bumper cars, but within a few minutes and after about 15 million simulations, which can be run that quickly, is that the cars start driving better than humans. And so I think that's a really phenomenal step, is the fact that you're able to get to a point where you can train a system how to drive and give them 15 man years of experience in a matter of minutes by the scale-out compute systems. >> Right, 'cause then you can put in new variables, right? You can change that training and modify it over time as conditions change, throw in snow or throw in urban environments and other things. >> Absolutely, right. And so we're not pretending that our machine learning, that application we're showing here is an end-all solution. But as you bring in other factors like pedestrians, deer, other cars running different algorithms or crazy drivers, is that you want to expose the system to those conditions as well. And so one of the questions that came up to us was, "What machine learning application are you running?" So we're showing all 25 cars running one machine learned application and that's incrementally getting better as they learn to drive, but we could also have every car running a different machine learning application and see how different AIs interact with each other. And I think that's what you're going to see on the highway as we have more self-driving cars running different algorithms, we have to make sure they all place nice with each other. >> Right, but it's really a different way of looking at the world, right, using machine learning, machine-to-machine versus single person or a team of people writing a piece of software to instruct something to do something and then you got to go back and change it. This is a much more dynamic realtime environment that we're entering into with IOT. >> Right, I mean the machine-to-human, which was kind of last year and years before, were, "How do you make interactions "between the computers better than humans?" But now it's about machine-to-machine and it's,"How do you make machines interact better "with other machines?" And that's where it gets really competitive. I mean, you can imagine with drones for example, for applications where you have drones against drones, the drones that are faster are going to be the ones that win. >> Right, right, it's funny, we were just here last week at the commercial drone show and it's pretty interesting how they're designing the drones now into a three-part platform. So there's the platform that flies around. There's the payload, which can be different sensors or whatever it's carrying, could be herbicide if it's an agricultural drone. And then they've opened up the STKs, both on the control side as well as the mobile side, in terms of the controls. So it's a very interesting way that all these things now, via software could tie together, but as you say, using machine learning you can train them to work together even better, quicker, faster. >> Right, I mean having a swarm or a cluster of these machines that work with each other, you could really do interesting things. >> Yeah, that's the whole next thing, right? Instead of one-to-one it's many-to-many. >> And then when swarms interact with other swarms, then I think that's really fascinating. >> So alright, is that what we're going to be talking about? So if we connect in 2018, what are we going to be talking about? The year's almost over. What are your top priorities for next year? >> Our top priorities are to see. We think that FPGA is just playing this important part. A GPU for example, became a very big part of the super computing systems here at this conference. But the other side of heterogeneous is the FPGA and the FPGA has seen almost, just very minimal adoption so far. But the FPGA has the capability of providing, especially when it comes to doing network IO transactions, it's speeding up realtime interactions, it has an ability to change the world again for HPC. And so I'm expecting that in a couple years, at this HPC conference, that what we'll be talking about, is the biggest top 500 super computers, is that how big of FPGAs do they have. Not how big of GPUs do they have. >> All right, time will tell. Well, John, thanks for taking a few minutes out of your day and stopping by. >> Okay, thanks Jeff, great to talk to you. >> All right, he's John Lockwood, I'm Jeff Frick. You're watching theCUBE from Super Computing 2017. Thanks for watching. >> Bye. (electronic music)

Published Date : Nov 14 2017

SUMMARY :

Covering Super Computing '17, brought to you by Intel. A lot of heavy science in terms of the keynotes. that aren't familiar with the company, and by doing that we make them go faster. still leveraging the speed that you get in hardware. And so with the acquisition of Altera, And you said one of your big customer sets Right, so Algo-Logic we have a whole product line Right, which is a game change, right? And so we make a product that enables all the trading firms Right, so it's interesting because the adoption of Flash And so this notion of having compute that goes fast and just the whole IOT space and teach cars how to drive in a few minutes. That's not somebody programming the commands. and that in the Markov Decision Process, Right, 'cause then you can put in new variables, right? And so one of the questions that came up to us was, of looking at the world, right, using machine learning, Right, I mean the machine-to-human, in terms of the controls. you could really do interesting things. Yeah, that's the whole next thing, right? And then when swarms interact with other swarms, So alright, is that what we're going to be talking about? And so I'm expecting that in a couple years, All right, time will tell. All right, he's John Lockwood, I'm Jeff Frick. (electronic music)

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
JeffPERSON

0.99+

John LockwoodPERSON

0.99+

ChicagoLOCATION

0.99+

Jeff FrickPERSON

0.99+

JohnPERSON

0.99+

2018DATE

0.99+

millionsQUANTITY

0.99+

25 carsQUANTITY

0.99+

last weekDATE

0.99+

Algo-LogicORGANIZATION

0.99+

last yearDATE

0.99+

billions of dollarsQUANTITY

0.99+

12,000 peopleQUANTITY

0.99+

Wall StreetLOCATION

0.99+

Denver, ColoradoLOCATION

0.99+

AlteraORGANIZATION

0.99+

next yearDATE

0.99+

this yearDATE

0.99+

Algo Logic SystemsORGANIZATION

0.99+

first tripQUANTITY

0.98+

under a microsecondQUANTITY

0.98+

oneQUANTITY

0.98+

IntelORGANIZATION

0.98+

dozens of racksQUANTITY

0.98+

firstQUANTITY

0.98+

Super Computing 2017EVENT

0.97+

FirstQUANTITY

0.97+

bothQUANTITY

0.97+

under a millionth of secondQUANTITY

0.96+

500 super computersQUANTITY

0.96+

Super Computing '17EVENT

0.94+

15 man yearsQUANTITY

0.94+

about 15 million simulationsQUANTITY

0.93+

three-part platformQUANTITY

0.89+

minutesQUANTITY

0.88+

XeonORGANIZATION

0.84+

theCUBEORGANIZATION

0.82+

single personQUANTITY

0.78+

one ofQUANTITY

0.75+

last several yearsDATE

0.74+

Key Value StoreORGANIZATION

0.72+

coupleQUANTITY

0.63+

couple yearsQUANTITY

0.61+

FlashTITLE

0.61+

yearsDATE

0.55+

Xeon PhisCOMMERCIAL_ITEM

0.51+

machinesQUANTITY

0.5+

questionsQUANTITY

0.5+

Value StoreORGANIZATION

0.49+

KeyTITLE

0.47+

XeonsORGANIZATION

0.4+

MarkovORGANIZATION

0.39+

CMETITLE

0.39+

Stephane Monoboisset, Accelize | Super Computing 2017


 

>> Voiceover: From Denver, Colorado, it's theCUBE covering Super Computing '17, brought to you by Intel. Hey, welcome back, everybody. Jeff Frick, here, with theCUBE. We're in Denver, Colorado at Super Computing 2017. It's all things heavy lifting, big iron, 12,000 people. I think it's the 20th anniversary of the conference. A lot of academics, really talking about big iron, doin' big computing. And we're excited to have our next guest, talking about speed, he's Stephane Monoboisset. Did I get that right? That's right. He's a director of marketing and partnerships for Accelize. Welcome. Thank you. So, for folks that aren't familiar with Accelize, give them kind of the quick overview. Okay, so Accelize is a French startup. Actually, a spinoff for a company called PLDA that has been around for 20 years, doing PCI express IP. And about a few years ago, we started initiative to basically bring FPGA acceleration to the cloud industry. So what we say is, we basically enable FPGA acceleration as a service. So did it not exist in cloud service providers before that, or what was kind of the opportunity that you saw there? So, FPGAs have been used in data centers in many different ways. They're starting to make their way into, as a service type of approach. But one of the thing that the industry, one of the buzzword that the industry's using, is FPGA as a service. And the industry usually refers to it as the way to bring FPGA to the end users. But when you think about it, end users don't really want FPGA as a service. Most of the cloud end users are not FPGA experts. So they couldn't care less whether it's an FPGA or something else. What they really want is the acceleration benefits. Hence the term, FPGA acceleration as a service. So, in order to do that, instead of just going and offering an FPGA platform, and giving them the tools, even if they are easy to use and develop the FPGAs, our objective is to propose to provide a marketplace of accelerators that they can use as a service, without even thinking that it's an FPGA on the background. So that's a really interesting concept. Because that also leverages an ecosystem. And one thing we know that's important, if you have any kind of a platform playing, you need an ecosystem that brings a much broader breadth of applications, and solution suites, and there's a lot of talk about solutions. So that was pretty insightful, 'cause now you open it up to this much broader set of applications. Well, absolutely. The ecosystem is the essential part of the offering because obviously, as a company, we cannot be expert in every single domain. And to a certain extent, even FPGA designers, they are what, about maybe 10, 15,000 FPGA designers in the world. They are not really expert in the end application. So one of the challenges that we're trying to address is how do we make application developers, the people who are already playing in the cloud, the ISVs, for example, who have the expertise of what the end user wants, being able to develop something that is efficient to the end user in FPGAs. And this is why we've created a tool called Quick Play, which basically enables what we call the accelerator function developers, the guys who have the application expertise, to leverage an ecosystem of IP providers in the FPGA space that have built efficient building blocks, like encryption, compression, video transcoding. Right. These sort of things. So what you have is an ecosystem of cloud service providers. You have an ecosystem of IP providers. And we have this growing ecosystem of accelerator developers that develop all these accelerators that are sold as a service. And that really opens up the number of people that are qualified to play in the space. 'Cause you're kind of hiding the complexity into the hardcore, harder engineers and really making it more kind of a traditional software application space. Is that right? Yeah, you're absolutely right. And we're doing that on the technical front, but we're also doing that on the business model front. Because one thing with FPGAs is that FPGAs has relied heavily over the years on the IP industry. And the IP industry for FPGAs, and it's the same for ASIGs, have been also relying on the business model, which is based on very high up-front cost. So let me give you an example. Let's say I want to develop an accelerator, right, for database. And what I need to do is to get the stream of data coming in. It's most likely encrypted, so I need to decrypt this data, then I want to do some search algorithm on it to extract certain functions. I'm going to do some processing on it, and maybe the last thing I want to do is, I want to compress because I want to store the result of that data. If I'm doing that with a traditional IP business model, what I need to do is basically go and talk to every single one of those IP providers and ask them to sell me the IP. In the traditional IP business model, I'm looking at somewhere between 200,000 to 500,000 up front cost. And I want to sell this accelerator for maybe a couple of dollars on one of the marketplace. There's something that doesn't play out. So what we've done, also, is we've introduced a pay-per-use business model that allows us to track those IPs that are being used by the accelerators so we can propagate the as-a-service business model throughout the industry, the supply chain. Which is huge, right? 'Cause as much as cloud is about flexibility and extensibility, it's about the business model as well. About paying what you use when you use it, turning it on, turning it off. So that's a pretty critical success factor. Absolutely, I mean, you can imagine that there's, I don't know, millions of users in the cloud. There's maybe hundreds of thousands of different type of ways they're processing their data. So we also need a very agile ecosystem that can develop very quickly. And we also need them to do it in a way that doesn't cost too much money, right? Think about it, and think about the app store when it was launched, right? Right. When Apple launched the iPhone back about 10 years ago, right, they didn't have much application. And they didn't, I don't think they quite knew, exactly, how it was going to be used. But what they did, which completely changed the industry, is they opened up the SDK that they sold for very small amount of money and enabled a huge community to come up with a lot of a lot of application. And now you go there and you can find application that really meats your need. That's kind of the similar concept that we're trying to develop here. Right. So how's been the uptake? I mean, so where are you, kind of, in the life cycle of this project? 'Cause it's a relatively new spinout of the larger company? Yes, so it's relatively new. We did the spinout because we really want to give that product its own life. Right, right. Right? But we are still at the beginning. So we started a developing partnership with cloud service providers. The two ones that we've announced is Amazon Web Services and OVH, the cloud service provider in France. And we have recruited, I think, about a dozen IP partners. And now we're also working with accelerator developer, accelerator functions developers. Okay. So it's a work in progress. And our main goal right now is to, really to evangelize, and to show them how much money they can do and how they can serve this market of FPGA acceleration as a service. The cloud providers, or the application providers? Who do you really have to convince the most? So the one we have to convince today are really the application developers. Okay, okay. Because without content, your marketplace doesn't mean much. So this is the main thing we're focusing on right now. Okay, great. So, 2017's coming to an end, which is hard to believe. So as you look forward to 2018, of those things you just outlined, kind of what are some of the top priorities for 2018? So, top priorities will be to strengthen our relationship with the key cloud service providers we work with. We have a couple of other discussions ongoing to try to offer a platform on more cloud service providers. We also want to strengthen our relationship with Intel. And we'll continue the evangelization to really onboard all the IP providers and the accelerator developers so that the marketplace becomes filled with valuable accelerators that people can use. And that's going to be a long process, but we are focusing right now on key application space that we know people can leverage in the application. Exciting times. Oh yeah, it is. You know, it's 10 years since the app store launched, I think, so I look at acceleration as a service in cloud service providers, this sounds like a terrific opportunity. It is, it is a huge opportunity. Everybody's talking about it. We just need to materialize it now. All right, well, congratulations and thanks for taking a couple minutes out of your day. Oh, thanks for your time. All right, he's Stephane, I'm Jeff Frick. You're watching theCUBE from Super Computing 2017. Thanks for watching. (upbeat music)

Published Date : Nov 14 2017

SUMMARY :

So one of the challenges that we're trying to address

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Stephane MonoboissetPERSON

0.99+

AnthonyPERSON

0.99+

TeresaPERSON

0.99+

AWSORGANIZATION

0.99+

RebeccaPERSON

0.99+

InformaticaORGANIZATION

0.99+

JeffPERSON

0.99+

Lisa MartinPERSON

0.99+

Teresa TungPERSON

0.99+

Keith TownsendPERSON

0.99+

Jeff FrickPERSON

0.99+

Peter BurrisPERSON

0.99+

Rebecca KnightPERSON

0.99+

MarkPERSON

0.99+

SamsungORGANIZATION

0.99+

DeloitteORGANIZATION

0.99+

JamiePERSON

0.99+

John FurrierPERSON

0.99+

Jamie SharathPERSON

0.99+

RajeevPERSON

0.99+

AmazonORGANIZATION

0.99+

JeremyPERSON

0.99+

Ramin SayarPERSON

0.99+

HollandLOCATION

0.99+

Abhiman MatlapudiPERSON

0.99+

2014DATE

0.99+

RajeemPERSON

0.99+

Jeff RickPERSON

0.99+

SavannahPERSON

0.99+

Rajeev KrishnanPERSON

0.99+

threeQUANTITY

0.99+

Savannah PetersonPERSON

0.99+

FranceLOCATION

0.99+

Sally JenkinsPERSON

0.99+

GeorgePERSON

0.99+

StephanePERSON

0.99+

John FarerPERSON

0.99+

JamaicaLOCATION

0.99+

EuropeLOCATION

0.99+

AbhimanPERSON

0.99+

YahooORGANIZATION

0.99+

130%QUANTITY

0.99+

Amazon Web ServicesORGANIZATION

0.99+

2018DATE

0.99+

30 daysQUANTITY

0.99+

ClouderaORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

183%QUANTITY

0.99+

14 millionQUANTITY

0.99+

AsiaLOCATION

0.99+

38%QUANTITY

0.99+

TomPERSON

0.99+

24 millionQUANTITY

0.99+

TheresaPERSON

0.99+

AccentureORGANIZATION

0.99+

AccelizeORGANIZATION

0.99+

32 millionQUANTITY

0.99+

Karsten Ronner, Swarm64 | Super Computing 2017


 

>> Announcer: On Denver, Colorado, it's theCUBE, covering SuperComputing '17, brought to you by Intel. >> Hey, welcome back, everybody. Jeff Frick here with theCUBE. We're in Denver, Colorado at this SuperComputing conference 2017. I think there's 12,000 people. Our first time being here is pretty amazing. A lot of academics, a lot of conversations about space and genomes and you know, heavy-lifting computing stuff. It's fun to be here, and we're really excited. Our next guest, Karsten Ronner. He's the CEO of Swarm64. So Karsten, great to see you. >> Yeah, thank you very much for this opportunity. >> Absolutely. So for people that aren't familiar with Swarm64, give us kind of the quick eye-level. >> Yeah. Well, in a nutshell, Swarm64 is accelerating relational databases, and we allow them to ingest data so much faster, 50 times faster than a relational database. And we can also then query that data 10, 20 times faster than relational database. And that is very important for many new applications in IoT and in netbanking and in finance, and so on. >> So you're in a good space. So beyond just general or better performance, faster, faster, faster, you know, we're seeing all these movements now in real-time analytics and real-time applications, which is only going to get crazier with IoT and Internet of Things. So how do you do this? Where do you do this? What are some of the examples you could share with us? >> Yeah, so all our solution is a combination of a software wrapper that attaches our solution to existing databases. And inside, there's an FPGA from Intel, the Arria 10. And we are combining both, such that they actually plug into standard interfaces of existing databases, like in PostgreSQL, Foreign Data Wrappers, the storage engine in MySQL, and MariaDB and so on. And with that mechanism, we ensure that the database, the application doesn't see us. For the application, there's just fast database but we're invisible and also the functionality of the database remains what it was. That's the net of what we're doing. >> So that's so important because we talked a little bit about offline, you said you had a banking customer that said they have every database that's ever been created. They've been buying them all along so they've got embedded systems, you can't just rip and replace. You have to work with existing infrastructure. At the same time, they want to go faster. >> Yeah, absolutely right. Absolutely right. And there's a huge code base, which has been verified, which has been debugged, and in banking, it's also about compliance so you can't just rip out your old code base and do something new, because again, you would have to go through compliance. Therefore, customers really, really, really want their existing databases faster. >> Right. Now the other interesting part, and we've talked to some of the other Intel execs, is kind of this combination hybrid of the Hardware Software Solution in the FPGA, and you're really opening up an ecosystem for people to build more software-based solutions that leverage that combination of the hardware software power. Where do you see that kind of evolving? How's that going to help your company? >> Yeah. We are a little bit unique in that we are hiding that FPGA from the user, and we're not exposing it. Many people, actually, many applications expose it to the user, but apart from that, we are benefiting a lot from what Intel is doing. Intel is providing the entire environment, including virtualization, all those things that help us then to be able to get into Cloud service providers or into proprietary virtualized environments and things like that. So it is really a very close cooperation with Intel that helps us and enables us to do what we're doing. >> Okay. And I'm curious because you spend a lot of time with customers, you said a lot of legacy customers. So as they see the challenges of this new real-time environment, what are some of their concerns, what are some of the things that they're excited that they can do now with real-time, versus bash and data lake. And I think it's always funny, right? We used to make decisions based on stuff that happened in the past. And we're kind of querying now really the desires just to make action on stuff that's happening now, it's a fundamentally different way to address a problem. >> Yeah, absolutely. And a very, very key element of our solution is that we can not only insert these very, very large amounts of data that also other solutions can do, massively parallel solutions, streaming solutions, you know them all. They can do that too. However, the difference is that we can make that data available within less than 10 microseconds. >> Jeff: 10 microseconds? >> So dataset arrives within less than 10 microseconds, that dataset is part of the next query and that is a game changer. That allows you to do controlled loop processing of data in machine-to-machine environments, and autonomous, for autonomous applications and all those solutions where you just can't wait. If your car is driving down the street, you better know what has happened, right? And you can react to it. As an example, it could be a robot in a plant or things like that, where you really want to react immediately. >> I'm curious as to the kind of value unlocking that that provides to those old applications that were working with what they think is an old database. Now, you said, you know, you're accelerating it. To the application, it looks just the same as it looked before. How does that change those performances of those applications? I would imagine there's a whole other layer of value unlocking in those entrenched applications with this vast data. >> Yeah. That is actually true, and it's on a business level, the applications enable customers to do things they were not capable of doing before, and look for example in finance. If you can analyze the market data much quicker, if you can analyze past trades much quicker, then obviously you're generating value for the firm because you can react to market trends more accurately, you can mirror them in a more tighter fashion, and if you can do that, then you can reduce the margin of error with which you're estimating what's happening, and all of that is money. It's really pure money in the bank account of the customer, so to speak. >> Right. And the other big trend we talked about, besides faster, is you know, sampling versus not sampling. In the old days, we sampled old data and made decisions. Now we don't want to sample, we want all of the data, we want to make decisions on all the data, so again that's opening up another level of application performance because it's all the data, not a sample. >> For sure. Because before, you were aggregating. When you aggregate, you reduce the amount of information available. Now, of course, when you have the full set of information available, your decision-making is just so much smarter. And that's what we're enabling. >> And it's funny because in finance, you mentioned a couple of times, they've been doing that forever, right. The value of a few units of time, however small, is tremendous, but now we're seeing it in other industries as well that realize the value of real-time, aggregated, streaming data versus a sampling of old. Really opens up new types of opportunities. >> Absolutely, yes, yes. Yeah, finance, as I mentioned is an example, but then also IoT, machine-to-machine communication, everything which is real-time, logging, data logging, security and network monitoring. If you want to really understand what's flowing through your network, is there anything malicious, is there any actor on my network that should not be there? And you want to react so quickly that you can prevent that bad actor from doing anything to your data, this is where we come in. >> Right. And security's so big, right? It in everywhere. Especially with IoT and machine learning. >> Absolutely. >> All right, Karsten, I'm going to put you on the spot. So we're November 2017, hard to believe. As you look forward to 2018, what are some of your priorities? If we're standing here next year, at SuperComputing 2018, what are we going to be talking about? >> Okay, what we're going to talk about really is that we will, right now we're accelerating single-server solutions and we are working very, very hard on massively parallel systems, while retaining the real-time components. So we will not only then accelerate a single server, by then, allowing horizontal scaling, we will then bring a completely new level of analytics performance to customers. So that's what I'm happy to talk to you about next year. >> All right, we'll see you next year, I think it's in Texas. >> Wonderful, yeah, great. >> So thanks for stopping by. >> Thank you. >> He's Karsten, I'm Jeff. You're watching TheCUBE, from SuperComputing 2017. Thanks for watching.

Published Date : Nov 14 2017

SUMMARY :

brought to you by Intel. and genomes and you know, Yeah, thank you very of the quick eye-level. And that is very important for So how do you do this? ensure that the database, about offline, you said about compliance so you can't just rip out How's that going to help your company? that FPGA from the user, stuff that happened in the past. is that we can make the street, you better know that provides to those and if you can do that, then you can And the other big trend we talked about, Now, of course, when you have the in finance, you mentioned quickly that you can prevent And security's so big, right? going to put you on the spot. talk to you about next year. All right, we'll see you next Thanks for watching.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Karsten RonnerPERSON

0.99+

KarstenPERSON

0.99+

TexasLOCATION

0.99+

JeffPERSON

0.99+

Jeff FrickPERSON

0.99+

50 timesQUANTITY

0.99+

November 2017DATE

0.99+

MySQLTITLE

0.99+

next yearDATE

0.99+

2018DATE

0.99+

next yearDATE

0.99+

less than 10 microsecondsQUANTITY

0.99+

Swarm64ORGANIZATION

0.99+

10 microsecondsQUANTITY

0.99+

12,000 peopleQUANTITY

0.99+

Denver, ColoradoLOCATION

0.99+

PostgreSQLTITLE

0.99+

10QUANTITY

0.99+

MariaDBTITLE

0.99+

IntelORGANIZATION

0.99+

Denver, ColoradoLOCATION

0.99+

20 timesQUANTITY

0.99+

bothQUANTITY

0.99+

first timeQUANTITY

0.98+

Swarm64TITLE

0.97+

SuperComputingEVENT

0.97+

single serverQUANTITY

0.95+

SuperComputing '17EVENT

0.91+

theCUBEORGANIZATION

0.85+

Super Computing 2017EVENT

0.83+

TheCUBETITLE

0.81+

singleQUANTITY

0.65+

2017EVENT

0.63+

bashTITLE

0.6+

Foreign Data WrappersTITLE

0.54+

SuperComputingTITLE

0.54+

Arria 10ORGANIZATION

0.52+

2017DATE

0.4+

Bill Jenkins, Intel | Super Computing 2017


 

>> Narrator: From Denver, Colorado, it's theCUBE. Covering Super Computing 17. Brought to you by Intel. (techno music) Hey, welcome back, everybody. Jeff Frick here with theCUBE. We're in Denver, Colorado at the Super Computing Conference 2017. About 12 thousand people, talking about the outer edges of computing. It's pretty amazing. The keynote was huge. The square kilometer array, a new vocabulary word I learned today. It's pretty exciting times, and we're excited to have our next guest. He's Bill Jenkins. He's a Product Line Manager for AI on FPGAs at Intel. Bill, welcome. Thank you very much for having me. Nice to meet you, and nice to talk to you today. So you're right in the middle of this machine-learning AI storm, which we keep hearing more and more about. Kind of the next generation of big data, if you will. That's right. It's the most dynamic industry I've seen since the telecom industry back in the 90s. It's evolving every day, every month. Intel's been making some announcements. Using this combination of software programming and FPGAs on the acceleration stack to get more performance out of the data center. Did I get that right? Sure, yeah, yeah. Pretty exciting. The use of both hardware, as well as software on top of it, to open up the solution stack, open up the ecosystem. What of those things are you working on specifically? I really build first the enabling technology that brings the FPGA into that Intel ecosystem. Where Intel is trying to provide that solution from top to bottom to deliver AI products. >> Jeff: Right. Into that market. FPGAs are a key piece of that because we provide a different way to accelerate those machine-learning and AI workloads. Where we can be an offload engine to a CPU. We can be inline analytics to offload the system, and get higher performance that way. We tie into that overall Intel ecosystem of tools and products. Right. So that's a pretty interesting piece because the real-time streaming data is all the rage now, right? Not in batch. You want to get it now. So how do you get it in? How do you get it written to the database? How do you get it into the micro-processor? That's a really, really important piece. That's different than even two years ago. You didn't really hear much about real-time. I think it's, like I said, it's evolving quite a bit. Now, a lot of people deal with training. It's the science behind it. The data scientists work to figure out what topologies they want to deploy and how they want to deploy 'em. But now, people are building products around it. >> Jeff: Right. And once they start deploying these technologies into products, they realize that they don't want to compensate for limitations in hardware. They want to work around them. A lot of this evolution that we're building is to try to find ways to more efficiently do that compute. What we call inferencing, the actual deployed machine-learning scoring, as they will. >> Jeff: Right. In a product, it's all about how quickly can I get the data out. It's not about waiting two seconds to start the processing. You know, in an autonomous-driven car where someone's crossing the road, I'm not waiting two seconds to figure out it's a person. Right, right. I need it right away. So I need to be able to do that with video feeds, right off a disk drive, from the ethernet data coming in. I want to do that directly in line, so that my processor can do what it's good at, and we offload that processor to get better system performance. Right. And then on the machine-learning specifically, 'cause that is all the rage. And it is learning. So there is a real-time aspect to it. You talked about autonomous vehicles. But there's also continuous learning over time, that's not necessarily dependent on learning immediately. Right. But continuous improvement over time. What are some of the unique challenges in machine-learning? And what are some of the ways that you guys are trying to address those? Once you've trained the network, people always have to go back and retrain. They say okay, I've got a good accuracy, but I want better performance. Then they start lowering the precision, and they say well, today we're at 32-bit, maybe 16-bit. Then they start looking into eight. But the problem is, their accuracy drops. So they retrain that into eight topology, that network, to get the performance benefit, but with the higher accuracy. The flexibility of FPGA actually allows people to take that network at 32-bit, with the 32-bit trained weights, but deploy it in lower precision. So we can abstract away the fact that the hardware's so flexible, we can do what we call floating point 11-bit floating point. Or even 8-bit floating point. Even here today at the show, we've got a binary and ternary demo, showcasing the flexibility that the FPGA can provide today with that building block piece of hardware that the FPGA can be. And really provide, not only the topologies that people are trying to build today, but tomorrow. >> Jeff: Right. Future proofing their hardware. But then the precisions that they may want to do. So that they don't have to retrain. They can get less than a 1% accuracy loss, but they can lower that precision to get all the performance benefits of that data scientist's work to come up with a new architecture. Right. But it's interesting 'cause there's trade-offs, right? >> Bill: Sure. There's no optimum solution. It's optimum as to what you're trying to optimize for. >> Bill: Right. So really, the ability to change the ability to continue to work on those learning algorithms, to be able to change your priority, is pretty key. Yeah, a lot of times today, you want this. So this has been the mantra of the FPGA for 30 plus years. You deploy it today, and it works fine. Maybe you build an ASIC out of it. But what you want tomorrow is going to be different. So maybe if it's changing so rapidly, you build the ASIC because there's runway to that. But if there isn't, you may just say, I have the FPGA, I can just reprogram it to do what's the next architecture, the next methodology. Right. So it gives you that future proofing. That capability to sustain different topologies. Different architectures, different precisions. To kind of keep people going with the same piece of hardware. Without having to say, spin up a new ASIC every year. >> Jeff: Right, right. Which, even then, it's so dynamic it's probably faster then, every year, the way things are going today. So the other thing you mentioned is topography, and it's not the same topography you mentioned, but this whole idea of edge. Sure. So moving more and more compute, and store, and smarts to the edge. 'Cause there's just not going to be time, you mentioned autonomous vehicles, a lot of applications to get everything back up into the cloud. Back into the data center. You guys are pushing this technology, not only in the data center, but progressively closer and closer to the edge. Absolutely. The data center has a need. It's always going to be there, but they're getting big. The amount of data that we're trying to process every day is growing. I always say that the telecom industry started the Information Age. Well, the Information Age has done a great job of collecting a lot of data. We have to process that. If you think about where, maybe I'll allude back to autonomous vehicles. You're talking about thousands of gigabytes, per day, of data generated. Smart factories. Exabytes of data generated a day. What are you going to do with all that? It has to be processed. We need that compute in the data center. But we have to start pushing it out into the edge, where I start thinking, well even a show like this, I want security. So, I want to do real-time weapons detection, right? Security prevention. I want to do smart city applications. Just monitoring how traffic moves through a mall, so that I can control lighting and heating. All of these things at the edge, in the camera, that's deployed on the street. In the camera that's deployed in a mall. All of that, we want to make those smarter, so that we can do more compute. To offload the amount of data that needs to be sent back to the data center. >> Jeff: Right. As much as possible. Relevant data gets sent back. No shortage of demand for compute store networking, is there? No, no. It's really a heterogeneous world, right? We need all the different compute. We need all the different aspects of transmission of the data with 5G. We need disk space to store it. >> Jeff: Right. We need cooling to cool it. It's really becoming a heterogeneous world. All right, well, I'm going to give you the last word. I can't believe we're in November of 2017. Yeah. Which is bananas. What are you working on for 2018? What are some of your priorities? If we talk a year from now, what are we going to be talking about? Intel's acquired a lot of companies over the past couple years now on AI. You're seeing a lot of merging of the FPGA into that ecosystem. We've got the Nervana. We've got Movidius. We've got Mobileye acquisitions. Saffron Technologies. All of these things, when the FPGA is kind of a key piece of that because it gives you that flexibility of the hardware, to extend those pieces. You're going to see a lot more stuff in the cloud. A lot more stuff with partners next year. And really enabling that edge to data center compute, with things like binary neural networks, ternary neural networks. All the different next generation of topologies to kind of keep that leading edge flexibility that the FPGA can provide for people's products tomorrow. >> Jeff: Exciting times. Yeah, great. All right, Bill Jenkins. There's a lot going on in computing. If you're not getting your computer science degree, kids, think about it again. He's Bill Jenkins. I'm Jeff Frick. You're watching theCUBE from Super Computing 2017. Thanks for watching. Thank you. (techno music)

Published Date : Nov 14 2017

SUMMARY :

Kind of the next generation of big data, if you will. We can be inline analytics to offload the system, A lot of this evolution that we're building is to try to of hardware that the FPGA can be. So that they don't have to retrain. It's optimum as to what you're trying to optimize for. So really, the ability to change the ability to continue We need that compute in the data center. We need all the different aspects of of the hardware, to extend those pieces. There's a lot going on in computing.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
JeffPERSON

0.99+

Jeff FrickPERSON

0.99+

Bill JenkinsPERSON

0.99+

two secondsQUANTITY

0.99+

2018DATE

0.99+

November of 2017DATE

0.99+

8-bitQUANTITY

0.99+

16-bitQUANTITY

0.99+

32-bitQUANTITY

0.99+

todayDATE

0.99+

next yearDATE

0.99+

BillPERSON

0.99+

30 plus yearsQUANTITY

0.99+

11-bitQUANTITY

0.99+

tomorrowDATE

0.99+

Denver, ColoradoLOCATION

0.99+

IntelORGANIZATION

0.98+

eightQUANTITY

0.98+

MovidiusORGANIZATION

0.98+

Super Computing Conference 2017EVENT

0.98+

a dayQUANTITY

0.96+

Saffron TechnologiesORGANIZATION

0.96+

thousands of gigabytesQUANTITY

0.95+

MobileyeORGANIZATION

0.95+

About 12 thousand peopleQUANTITY

0.95+

two years agoDATE

0.95+

90sDATE

0.94+

less than a 1%QUANTITY

0.94+

NervanaPERSON

0.94+

FPGAORGANIZATION

0.9+

both hardwareQUANTITY

0.89+

firstQUANTITY

0.84+

Exabytes of dataQUANTITY

0.76+

Super Computing 2017EVENT

0.75+

past couple yearsDATE

0.73+

every yearQUANTITY

0.69+

yearQUANTITY

0.69+

per dayQUANTITY

0.6+

5GQUANTITY

0.58+

Super Computing 17EVENT

0.55+

theCUBEORGANIZATION

0.52+

FPGATITLE

0.42+

Bernhard Friebe, Intel Programmable Solutions Group | Super Computing 2017


 

>> Announcer: From Denver, Colorado, it's theCUBE. Covering Super Computing 2017 brought to you by Intel. (upbeat music) >> Hey, welcome back everybody. Jeffrey Frick here with theCube. We're in Denver, Colorado at Super Computing 17. I think it's the 20th year of the convention. 12,000 people. We've never been here before. It's pretty amazing. Amazing keynote, really talking about space, and really big, big, big computing projects, so, excited to be here, and we've got our first guest of the day. He's Bernard Friebe, he is the Senior Director of FPGA, I'll get that good by the end of the day, Software Solutions for Intel Programmable group. First off, welcome, Bernard. >> Thank you. I'm glad to be here. >> Absolutely. So, have you been to this conference before? >> Yeah, a couple of times before. It's always a big event. Always a big show for us, so I'm excited. >> Yeah, and it's different, too, cuz it's got a lot of academic influence, as well, as you walk around the outside. It's pretty hardcore. >> Yes, it's wonderful, and you see a lot of innovation going on, and we need to move fast. We need to move faster. That's what it is. And accelerate. >> And that's what you're all about, acceleration, so, Intel's making a lot of announcements, really, about acceleration at FPGA. For acceleration and in data centers and in big data, and all these big applications. So, explain just a little bit how that seed is evolving and what some of the recent announcements are all about. >> The world of computing must accelerate. I think we all agree on that. We all see that that's a key requirement. And FPGA's are a truly versatile, multi-function accelerator. It accelerates so many workloads in the high-performance computing space, may it be financial, genomics, oil and gas, data analytics, and the list goes on. Machine learning is a very big one. The list goes on and on. And, so, we're investing heavily in providing solutions which makes it much easier for our users to develop and deploy FPGA in a high-performance computing environment. >> You guys are taking a lot of steps to make the software programming at FPGA a lot easier, so you don't have to be a hardcore hardware engineer, so you can open it up to a broader ecosystem and get a broader solution set. Is that right? >> That's right, and it's not just the hardware. How do you unlock the benefits of FPGA as a versatile accelerator, so their parallelism, their ability to do real-time, low-latency, acceleration of many different workloads, and how do you enable that in an environment which is truly dynamic and multi-function, like a data center. And so, the product we've recently announced is the acceleration stack for xeon with FPGA, which enables that use more. >> So, what are the components for that stack? >> It starts with hardware. So, we are building a hardware accelerator card, it's a pc express plugin card, it's called programmable accelerator card. We have integrated solutions where you have everything on an FPGA in package, but what's common is a software framework solution stack, which sits on top of these different hardware implementation, which really makes it easy for a developer to develop an accelerator, for a user to then deploy that accelerator and run it in their environment, and it also enables a data center operator to basically enable the FPGA like any other computer resources by integrating it into their orchestration framework. So, multiple levels taking care of all those needs. >> It's interesting, because there's a lot of big trends that you guys are taking advantage of. Obviously, we're at Super Computing, but big data, streaming analytics, is all the rage now, so more data faster, reading it in real time, pumping it into the database in real time, and then, right around the corner, we have IoT and internet of things and all these connected devices. So the demand for increased speed, to get that data in, get that data processed, get the analytics back out, is only growing exponentially. >> That's right, and FPGAs, due to their flexibility, have distinct advantages there. The traditional model is look aside of offload, where you have a processor, and then you offload your tasks to your accelerator. The FPGA, with their flexible I/Os and flexible core can actually run directly in the data path, so that's what we call in-line processing. And what that allows people to do is, whatever the source is, may it be cameras, may it be storage, may it be through the network, through ethernet, can stream directly into the FPGA and do your acceleration as the data comes in in a streaming way. And FPGAs provide really unique advantages there versus other types of accelerators. Low-latency, very high band-width, and they're flexible in a sense that our customers can build different interfaces, different connectivity around those FPGAs. So, it's really amazing how versatile the usage of FPGA has become. >> It is pretty interesting, because you're using all the benefits that come from hardware, hardware-based solutions, which you just get a lot of benefits when things are hardwired, with the software component and enabling a broader ecosystem to write ready-made solutions and integrations to their existing solutions that they already have. Great approach. >> The acceleration stack provides a consistent interface to the developer and the user of the FPGA. What that allows our ecosystem and our customers to do is to define these accelerators based on this framework, and then they can easily migrate those between different hardware platforms, so we're building in future improvements of the solution, and the consistent interfaces then allow our customers and partners to build their software stacks on top of it. So, their investment, once they do it and we target our Arria 10 programmable accelerator card can easily be leveraged and moved forward into the next generation strategy, and beyond. We enable, really, and encourage a broad ecosystem, to build solutions. You'll see that here at the show, many partners now have demos, and they show their solutions built on Intel FPGA hardware and the acceleration stack. >> OK, so I'm going to put you on the spot. So, these are announced, what's the current state of the general availability? >> We're sampling now on the cards, the acceleration stack is available for delivery to customers. A lot of it is open source, by the way, so it can already be downloaded from GitHub And the partners are developing the solutions they are demonstrating today. The product will go into volume production in the first half of next year. So, we're very close. >> All right, very good. Well, Bernard, thanks for taking a few minutes to stop by. >> Oh, it's my pleasure. >> All right. He's Bernard, I'm Jeff. You're watching theCUBE from Super Computing 17. Thanks for watching. (upbeat music)

Published Date : Nov 14 2017

SUMMARY :

brought to you by Intel. I'll get that good by the end of the day, I'm glad to be here. So, have you been to this conference before? Yeah, a couple of times before. Yeah, and it's different, too, and you see a lot of innovation going on, For acceleration and in data centers and the list goes on. and get a broader solution set. and how do you enable that in an environment and run it in their environment, and all these connected devices. and FPGAs, due to their flexibility, and enabling a broader ecosystem and the consistent interfaces then OK, so I'm going to put you on the spot. A lot of it is open source, by the way, Well, Bernard, thanks for taking a few minutes to stop by. Thanks for watching.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
BernardPERSON

0.99+

Bernard FriebePERSON

0.99+

Bernhard FriebePERSON

0.99+

Jeffrey FrickPERSON

0.99+

JeffPERSON

0.99+

Intel Programmable Solutions GroupORGANIZATION

0.99+

12,000 peopleQUANTITY

0.99+

Denver, ColoradoLOCATION

0.99+

20th yearQUANTITY

0.98+

Super Computing 17EVENT

0.97+

FPGAORGANIZATION

0.97+

Super Computing 2017EVENT

0.97+

todayDATE

0.96+

FirstQUANTITY

0.96+

GitHubORGANIZATION

0.95+

first half of next yearDATE

0.95+

first guestQUANTITY

0.95+

IntelORGANIZATION

0.95+

FPGATITLE

0.85+

theCubeORGANIZATION

0.84+

Arria 10COMMERCIAL_ITEM

0.73+

theCUBEORGANIZATION

0.54+

SuperEVENT

0.41+

ComputingTITLE

0.39+

17EVENT

0.36+

Karen Lu, Alibaba Group | The Computing Conference


 

>> Narrator: Silicon Angle Media presents TheCUBE! Covering Alibaba Cloud's annual conference. Brought to you by Intel. Now, here's John Furrier.... >> Hi, I'm John Furrier of Silicon Angle Media based in the United States in Silicon Valley in Palo Alto, California. I'm also co-host of TheCUBE where we go out through the event and extract the signal from the noise. We're here in China, we are here with a business development director of America's for Alibaba Cloud International, Karen Lu. Thanks for taking the time. >> Karen: Sure, absolutely. >> So, it's exciting for us from the US to come to China to hear the (mumbles), but I'm blown away by the culture. It's not a B-to-B tech conference. It's not boring. It's exciting. Talk about the Alibaba Cloud. What's so special about Alibaba Cloud? >> Sure. Alibaba Cloud is actually the encumbered cloud provider in China, and further more we extend our reach into global market since two years ago, and our strategy for globalize our cloud services is really to bridge up the business communities from overseas to China, from US to China, from US to Asia-Pacific, and to connect the rest of the world as well. Our goal is set up the platform to enable our enterprise customers, our SMEs, small and medium customer base be able to utilize our platform to develop their applications, their vertical solutions to benefit their end users. >> Alibaba Cloud has come such a long way since 2009. So much has happened, Alibaba grew up as a company. It's not just e-commerce. It's intersecting e-commerce, entertainment and web services, which is the magical formula that consumers want. They don't want just a business solution or just do e-commerce. You guys have weaved that formula together. What's special about that formula, and why is Alibaba important to the folks in the United States? >> I think it's all about the ecosystem and what makes the people, the people's community, and business community benefit from the services we provided to the world, right? Not just the e-commerce platform that have been running for the past 18 years, but also entertaining, to the map services, location services, the data services like Ali Cloud is providing, and be able to put out those elements together, and benefit people's lives, and help to improve users' experience from globally. >> It's been impressive here in China. Now as you go outside of China in the globalization plan, what's the strategy, what's the tactics? What are you going to do? >> I think our value is to, as I mentioned earlier, bridge up the business communities, especially to enable the outside world benefit a huge market from Mainland China and rest of the world as well, so I think I think our key value is to enable the business communities and be able to help them reach out outside the world. That being said, one of our key globalization strategy is to be able to help the SME's, small and medium companies to benefit the new technologies to the level that they won't be able to get in the past. It's the old technologies. >> John: What's some of the statistics or facts, fun facts, or Alibaba stats in the US, North America, your presence there, can you share what the current situation is? >> Sure. I think things about two years ago, when we extend our reach in two your market, we now have more than two thousand customers from individual to startup, to medium enterprises, and to some very large enterprises in the world as well. People are from the communities get to know Alibaba Cloud and get to know Alibaba not only provide to the e-commerce services, the EWTP platform to the world. We're also brought the data technologies. We also provided the technologies to the world that benefit their reach to the world. >> Everyone talks about data-driven. You guys have a very specific data formula, data fueling, not just getting the data from engagement data and user data, but fueling data in for user experience. The question is as you go outside of China into the US, certainly you have a developer ecosystem, you have a business ecosystem. >> Correct. >> How do the folks benefit locally in the US, to our business, do they have have access to China? Is it the services, is it the technology? Can you share the benefits to the developers and to the businesses? >> Sure absolutely. We ran a program called the China Connect, and that's the program we help the business communities you have, from the IVs, the independent after vendors, from the sales providers and developers' opportunity of communities to be able to develop their applications and software, and bring those benefits to China market. Through this process, it's hard to navigate a brand new market, especially in China, without knowing the people, the communities, the culture, the business practices here, right? We actually provide a platform, a program to help them to get to know the market, and help them to land their business in China through this program, and help them, of course, expertise their business roles in China. >> A lot of people want to know what's inside their cloud. It's one of those things where this mysterious cloud. The security's a concern, but partnerships are critical. Talk about what's inside your cloud. Intel's a big partner. What's the Alibaba-Intel partnership like? >> It's a fantastic partnership. We have been established over the past years, and Intel is one of our strategic alliance in the marketplace. They provide us a lot from hardware to technology, in terms of helping us to establish the platform with the business communities, not only China, but globally, so we really appreciate Intel's partnership, and moving forward we are looking for more reciprocal partnership with Intel to be able to form more strategic partnership to be able to benefit the business communities, and people's communities as well. >> For the folks in the US, I'll say that this is an amazing conference. It's got a million people here. I don't even know the numbers. I'm sure you have the numbers handy, but it's a mix of developers. You have a crowded house here with developers, but you also have some business people. You have key partners. I saw some US companies here. What's the vibe at the event? What's the feeling here? You got a music festival three nights. It's not a boring tech conference. Is that by design? Share the stats, how many people are here? >> I guess this is the excitement of this, the conference, annually, we actually invite a lot of our customers from US, and the rest of the world to join us to share the excitement from China, to share the experience from Alibaba. Just like Jack said, the vision for us is to make people's lives more healthier and happier. The 2H strategy from us, right, is not just the hardworking. It's also the fun. It's also the the excitement for us to share these technologies, to share this platform, and to enable people to enjoy this technology. >> The scene I see here is interesting. I've seen at Apple, in the late 90's when Steve Jobs transformed that company, he had the vision of technology meeting liberal arts. That became their calling card. You guys have art and science come in together. It's not just scientists and developers. You have artists here because user experience is super important >> exactly. >> Is that part of the culture as science and art comes together because Jack is a charismatic leader. He's a culptive personality. Young Company. >> Karen: It is. >> Share the culture. >> It is. Just like Jack and other topic executives has been sharing with the community, we want to make sure technology is inclusive elements to everyone in the community, not just for the programmers or developers, or the very high-tech companies, right? It should benefit the entire society, and fun, of course, always as part of it to make people's life happier, and to make users' experience more satisfied. >> You had a career in international technology industry for a while. You see how it's played out in the past. We're in a different now. It's a global world. The internet has opened up a lot of good things, and sometimes not so good things. The US have the selection in fake news, but as the culture starts connecting, a new kind of normal is evolving. How does Alibaba see themselves in this new world order? >> I think we see ourselves as the enabler and platform to bring the technology, and bring the people, and bring the happiness together to benefit everyone in the world, not just the tech sectors, or just the e-commerce sectors, or just one of the single verticals. We are trying to bring the technologies, and the enablement, the platform that everybody can enjoy. That's the core value for us as the inclusive technology provider. >> For the folks in the United States that will see this video, share something that they may not know about Alibaba. Might be the first time in getting to see some of the culture and some of the commentary, what should they know about Alibaba as you guys move in and become global? They're going to see some services. Is it the services, is it the people, the culture, what should they engage with Alibaba at cloud? How should they see Alibaba Cloud? >> First of all, we are one of the top three cloud providers in the world. If you look at the latest (speaks in foreign language) released a couple of weeks ago, and that's why globalization is critical for us, and we want to be able to reach out to the overseas communities, and we want to build up the trust and the confidence with the local business communities, like the rating, where in US market for instance. For us, become the global family is critical for us, and this is our vision to bring the values to them as well. >> That's fantastic, spectacular culture, and the ecosystem is just now growing, open-source software is growing exponentially, global fabric of communities developing. It is opportunities for US companies and developers to access China. Talk a little bit more about the potential that entrepreneurs and businesses could have in this global framework. >> Sure. The beauty of cloud is actually the ecosystem. It's not just one company or one vertical. For us, for instance, we try to enable the small business, especially those startup business by offering them the free resources from our infrastructure at global level, be able to enable those young peoples, especially, to create their own ideas, to be innovative, and to utilize our resources, be able to access the technologies like the way the big companies has been invested into. This is, I think, as an example for us to commit to this global market. I think for us to be part of that family, especially in Silicon Valley is critical because of the technologies, because of innovations, and because of the mindset in Silicon Valley. That's why we set up our R&D centers, we set up our frontend back office in Silicon Valley as well be able to part of that reach in, and not only to learn the technologies, but sense the mindset in our reach in. I think that's critical for us as well as the Chinese headquarter of the company, but with a global vision. >> And where in Silicon Valley is your office? >> We're headquartered in San Mateo, California for US operations. >> And entrepreneurship is changing, and it's global. It's exciting. What's the benefit to entrepreneurship? Certainly, ventured capitalists are highly interested in the China market. They've been in here for a while. Is it coming together? >> Yes, it is indeed. Actually, not only we funding a lot of the new tech companies, we also been able to help them to find their partners to build up a extended ecosystem. In Silicon Valley, in West Coast reach ins, as well as extend from the inner US, in mid-western reach in, Chicago for instance, to New York coastal areas as well. >> I noticed on the sponsorship list and partner list in your ecosystem, a logo that is new, but it's super important in the US. It's growing like crazy. The Cloud Native Compute Foundation's here, and that's the Linux Foundation. They're partnering with you. The cloud native developer market is evolving very, very quickly. They're different than the old classic IT developers. A new generation, it's not IT anymore. It's data that's driving it, and it's open-source. How do you guys engage with that community because, clearly, they win with you. >> Yes. We're actually working with a lot of open-source partners like Docker, (mumbles), and others, be able to help them to bring the communities to bring their customers onto our infrastructures and create this platform to help the developer communities to develop their applications. It's a lot of vertical focus, the solution department tasks right now. >> Excuse me, you mentioned small, medium size enterprises and business, but the big enterprises are transforming as well. How do you see Alibaba helping them because they're going cloud native? They're going private cloud on premise. You have quantum computing. You even have IOT. You have a lot of things. How's the digital transformation message for enterprises and for small businesses that don't want to pay the technology tax. >> I think for large enterprises, the most strategy you have been seeing from the marketplace, one is multi-cloud strategy. People need redundancy. People want to reduce the dependencies for one or two cloud providers, and we work with other cloud providers in the community to provide interval qualities to support this multi-cloud strategies. On the other side, couple years back, people didn't know what's in a cloud. And then, people rush to cloud for everything. And now, people come back and review the strategies and find out hybrid-cloud strategy is more suitable for large enterprises. They have their on-prem architect and infrastructure. Meanwhile, they move some of their applications to cloud. It's a good combination of on-prem physical infrastructure cloud topology. We have been seeing a trend for both for large enterprise clients. For small business, especially for small business, they don't have the upfront huge investment paying to the infrastructure, and we provide them the instant access to the infrastructure, not only from computing storage network and the database perspective, more importantly from security perspective. >> The Alibaba Infrastructure services, I saw a part of the display here, very prominent in that equation. You guys have the scale. What can you share about the under-the-hood? What's the technology look like? What's the engine of Alibaba Cloud? How mature is it? What's to do? Where's the strategic direction? Block-chain is important, but now, that's changing everything It's all this new wave's coming. >> Just like the (speaks in foreign language) indicated two months ago, if you look at the overall qualifications to be a world lead cloud provider, we're number four, after AWS, Microsoft Azure, and Google Cloud, but if you look at market share and revenue, we're number three. That being said, we actually provide a very comprehensive technology, and the infrastructure to the business communities, and people's communities. For instance, from the global footprint perspective, right now, we have 14 reach ins, pretty much cover all the major market in the world. By end of this year into beginning of next year, we're going to activate two to three reach ins, make it 16 to 17 reach ins globally, that we can offer the global cloud solutions for the big and small businesses. >> That's exciting, and Silicon Valley certainly import our home base. Are you guys hiring, is there expanding? Share a little bit of a public service announcement on what's going on in the Silicon Valley area. You guys hiring, looking for engineers, what kind of people are you looking for? >> Yes, (laughs) great question. Actually, we are hiring, and we're looking for talented professionals join us from those marketing, business development, to cloud architect, to technical account management, to marketing premises, so we want to build up a business that we can truly build up the trust towards the local business communities. That's why we hire a lot of local talented young professionals, and to help them to be able to fit in to the culture, the unique culture of Alibaba, and also be able to contribute to this journey, very exciting journey... >> China has always been big. Everyone in the United States knows. The numbers are big here in terms of mobile deployment, app size. A lot of the people in the US look at China and say, "Wow, we can collaborate with China." It's a very nice distribution system, but they got to take care of their needs at home. >> Exactly. >> This is a big part of the undercurrent we're hearing. How do you guys help? >> Globalization is always critical for any business, even for some small business. Just like Jack Ma said this morning at his speech, even for small business, they need to globalize. They need to reach out to more business communities, and more customers. For us, because of the huge market in China, because of the EWTP platform we set up globally, because Alibaba Cloud Infrastructure and our global footprint, we're actually being able to help our customers, not only access the infrastructure from cloud perspective, but also help them to leverage our ecosystem from different business unit, and more partnership, to be able to help them to expertise their business in China and globally. >> That's exciting. Finally, developers are a big hot button. Everyone always says, I hear comments like, "We have to own the developer community," not that you could own the developer. No one wants to be owned, but what they mean is they want to win over the hearts and minds of developers. A lot of competition, and developers want programmable infrastructure. In dev ops world, that's called dev ops. That is really the new normal in developer community. How do you guys attack that developer market? >> We actually want to enable the developers community, not own or just win over. We want constantly enable them with the new platform, the new business models, the new programs that we can bring them together. That's our mission, enablement. >> Congratulations on a spectacular formula. Thanks for having us here, TheCUBE and Silicon Angle, and thanks for your time. >> Thank you so much for the opportunity. >> Karen Lu here in China with TheCUBE. Exclusive coverage in China, bringing the stories of the most important trends and tech in Alibaba Cloud. Really changing the game with their formula of e-commerce, entertainment, and entertainment. This is not B-to-B, boring to boring. It's exciting, in a music festival. 60 thousand people are here at this conference. Developers in the world watching, I'm John Furrier with Silicon Angle. Thanks for watching. (techno music)

Published Date : Oct 26 2017

SUMMARY :

Brought to you by and extract the signal from the noise. Talk about the Alibaba Cloud. and to connect the rest of the world as well. in the United States? and business community benefit from the services It's been impressive here in China. the new technologies to the level We also provided the technologies to the world not just getting the data and that's the program we help What's the Alibaba-Intel partnership like? in the marketplace. For the folks in the US, It's also the the excitement he had the vision of technology meeting liberal arts. Is that part of the culture and to make users' experience more satisfied. The US have the selection in fake news, and the enablement, the platform and some of the commentary, the overseas communities, and we want to build up and the ecosystem is just now growing, and because of the mindset in Silicon Valley. We're headquartered in San Mateo, California What's the benefit to entrepreneurship? a lot of the new tech companies, and that's the Linux Foundation. and create this platform to help the developer communities but the big enterprises are transforming as well. the most strategy you have been seeing from the marketplace, You guys have the scale. and the infrastructure to the business communities, Share a little bit of a public service announcement and also be able to contribute to this journey, A lot of the people in the US look at China and say, of the undercurrent we're hearing. because of the EWTP platform we set up globally, That is really the new normal in developer community. the new business models, the new programs and thanks for your time. Developers in the world watching,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
JackPERSON

0.99+

AlibabaORGANIZATION

0.99+

Steve JobsPERSON

0.99+

JohnPERSON

0.99+

ChinaLOCATION

0.99+

oneQUANTITY

0.99+

USLOCATION

0.99+

AsiaLOCATION

0.99+

Karen LuPERSON

0.99+

Jack MaPERSON

0.99+

John FurrierPERSON

0.99+

Silicon Angle MediaORGANIZATION

0.99+

KarenPERSON

0.99+

Silicon ValleyLOCATION

0.99+

New YorkLOCATION

0.99+

Linux FoundationORGANIZATION

0.99+

twoQUANTITY

0.99+

three nightsQUANTITY

0.99+

United StatesLOCATION

0.99+

AWSORGANIZATION

0.99+

IntelORGANIZATION

0.99+

Cloud Native Compute FoundationORGANIZATION

0.99+

Silicon AngleORGANIZATION

0.99+

North AmericaLOCATION

0.99+

2009DATE

0.99+

16QUANTITY

0.99+

Alibaba GroupORGANIZATION

0.99+

60 thousand peopleQUANTITY

0.99+

San Mateo, CaliforniaLOCATION

0.99+

Palo Alto, CaliforniaLOCATION

0.99+

TheCUBEORGANIZATION

0.99+

ChicagoLOCATION

0.99+

threeQUANTITY

0.99+

one companyQUANTITY

0.99+

first timeQUANTITY

0.98+

Mainland ChinaLOCATION

0.98+

Dr Min Wanli, Alibaba | The Computing Conference


 

>> Announcer: SiliconANGLE Media presents theCUBE! covering Alibaba Cloud's annual conference, brought to you by Intel. Now here's John Furrier.... >> Hi I'm John Furrier, with SiliconANGLE, Wikibon and theCUBE. I'm the co-founder based in Silicon Valley, California, Palo Alto, California, and I am here in Hangzhou, China for the Alibaba Cloud conference in Cloud City, it's the biggest cloud computing conference here in China. I'm excited to be here with Dr. Min Wanli, who's the Chief Data Scientist and General Manager of Big Data division at Alibaba Cloud. Dr. Wanli, thank you for spending time. >> Thank you for having me. >> We have seen a lot of data in the conversation here at the show, data technology's a big part of this new revolution, it's an industrial revolution that we've never seen before, a whole 'nother generation of technology. What does data technology mean to Alibaba? >> Okay, it means everything. So first off, our internal technical speaking, it's technology handling massive real-time data and streaming data, and that's of different variety. For instance the app for the mobile app, for system knock, the customer behavior, they click, and they click browsing of the digital image of each merchant and asking for the price and compare against another similar product. All these behaviors are translated as data, and this data will be further merged with the archived data and try to update the profile of this customer's interests, and then try to detect whether there's a good match of they current merchant with the customer intent. If the match is good, then will flash this to the top priority, the top spot. So that try to increase the conversion rate. So if the conversion rate is high, then our sales is high. So DT, data technology means everything to Alibaba. >> It's interesting, I find my observation here, it's so fascinating because in the old days, applications produced data, that was stored on drives. They'd go to data warehouses, and they'd analyze them. You guys, in Alibaba Cloud are doing something fundamentally different, that's exciting in the sense that you have data, people call it data exhaust or data in general, but you're reusing the data in the development in real-time. So it's not just data exhaust, or data from an application. You're using the data to make a better user experience and make the systems smarter and more intelligent. Did I get that right? >> Exactly, exactly. This is a positive feedback loop, in a way, so in the old-fashioned way, you archived the data for offline analysis and for post-event analysis, and trying to identify whether there's any room for improvement. But that's fine. But now people cannot wait, and we cannot wait. Offline is not enough. So we have to do this in real time, online, in a feedback version, in search of a way that we could capture exactly at the right moment, understand the intent of the customer, and then try to deliver the right content to the customer on the fly. >> Jackie Ma, or Jack Ma, your boss, and also Dr. Wong who I spoke with yesterday, talk about two things. Jack Ma talks about a new revolution, a new kind of industrial revolution, a smarter world, a better society. Dr. Wong talks about data flowing like a river, and you hear Hangzhou as an example, but it highlights something that's happening across the world. We're moving from a batch, slow world with data to one that's in motion and always real time. They're not necessarily mutually exclusive, but they're different. A data lake or a data river, whatever word you want, I don't really like the word data lake personally, I think it means, it's batch to me. But batch has been around for a while. Real time mixed streaming. This is something that's happening, and it's impacting the architecture and the value proposition of applications, and it's highlighted in Internet of Things, it's highlighted in examples that we're seeing that's exciting like the ET Brains. Can you share your view in your project around ET Brains, because that is not just one one vertical. It's healthcare, it's industrial, it's transportation, it's consumer, it's everything. >> Yeah, good question, so first of all I concur with you that data lake already exists, will continue to exist, because it's got its own value because our ET Brain for example, actually emerged from data lake, because it has to learn all the benchmark, the baseline model, the basic knowledge from the existing archive data, which is a data lake. However, that's not enough. Once you have the knowledge, you have the capability but you need to put that in action. So we are talking about data in motion, data in action. How do we do that? So once you have the training example, all the training data from data lake, and you train the brain, the brain is mature enough and in the next step you want to push the brain coupled with real-time streaming data, and then to generate real-time action in real-time manner in preemptive way, rather than posting in a reactive way. So for example, in transportation and travel, T and T, travel and transportation, and traffic management. So currently, all the authorities, they have access to real-time information, and then they do a post-event analysis if there's a traffic jam, and then they want to do some mitigation. However, the best scenario is, if you can prevent the traffic jam from happening in the first place, right, how can you foresee there will be, there would be, there could be traffic jam happen in 10 minutes from now, and then you take a preemptive strike, and then try to prevent that from happening. That's the goal ET Brain, in traffic management want to achieve. Like for example, you see the ambulance case, and once the ET Brain receives the message say the ambulance is going to go to Point A, pick up a patient, and carry that patient, rush them to Hospital B, and then it immediately calculates the right routing, the driving direction, and the calculate the ETA to every intermediate intersection and then try to coordinate with the traffic lights, traffic signal. All this systematic integration will create on demand a green wave for ambulance, but in the past ambulance is just by the siren, right. >> Yeah, this is fascinating, and also I'd like to get your thoughts, because you bring us something that's important, and that is, and I'd like to connect the dots for the audience, and that is, real time matters. If you're crossing the street, you can't be near real time, because you could get hit by a car. But also latency's important, also the quality of the data is good. I was talking to an executive who's laying out an architecture for a smart city, and he said, "I want the data in real time," and the IT department said, "Here it is, "it's in real time", and he says, "No, that's last year's data." And so the data has to be real time and the latency has to be low. >> Exactly. I completely agree. The latency has to be low. Unfortunately, in the current IT infrastructure, very often the latency exist. You cannot eliminate that, right? And then you have to live with that, so the ET Brain acknowledge the fact, in fact we have our own algorithm designed in a way that it can make a shortened prediction. So based on five minutes ago data, the data collected five minutes ago, and then it can project the next five minutes, next 10 minutes, what would be the data, and then use that to mitigate, or to conquer, to offset the latency. So we find that to be a good strategy, because it's relatively easy to implement, and it is fast and efficient. >> Dr. Wanli, fascinating conversation. I'd like to get your thoughts on connecting that big data conversation or data conversation to this event. This is a cloud computing event. We at theCUBE and SiliconANGLE and our Wikibon research team we go to all the events. But sometimes the big data events are about big data, Hadoop, whatever, and then you have cloud, talking about DevOPs, and virtual machines. This conference is not just a siloed topic. You have cloud computing, which is the compute, it's the energy, it's the unlimited compute potential, but it's also got a lot of data. You guys are blending it in. >> Exactly. >> Is that by design, and why is that important? >> It's by design. Actually, you cannot separate cloud from data, big data. Or you cannot talk big data without referring to cloud, because once the data is big, you need a huge computation power. Where does that come from? Cloud computing. So that means that data intelligence, all the value has to require a good technological tool to unleash the value. What's the tool? Cloud computing. For example, the first time IBM come up with a smart plan, a smart city, that's 2005 or 2006, around that time, there's no cloud computing yet, at the earliest emerging stage. And then we see what happens. And the smarter city and then gradually become IT infrastructure construction. But it's not DT, data technology. So they invested billions of dollars in the infrastructure level, and they collect so much data, but all the data become a burden to the government, to save, to archive the data or protect the data from hacking, right. Now, these days, if you have the cloud computing available, you can do real-time analytics to unleash the value in the first place, at the first moment you receive the data and then later on you know which data is more valuable, which data is of less value, and then you know how much you want to archive. >> Our Wikibon research team put out research this past year that said IT is no longer a department, it's everywhere, >> It's everywhere >> it supports your DT, data technology, it's a fabric. But one thing that's interesting going back to 2005 to now is not only the possibility for unlimited compute, is that now you're seeing wireless technologies significantly exploding in a good way, it's really happening. That's also going to be a catalyst for change. >> Definitely. >> What's your thoughts on how wireless connectivity, 'cause you have all these networks, you have to move data around, it has to be addressable, you have to manage security. That's a heavy load.\ what do you do, how are you guys doing that? >> Okay, very good question. We faced this challenge a couple of years ago, we realized that, because in Chinese domestic market, the users they are migrating from PC to mobile, and this create the mobile phone has wi-fi, right, so interacts with another AP, Access Point, right. So then how do we recognize our tracking, and recognizes ID identification, all this stuff, create huge headache to us, and this time, in this conference, we announce our solution for mobile, for mobile cloud. So what does that mean? So essentially, we have a cloud infrastructure product designed in order to do a real-time integration and do a data cleansing of the mobile data. I mean by mobile, and wireless as well. Wireless means even bluetooth, or even IoT, IoT solution also supported there. So this is a evolving process in the way. The first solution probably is less than perfect, but gradually, as we are expanding into more and more application scenario, and then we will amalgamate the solution and try to make it more robust. >> You guys have a good opportunity, and Alibaba Cloud certainly met with Karen Liu about the opportunity in North America and United States where I'm from. But Alibaba Cloud, and Alibaba Group, in the Alibaba Cloud has had a great opportunity, almost a green field, almost a clean sheet of paper, but you have a very demanding consumer base here in China. They're heavily on mobile as you pointed out, but they love applications. So the question I want to ask you is, and I'd love your thoughts on this. How do you bring that consumerization, its velocity, the acceleration of the changing landscape of the consumer expectation and their experience to small businesses and to enterprises? >> Ok, very good question. So user not just customer base, and the demanding customers in China trying to help us to harden our product, harden our solution, and to reduce the cost, the overall cost, and the economy of mass scale, economy of scale, and then once we reach that critical point, and then our service is inexpensive enough, and then the small and medium, SMB, small and medium business they could afford that. And in old days, SMB, they want to have access to high performance computing, but they do not have enough budget to afford the supercomputer. But these days now, because our product, our computation product, cloud product, big data product is efficient enough, so the total cost is affordable. And then you see that 80% of our customers of Alibaba, at least 80%, are actually SMB. So we believe the same practice can be applied to overseas market as well. >> You bring the best practices of the consumer and the scale of Alibaba Cloud to the small and medium-sized enterprises, and they buy as they grow. >> Exactly. >> They don't buy a lot upfront. >> Yeah, yeah, they buy on demand, as they need. >> That's the cloud, the benefit of the cloud. >> Exactly. >> Okay, the compute is great, you've got greatness with the compute power, it's going to create a renaissance of big data applications where you see that. What is your relationship with Intel and the ecosystem, because we see, you guys have the same playbook as a lot of successful companies in this open source era, you need horsepower and you need open source, what is Alibaba's strategy around the ecosystem, relationship with Intel, and how are you guys going to deal with partners? >> Yeah, first of all, so we really happy that we have Intel as our partner. In our most recent big data hackathon for the medical AI competition, and we just closed that competition, that data hackathon. Okay, very fascinating event, okay. Intel provided a lot of support. All the participants of this data hackathon, they do their computing leveraging on the Intel's products, because they do their image process. And then we provided the overall computing platform. Okay, this is a perfect example of how we collaborated with our technology partners. Beyond Intel, in terms of the ecosystem, first of all, we are open. We are building our ecosystem. We need partners. We need partners from pure technology perspective, and we also need partners from the traditional vertical sectors as well, because they provide us domain knowhow. Once we couple our cloud computing and big data technology with the domain knowhow, the subject matter expertise, well together the marriage will generate a huge value. >> That's fantastic, and believe me, open source is going to grow exponentially, and by 2025 we predict that it's going to look like a hockey stick. From the Linux foundation that's doing amazing work, you're seeing the Cloud Native Foundation. I want to get your thoughts on the future generation. >> Yeah, you mean open source? >> The future generation that's using open source, they're younger, you guys have tracked, you know the demographics in your employee base, you have a cloud native developer now emerging. They want to program the infrastructure as they go. They don't want to provision servers, they want the street lights to just work, whatever the project, the brains have to be in the infrastructure, but they want to be creative. You're bringing two cultures together. And you've got AI, it's a wonderful trend, machine learning is doing very well. How do you guys train the younger generation, what's your advice to people looking at Alibaba Cloud, that want to play with all the good toys? You got machine learning, you got AI, they don't want to necessarily baby, they don't want to program either. They don't want to configure switches. >> Yeah, very good question. Actually this is related to our product strategy. So in a way, like today we announce our ET Brain, so we are going to release this and share this as a platform to nurture all the creative mind, creative brains, okay, people, trying to leverage on this brain and then do the creative job, rather than worry about the underlying infrastructure, the basic stuff. So this is that part which we want to share with the young generation, tell them that unleash your creativity, unleash your imagination, don't worry about the hard coding part, and we already build the infrastructure, the backbone for you. And then image anything you think possible and then try to use ET Brain, try to explore that. And we provide the necessary tool and building blocks. >> And the APIs. >> And the APIs as well, yes. >> Okay, so I want to get your thoughts on something important to our audience, and that is machine learning, the gateway to AI. AI, what is AI? AI software, using cloud. Some will argue that AI hasn't really yet come on the scene but it's coming. We love AI, but machine learning is really where the action is right now, and they want to learn about how to get involved in machine learning. So what's your view on the role of machine learning, because now you have the opportunity for a new kind of software development, a lot of math involved, that's something that you know a lot about. So is there going to be more libraries? What's your vision on how machine learning moves from a bounded use case to more unbounded opportunities, because, I'm a developer, I want the horizontally scalable resource of the cloud, but I'm going to have domain expertise in a vertical application. So I need to have a little bit of specialism, and I want the scalability. So data's got to move this way and it's got to be up this way. >> Yes, yeah, okay, let me put it this way. So first off, for people who are really interested in AI, or they want to work on AI, my recommendation first of all, you got to learn some mathematics. Why, because all the AIs and machine learnings is talking about algorithms, and those algorithms are actually all about math, mathematics, the formula, and also the optimization, how to speed up the convergence of the algorithms, right. So all this maths is important, okay. And then if you have that math background, and then you have the capability to judge or to see next, which algorithm, or which machine software is suitable to solve the vertical problems. Very often the most popular algorithm may not be the right one to solve the specific vertical problems. So you're going to the way, capability to differentiate and to see that and make the right choice. That's the first recommendation. The second recommendation, try to do as many type of examples as possible, try to get your hands on, don't stop at looking at the function specification and oh, this is a function and input, output, da da da, but you need to get your hands dirty, get your hands on the real problem, the real data. So that you can have a feeling about how powerful it is or how bad or how good it is. Once you have this kind of experience, and then you do have capability, you gradually build up a cumulative capability to make a right choice. >> This is fascinating, Dr. Wanli, this is fantastic. I want to follow up on that because you're bringing up, in my mind I can almost see all these tools. There's an artisan culture coming on. You're seeing that. Dr. Wong discussed that with me yesterday. Artisans meeting technologists, scientists and creatives. UI, we're seeing evolutions in user experience that's more art. And so culture's important. But the machine learners of the algorithms, sometimes you have to have a lot of tools. If you have one tool, you shouldn't try to use tools for other jobs. So bring this up. How should a company who's architecting their business or their application look at tooling, because on one hand, there's the right tool for the right job, but you don't want to use a tool for a job that it's not designed for. To your point. Tools, what's your advice and philosophy on the kinds of toolings and when to engage platforms, relationship between platforms and tools. >> Okay, then put it this way. So, this is a decision based on a mixture of different criteria together. So first of all, from technology perspective, and secondly from the business perspective. From technology perspective I would say if your company's critical competence is technical stuff, and then you've got to have your own tool, your own version. If you only rely on some existing tool from other companies, your whole business actually is dependent on that, and this is the weakest link, the most dangerous link. But however, very often to develop your own version of the tool takes forever, and market wouldn't give you so much time. And then you need too strike a balance, how much I want to get involved for self development and how much for in-house development, and it's how much I want to buy in. >> And time. >> And time as well, yes. And another one is that you've got to look at the competitive landscape. If this tool actually has already existed for many years and many similar product in the market, and the problem is not a good idea to reproduce or reinvent, and then you're going to why not buy it, you take that for granted. And it think that's a fact, and then you build a new fact, right. So this is another in terms of the maturity of the tool, and then you need to strike a balance. And in the end, in the extreme case, if your business, your company is doing a extremely new, innovative, first of a kind study or service, you probably need some differentiate, and that differentiator probably is a new tool. >> Final question for you. For the audience in America, in Silicon Valley, what would you like to share from your personal perspective about Alibaba Cloud that they should know about? Or they might not know about and should know about. >> Okay, 'cause I worked in the US for 16 years. To be frank, I knew nothing about Alibaba until I came back. So as a Chinese overseas, I'm so ignorance about Alibaba until I came back. So I can predict, I can guess, more or less, in the overseas market, in US customers, they probably know not that much about Alibaba or Alibaba Cloud. So my advice and from my personal experience, I say, first of all, Alibaba is a global company, and Alibaba Cloud is a global company. We are going to go global. It's not only a Chinese company, for example. We are going to serve customers overseas market in Europe and North America and Southeast Asia. So we want to go global first. And second, we are not only doing the cloud. We are doing blending of cloud and big data and vertical solutions. I call this VIP. V for vertical, I for innovation. P for product. So VIP is our strategy. And the innovation is based upon our cloud product and big data product. >> And data's at the center of it. >> Data is the center of this, and we already got our data technique, our data practice from our own business, which is e-commerce. >> And you're solving some hard problems, the ET Brain's a great playground of AI opportunity. You must be super-excited. >> Yeah, yeah, right, right, okay. >> Are you having fun? >> Yes, a lot of fun. Very rewarding experience. A lot of dreams really come true. >> Well, certainly when you come to Silicon Valley, I know you have a San Mateo office, we're in Palo Alto, and this is theCUBE coverage of Alibaba Cloud. I'm John Furrier, co-founder of SiliconANGLE, Wikibon research and theCUBE, here in China covering the Alibaba Cloud, with Dr. Wanli, thanks for watching.

Published Date : Oct 26 2017

SUMMARY :

brought to you by Intel. it's the biggest cloud computing conference here in China. We have seen a lot of data in the conversation here So if the conversion rate is high, then our sales is high. and make the systems smarter and more intelligent. so in the old-fashioned way, you archived the data and it's impacting the architecture and in the next step you want to push the brain and the latency has to be low. And then you have to live with that, it's the energy, it's the unlimited compute potential, in the first place, at the first moment you receive the data That's also going to be a catalyst for change. it has to be addressable, you have to manage security. and do a data cleansing of the mobile data. So the question I want to ask you is, and the demanding customers in China and the scale of Alibaba Cloud to the because we see, you guys have the same playbook All the participants of this data hackathon, and by 2025 we predict that it's going to the infrastructure, but they want to be creative. and then try to use ET Brain, try to explore that. and that is machine learning, the gateway to AI. and then you have the capability to judge for the right job, but you don't want to use a tool and secondly from the business perspective. and the problem is not a good idea to reproduce what would you like to share from your personal perspective And the innovation is based upon our cloud product and we already got our data technique, the ET Brain's a great playground of AI opportunity. Yes, a lot of fun. here in China covering the Alibaba Cloud,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
AlibabaORGANIZATION

0.99+

AmericaLOCATION

0.99+

WanliPERSON

0.99+

IBMORGANIZATION

0.99+

USLOCATION

0.99+

2006DATE

0.99+

ChinaLOCATION

0.99+

Karen LiuPERSON

0.99+

John FurrierPERSON

0.99+

Palo AltoLOCATION

0.99+

Silicon ValleyLOCATION

0.99+

WongPERSON

0.99+

Alibaba CloudORGANIZATION

0.99+

Alibaba GroupORGANIZATION

0.99+

80%QUANTITY

0.99+

16 yearsQUANTITY

0.99+

2005DATE

0.99+

EuropeLOCATION

0.99+

United StatesLOCATION

0.99+

North AmericaLOCATION

0.99+

two thingsQUANTITY

0.99+

Min WanliPERSON

0.99+

SiliconANGLEORGANIZATION

0.99+

second recommendationQUANTITY

0.99+

Cloud Native FoundationORGANIZATION

0.99+

Jackie MaPERSON

0.99+

San MateoLOCATION

0.99+

Jack MaPERSON

0.99+

Southeast AsiaLOCATION

0.99+

WikibonORGANIZATION

0.99+

IntelORGANIZATION

0.99+

yesterdayDATE

0.99+

secondQUANTITY

0.99+

one toolQUANTITY

0.99+

Cloud CityLOCATION

0.99+

2025DATE

0.99+

last yearDATE

0.99+

Dr.PERSON

0.99+

billionsQUANTITY

0.99+

first recommendationQUANTITY

0.99+

five minutes agoDATE

0.99+

first timeQUANTITY

0.98+

theCUBEORGANIZATION

0.98+

HangzhouLOCATION

0.98+

Alibaba CloudEVENT

0.98+

two culturesQUANTITY

0.98+

DrPERSON

0.97+

firstQUANTITY

0.97+

Hangzhou, ChinaLOCATION

0.97+

first solutionQUANTITY

0.97+

todayDATE

0.97+

ET BrainORGANIZATION

0.96+

LinuxORGANIZATION

0.96+

SiliconANGLE MediaORGANIZATION

0.93+

Feature: Alibaba Cloud Founder; Dr Wang, Alibaba | The Computing Conference 2017


 

>> SiliconANGLE Media presents ... theCUBE! Covering AlibabaCloud's annual conference. Brought to you by Intel. Now, here's John Furrier... >> Hello everyone, I'm John Furrier, the co-founder of SiliconANGLE, Wikibon, and theCUBE. We are here for an exclusive Cube conversations at the Alibaba Cloud conference here in Hangzhou, China. We're here with Dr. Wang, who's the chairman of the Alibaba Group Technology Committee as well as the founder of Alibaba Cloud, here in the new Museum of Inspiration at the event. Thanks for spending the time with me. >> Thank you for coming. >> So before we talk about Alibaba Cloud and all the goodness going on here at the conference, talk about this Museum of Inspiration. It is new, and it has kind of a display theme. You kind of walk through time. What was the motivation and the inspiration for the museum? >> Yeah, I think the keyword for the museum, inspiration, is really the inspiration that started the museum. I would say that there's two, really the artists thinking about that. The first thing is really about when people, people take a lot of things for granted. One of the goals for this museum, it just shows the people they probably see every day. But just let them, just, wow, okay, that's different from what I thought. I think a lot of people take for granted, but it's really a great invention, a great human contribution to the whole society. I think that one thing is really about that people understand why we got here today. So that's the first thing. The other thing is really about science and technology. When people are talking about science and technology, people often will say, whether we can combine science and technology. But I don't think that's the right way to describe the relationship between science and technology. I would say science and technology, really the two sides of the coin. I really want to see, let people to see two sides instead of mixing together and got one thing. So that's two things that's parallel, just like zero and one. They are two things. When they're put together in a computer, amazing things happen. If you mix the zero and the one, like half something, then it's just not that fun. So I really want to make sure it's the museum of science and art instead of the mixture of science and arts. So that's the one thing. The other thing is really about the inspiration of future. Most of the museum is really about the past, to show how we have in the past, and with less on the inspiration to help people to think about the future. This museum is really, when we think about everything over here, we did talk about the past, but we want to make sure people think about the future. That's the whole idea about the museum. >> And the computer industry is fairly young, if you go back to modern computing. But you kind of have a take here about how technology really is embedded in life. Talk more about that impact 'cause that seems to translate to the conference here at Alibaba, that technology isn't just about the speeds and the feeds, it's about the integration into life. >> Yeah, and I think that from this museum we can see actually I trace back the origin of all the technology. When people are talking about the computer technology, I really want to talk about the computing technology. And then we can trace back, see actually the human is the first signal computing device. Our Mother Nature created for us. If you look at the same things differently, you really can see the origin of that. I think in this museum we talk about two really original things. The first is about the nature origin of the Internet. When talking about Internet, people talk about our current technological infrastructure of Internet. When you look at the human history, how is when people walk, you create an Internet for Earth? You can see a lot of things can trace back. Then, with this kind of trace back, you can help us to think about what's going to happen next. The trace of the original idea is actually very important if you're thinking about technology. >> Talk about the story of Alibaba Cloud. That is not, It's new, Amazon has had it for around, early 2000's. But you guys came right after Amazon, 2009. Still young and growing. How does the Alibaba Cloud take the culture of this inspiration? What are some of the design principles of the Alibaba Cloud? >> Actually I would say the Alibaba Cloud is different from the Amazon Cloud. In the sense we have different vision about the future. Unfortunately though, we are put under the same umbrella called cloud computing by media, I will say that. So we are different, in the sense when the Amazon, actually I show great respect to Amazon. When Amazon started cloud computing, they are really talking about the utility. They're talking about how to cut the cost down. So basically, they start with the low cost of IT infrastructure. That's what I understand. When I started Alibaba Cloud, we know that actually cost is important for sure. But we know that actually the computing part is more important than the cost if you're thinking about the big data era. We started with thinking it's the acentric cloud computing. When you look at our first brochure and we put those words over there. That's almost nine years ago. We called it acentric cloud computing. Instead of the IT-centric cloud computing. This actually, it's not just an idea difference. It's actually, eventually, influenced of the underlying technology infrastructure. Our whole underlying technological infrastructure is designed for the data, instead of just for the IT deployment. >> Jack Ma was talking about this industrial revolution, this digital transformation. What strikes me is you guys have that same art and scientist dynamic, art and science coming together, reminds me of the Steve Jobs technology liberal arts thinking that spawns new creativity. Certainly the iPhone is a great example of that as one of the many things. But now the new generation is coming together. You have a big artist focus here at the event. Music festival, not just technology. How is that part of the focus at the event here? What does that mean for new developers? >> I think it's really the crossing behind that. If you're thinking about technology and now e-commerce, what's really the one thing behind that that's really changed the way of peoples' lives? Computing in that sense, computing is not just technology. It's really something that changes the way of life of every people. I think the e-commerce change the way of life of every people. In that sense, they are the same. If you look at the peoples' lives, they won't just live on technology. They won't just live on the arts. They need a life, love means everything. By nature, we have to make sure as consumers, they need something more than just one thing. I think we are very lucky we understand that. If you're thinking about the young people, I will give you a few numbers about this conference about young people. In China, we have a very specific word talking about the young people a couple of years ago. We call the 'badiho'. It basically means the generation born of the '80s. When people talk about 'jodiho', that basically means people born after '90s. And then people talking about the 'leniho' it's basically people born after 2000. I think that most of the visitors for this conference are 'leniho', 'jodiho', and 'badiho'. These are all young, all young people. >> The digital culture. >> It's a digital culture. I would rather use my own word in the book I would say instead of digital. For me, digital generation is already an old generation I would say. I would like to call this the online generation. They do everything online. Even the last generation do a lot of things digital because digital is everywhere. But I want to emphasize it's an online generation. They do everything online. >> Dr. Wang, talk about data. You mentioned that's the key ingredient, the fuel for innovation. That's impacting the city brain project you guys are doing. Talk about the city brain and the role of data and how that's impacting the societal users out there certainly here in China, the traffic is crowded. This is just an example of what else is out there. >> Okay. City brain actually it's, again it means different things based on the perspective. One thing that's probably important is the data. This is first time actually I think instead of using the big data, it's better to using what I call the data results. It's a better word than big data. I think the one fundamental thinking for the city brain is we find a human army. Humans finally realize actually that data results is the most precious resource for the city, instead of land and water supply. Because we already know that the land is limited. The water supply is limited. This is very important. It doesn't view data as a non-essential thing. It's just a part of your IT system. We finally realize that data is part of the city instead of part of your city IT system. I think it's a leap frog thinking, at least for me. When it got to that, and you realize that today all the existing IT systems cannot actually really embrace the data. IT system is just to support the people doing the work they used to do. And then you realize we need an infrastructure to really make the value from the data. Just like we have water supply for the city, then you can use the reservoir. Otherwise, the reservoir is useless for the city. I think city brain is just like a water supply system for the data. The city eventually can consume that. We start thinking it's a new infrastructure for the city just like water supply system, just like power grid, just like any way system. That's how we're thinking about it. This is the first thing. The reason we got to the traffic system is this is the problem every city has around the world. From my yesterday's presentation, I just joked about we build two roads for the city, which is too many. I was thinking a lot of people realize it. That's why Boston had the project. They want to get all the roads under the surface. Under surface. But it's still a road. It's still expensive. You know how much money they spend just to move all the roads. >> The big dig, I remember, that was the-- >> Yes, that's a big dig. I don't think that's, that's good for the transportation system, but I don't think that's the number one way for the growth of the city. I think probably most of the city don't have the money to do that. What the data city brain wants to do whether we can take the resource of data and we can optimize every aspect of the city so we can use less resource to support city growth. When we start with the traffic, it's just to make sure, you know that when we use the data to optimize the traffic lights, the idea behind that actually we use the data to optimize the time. How to just read the time. It's not just lights. And then if you're thinking, when we show the eventually, if you have enough data, then we can have less roads in the city but still got the same. >> So the Internet of Things is the hottest trend. 0bviously machine learning and artificial intelligence are part of that, and the cloud powers this new edge of the network, and the data has to flow. So the question that a lot of technologists who are architecting these solutions ask is how do you make the data go at a very low latency? That takes compute power. That takes a lot of technology. How does Alibaba Cloud think about the architecture? Obviously you have a strategic partner like Intel, Obviously with a lot of compute power. You got to think differently around making the data move. If it's like water, it needs to flow. So real time is really important, but self-driving cars, real time is down to the millisecond, nanosecond. How do you think about that as a technologist? >> I think the, if you go back to the Internet of things, I think it's still the Internet. I would say eventually, if you're thinking about the word cloud computing and people use edge computing and people talking about Internet of things. For me, it's just the computing of the Internet. Cloud computing is the computing of the Internet and edge computing is computing on the Internet. Even the IoT is the computing of the Internet. If you're talking about the data, I think eventually it's really about the data on the Internet. It's not data on the sensor. It's not data on the cloud. Basically it's data on the Internet. I would expect eventually the Internet infrastructure will be improved significantly. It's not an improved cloud. It's not improved edge computing. Or it's improvement of the IoT. But it's really, >> Together. >> it's together. >> So Intel, I was covering them, Mobile World Congress earlier in the year. And obviously five G. You need the mobile overlay, that's super important. You also have the end-to-end inside the cloud. Obviously Intel is a strategic partner. Can you talk about the relationship you have with Intel? And also your philosophy, technically speaking, with the ecosystem? Because it's not just Intel, it's everybody. There's a lot of people here at this event. American companies as well as international companies who are now going to be part of your ecosystem. >> Actually the, we certainly have a very good relationship with Intel. I think we share in some sense the same vision. I think that the number one thing is really about people learning about how important the computing is. For me, the Intel is not that, a chip selling company. Intel is really the provider of computing power. That's what I understand. And we can expect eventually the whole ecosystem is really about who is going to provide the computing power. Who is going to provide the infrastructure to make the data? Instead of just equipment supply, eventually the need for computing, and the need for data, will be the challenge for every company, including Alibaba Cloud. We are not, we are not immune from these challenges. We will feel the same challenge. What we want to do is really make sure that with all these partners, provide enough computing for the next 10 or 20 years. We want to make sure that there's enough data flow for the next 10 years. In that sense, it's not the traditional ecosystem as like you do this and you do that. It's basically how we can work together to really make sure we have the challenge for the data and computing in the next 10 years. >> Yesterday we covered the news that you guys announced 15, building and R&D over the next three years, which is a lot of money. Also it has a very international and global view. Academics with younger folks. Alibaba Cloud is going to be a part of that, I'm assuming. I'd love to get your thoughts on how you see that intersecting. But the question for you is the cloud world today is moving at very, very fast speed. We're seeing Amazon, for instance, has been the best in terms of new announcements every year. Not one or two, like a ton of announcements, a lot. How are you guys going to continue to keep the pace? To move faster because the city brain is a great project, but it's going to have more evolution. It's going to move fast. How are you guys keeping up with the pace? >> I think the only way, that's not just for the next 10 years. Actually when I started Alibaba Cloud, we take the same philosophy. Actually the user moves very fast than us. If you look at the users in China they move very fast probably than anywhere else around the world. If you use the city brain project, I would say city brain project is basically tell the people, we need the computing power more than any other task. You really can see that. People want you. If you can't satisfy their demand, then somebody else is going to do that. It's not something we want to move fast but >> You have to move fast. >> You have to move fast. That's why the China is special. I want to say China is not just a place for the market. China is the place that pushes you to move faster. That's more important than market size. >> You mentioned data technology and information technology kind of transferring to a new world. Software is also a big part of it. Software you have to compute, obviously with Intel and the relationships you have. But software is growing exponentially. Certainly in open source, we see Cloud Native Foundation here. They'll probably have Linux foundation. Open source is going to grow exponentially. Most of the code will be shipping. But you have more data growing exponentially. Software is eating the world, but data is eating software. That means data is greater than software. If you look at it that way, that's super important. As the new architects, you and I were just talking about how we've in the industry for a while. You certainly have an amazing career from Microsoft now at Alibaba. A new generation of architects and developers are going to create new innovations around this dynamic of data. What's your advice and how do you view that if you are 21 years old again right now and you were going to jump into studies and academic and or field. It's a whole new world. >> I think there's probably two suggestions. Not necessarily for the young generation, but I would say it's just a suggestion for the young generation to push that habit. The first thing you mentioned about the data eats software. Well, I would put it in a different perspective. I would say for the last generation, the last two or three generations, I would say the computer era, we are really talking about the computer software. That's pretty much in everything. For this generation, I would say we are talking about computing plus the data. That box is not important, but the computing power is more important. For the computing era, the box is important. >> There's no box. It's the world, it's the cloud. >> That's one thing. The implication for this, I want the young generation to push is, then we need the new infrastructure. Thinking about the build as a great vision, got to have the computer in every home. That's infrastructure. Today when you are in the computing process data era, the infrastructure is not there. I think the vision for the Alibaba Cloud is make sure that we have this infrastructure for the next 10 or 20 years so the young generation can take advantage of that and to do that innovation and inventions, just like computing in every home. >> That's very important. I think that also speaks to businesses, how enterprises, I remember my first start up, I had to buy all this equipment and put it into the telephone closet. Now, start ups and small businesses don't need IT departments. This has been a big growth area certainly for Alibaba Cloud. But now all businesses might have a small closet, not a big data center. This is going to change the nature of business. So work and play are coming together. This speaks to the Museum of Inspiration theme here where you can have work and play kind of integrate but yet still be separate in that analog digital world. What's your vision on this new dimension of everything doesn't have to be just digital? You can have an analog life and mix it with digital. >> Actually I was always sad. It's not, the world never has just one side. It always has two sides. The difference is which side is important at a particular time. Just like when people talk about digital and analog, the analog will exist forever. It's hard for you to kill. The question is whether you can find the most beauty from the digital at the same time you can most beautiful part of the analog. I would say that the people, just like when talking about software, people still loved the hardware. And people still loved the touch. The digital has to make sure it looks good. Will it work versus it looks good? I would say we want to make sure people live in a world with two sides, instead of just giving them one side of the world. >> You mentioned people still love hardware. I always say, a car drives but there's still an engine, and people like to understand the engine. There's a maker culture in the United States that's been growing over the past two decades. And now even more accelerated is the maker culture because of the edge and how technology has become part of the fabric of life. How do you see that maker culture being enabled by more cloud services? Because anyone can make a skateboard or motorcycle or a computer or a device now. Powering that with the cloud is an opportunity. How do you view that? >> I would say that eventually, if we have the broad definition of a cloud, I would say eventually, everything the maker makes will be part of the cloud. When talking about clouds, we're really talking about Internet, so every hardware, every piece of hardware will be part of Internet. I would say, if you look at the evolution of the Internet, Internet, it's just a backbone at the very beginning. Actually the first revolution the Internet made is really to make sure that every piece of software is a part of Internet. That's how we got the world wide web. I would say when talking about the maker culture, I would say eventually that every piece of hardware will be part of Internet. So Internet won't be complete without the hardware. In that sense, the cloud is a really essential part of that. >> There's some really interesting things happening here in China that I'm excited about. One of them is the nature of the user base and how close you guys are to that. In the US a similar scale but it's kind of spread over a bunch of other cloud providers. But the interesting phenomenon as data grows exponentially, as software grows exponentially in open source, things are becoming more decentralized. Without talking about the whole initial coin offerings, I know China has banned it and Russia's going to ban it. Other countries are putting a clamp down on crypto currency. Putting that aside, there's still blockchain as a potential disruptive enabler. You're seeing decentralization becoming a new architecture dynamic because you have to support the growth of these devices at the edge. Distributive computing has been around for a while, but now a decentralized architecture dynamic exists. How do you steer that technology direction? >> You have to separate from the the distributive architecture versus its physical location. I would say I like the blockchain idea very much. I think eventually it would be part of the Internet. It's not just something that sits on top of the Internet. It would be very fundamental, just like TCP and IP. This is low level, so this would be part of the Internet instead of standing on top of the Internet. Eventually, in that sense, Internet would be very distributed. By thinking that it's nothing, there's no decentralization exists. You still need, even though physically, it's in one place. >> It's almost decentralized, not 100 percent. >> Yeah, yeah. Obviously this would be different. Without Internet, without new software, that basically, just like PC. PC is really in a single box, and we use all software in a box. We distribute architecture. We could have decentralized, but everything actually is distributed. You still cannot trace that. You put like a meeting. A service in a data center. It's actually distributed over this one meeting service. In that sense, it's completely distributed. >> That server list too is a big trend where if you talk about the edge of the network, you got to move compute to the data sometimes. Or have compute on the edge. So this is going to be continued growth, you see that as well right? >> Yes, but I still think, if you use Silicon as a measure for this computing power, I would say if you can see there's more silicon on the edge, but I would say when we put one silicon on the edge, you probably have to put 100 silicons on the cloud. It's still kind of-- >> It's a relationship. >> It's a relationship, just like our body is very important but the brain consumes the most oxygen. >> It's important what's in the cloud then. You got to have the computing, have those ratios. It depends on the architecture. >> Yes, yes. >> Final question for you is as the folks in Silicon Valley, where we're based, and Palo Alto want to know is Alibaba, what it means to them? If you have a chance to say a few things about what Alibaba Cloud is to America, what would you like to say? >> I would say that actually they would just put the cloud computing aside. Just look at what it really means behind that. I think the cloud, we do have an understanding of what the cloud computing really means. At the very beginning actually, I wouldn't call the company a cloud computing company. I would call it a general computing company. It's really a fraction of what's thinking in China. Again, my comment is not just to view China as the market to sell your product. To view China as the place to inspire having a new product. >> And it's a global world now, the world is flat. >> Yes, just like United States, it's not, it's a place inspired. All the people around the world together to have a new idea. I think the people in China just love new things. They love to try new things. It really can shoot your size of your innovation. >> And it's a global collaboration, it's interesting. That phenomenon is going to continue. You've done amazing work here. Congratulations on the Museum of Inspiration and the projects you're working on. Personal question for you. What are you excited about now? We kind of joke about how old we are now, but the young people certainly have a great future ahead of them. But you have a lot of experience and you're steering Alibaba's technology committee across the group as well as being the founder of the cloud. What are you excited about right now, technically speaking? What's the big, or just impact? What's the big wave that you like? >> I think it's very exciting in a couple of things, three things I would say. The first is really about just look at technology itself. Just like when I described my book, it's really, really exciting in your life if you can see the Internet plus the computing and plus data, cause they're together. Just like you have this engine, you have the airplane, a couple of things, they're together wherever. This is a very, very exciting era. This is not just about a technology era. It's an era that all things happen at the same time, so that's very exciting. That's one thing. The second thing as you read about the city around over here, I think the the Alibaba the Hanzo, it's just a very special for Alibaba, but I think it's special for the other company as well. So this place is very special. Just to give you an idea where you are, this area has the most networked river in the past. If you look at the map, it's like Internet. I would say, all the people over here, just their mindset. It's just on an Internet mindset. Even goes back 100, 200 years ago because the river is the only way for them to travel, for the communications-- >> That's the data back then. >> That's exactly my point, see. If you look at the map, so this is very exciting. The other thing about that the Alibaba, for me, the Alibaba you know Alibaba, we have a very broad opinions. You can feel that. From a technology point of view, that basically means it's the place you can touch every aspect of technology. You have a very slight, very-- >> You have a great surface area aperture to look at impact of life. >> So you think about these three things together. It's hard to say the, you better get excited. >> It's a great time to be in technology, isn't it? Entertainment, e-commerce, web services. >> For me, when I work on the city brain project, it's just the beginning of machine learning. A lot of people, they are fighting for like, when people talk about speech recognition, they are fighting for the last one meter for the speech recognition. But if you're talking about city brain, it's the world. The most big AI project. And it's just the beginning. We just start with the one percent. >> It must be a lot of fun. You got a lot of data to work with. You have real life integration. It's super exciting. When are we going to see you in Silicon Valley? >> I appear regularly to Silicon Valley two or three times every year. We'll probably see sometime early next year. >> Thank you very much for the time, appreciate it. >> Thank you for coming to the conference and coming to the museum. >> Thank you very much for your inspiration. >> Thank you. >> Thank you.

Published Date : Oct 26 2017

SUMMARY :

Brought to you by Intel. We are here for an exclusive Cube conversations at the Alibaba Cloud conference here in Hangzhou, So before we talk about Alibaba Cloud and all the goodness going on here at the conference, Most of the museum is really about the past, to show how we have in the past, and with that technology isn't just about the speeds and the feeds, it's about the integration The first is about the nature origin of the Internet. How does the Alibaba Cloud take the culture of this inspiration? It's actually, eventually, influenced of the underlying technology infrastructure. How is that part of the focus at the event here? It's really something that changes the way of life of every people. Even the last generation do a lot of things digital because digital is everywhere. That's impacting the city brain project you guys are doing. We finally realize that data is part of the city instead of part of your city IT system. optimize every aspect of the city so we can use less resource to support city growth. So the Internet of Things is the hottest trend. Cloud computing is the computing of the Internet and edge computing is computing on the Internet. You also have the end-to-end inside the cloud. In that sense, it's not the traditional ecosystem as like you do this and you do that. But the question for you is the cloud world today is moving at very, very fast speed. Actually the user moves very fast than us. China is the place that pushes you to move faster. As the new architects, you and I were just talking about how we've in the industry for That box is not important, but the computing power is more important. It's the world, it's the cloud. I think the vision for the Alibaba Cloud is make sure that we have this infrastructure This speaks to the Museum of Inspiration theme here where you can have work and play kind It's not, the world never has just one side. And now even more accelerated is the maker culture because of the edge and how technology Actually the first revolution the Internet made is really to make sure that every piece Without talking about the whole initial coin offerings, I know China has banned it and I think eventually it would be part of the Internet. PC is really in a single box, and we use all software in a box. So this is going to be continued growth, you see that as well right? silicon on the edge, you probably have to put 100 silicons on the cloud. It's a relationship, just like our body is very important but the brain consumes the It depends on the architecture. I think the cloud, we do have an understanding of what the cloud computing really means. All the people around the world together to have a new idea. What's the big wave that you like? the Internet plus the computing and plus data, cause they're together. If you look at the map, so this is very exciting. It's hard to say the, you better get excited. It's a great time to be in technology, isn't it? And it's just the beginning. When are we going to see you in Silicon Valley? I appear regularly to Silicon Valley two or three times every year.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
AlibabaORGANIZATION

0.99+

AmazonORGANIZATION

0.99+

ChinaLOCATION

0.99+

AmericaLOCATION

0.99+

Silicon ValleyLOCATION

0.99+

Jack MaPERSON

0.99+

twoQUANTITY

0.99+

Steve JobsPERSON

0.99+

John FurrierPERSON

0.99+

two thingsQUANTITY

0.99+

OneQUANTITY

0.99+

MicrosoftORGANIZATION

0.99+

two sidesQUANTITY

0.99+

oneQUANTITY

0.99+

WangPERSON

0.99+

USLOCATION

0.99+

2009DATE

0.99+

Cloud Native FoundationORGANIZATION

0.99+

United StatesLOCATION

0.99+

one sideQUANTITY

0.99+

EarthLOCATION

0.99+

100 siliconsQUANTITY

0.99+

first timeQUANTITY

0.99+

Alibaba Group Technology CommitteeORGANIZATION

0.99+

IntelORGANIZATION

0.99+

second thingQUANTITY

0.99+

ntelORGANIZATION

0.99+

iPhoneCOMMERCIAL_ITEM

0.99+

firstQUANTITY

0.99+

first brochureQUANTITY

0.99+

Hangzhou, ChinaLOCATION

0.99+

two suggestionsQUANTITY

0.99+

one percentQUANTITY

0.99+

ChinaORGANIZATION

0.99+

100 percentQUANTITY

0.99+

first thingQUANTITY

0.99+

yesterdayDATE

0.99+

todayDATE

0.99+

SiliconANGLEORGANIZATION

0.99+

15QUANTITY

0.99+

RussiaORGANIZATION

0.98+

Mobile World CongressEVENT

0.98+

one thingQUANTITY

0.98+

one thingQUANTITY

0.98+

zeroQUANTITY

0.98+

WikibonORGANIZATION

0.98+

three thingsQUANTITY

0.98+

Alibaba CloudEVENT

0.98+

three generationsQUANTITY

0.98+

two roadsQUANTITY

0.97+

Alibaba CloudORGANIZATION

0.97+

One thingQUANTITY

0.97+