Robert Nishihara, Anyscale | AWS Startup Showcase S3 E1
(upbeat music) >> Hello everyone. Welcome to theCube's presentation of the "AWS Startup Showcase." The topic this episode is AI and machine learning, top startups building foundational model infrastructure. This is season three, episode one of the ongoing series covering exciting startups from the AWS ecosystem. And this time we're talking about AI and machine learning. I'm your host, John Furrier. I'm excited I'm joined today by Robert Nishihara, who's the co-founder and CEO of a hot startup called Anyscale. He's here to talk about Ray, the open source project, Anyscale's infrastructure for foundation as well. Robert, thank you for joining us today. >> Yeah, thanks so much as well. >> I've been following your company since the founding pre pandemic and you guys really had a great vision scaled up and in a perfect position for this big wave that we all see with ChatGPT and OpenAI that's gone mainstream. Finally, AI has broken out through the ropes and now gone mainstream, so I think you guys are really well positioned. I'm looking forward to to talking with you today. But before we get into it, introduce the core mission for Anyscale. Why do you guys exist? What is the North Star for Anyscale? >> Yeah, like you mentioned, there's a tremendous amount of excitement about AI right now. You know, I think a lot of us believe that AI can transform just every different industry. So one of the things that was clear to us when we started this company was that the amount of compute needed to do AI was just exploding. Like to actually succeed with AI, companies like OpenAI or Google or you know, these companies getting a lot of value from AI, were not just running these machine learning models on their laptops or on a single machine. They were scaling these applications across hundreds or thousands or more machines and GPUs and other resources in the Cloud. And so to actually succeed with AI, and this has been one of the biggest trends in computing, maybe the biggest trend in computing in, you know, in recent history, the amount of compute has been exploding. And so to actually succeed with that AI, to actually build these scalable applications and scale the AI applications, there's a tremendous software engineering lift to build the infrastructure to actually run these scalable applications. And that's very hard to do. So one of the reasons many AI projects and initiatives fail is that, or don't make it to production, is the need for this scale, the infrastructure lift, to actually make it happen. So our goal here with Anyscale and Ray, is to make that easy, is to make scalable computing easy. So that as a developer or as a business, if you want to do AI, if you want to get value out of AI, all you need to know is how to program on your laptop. Like, all you need to know is how to program in Python. And if you can do that, then you're good to go. Then you can do what companies like OpenAI or Google do and get value out of machine learning. >> That programming example of how easy it is with Python reminds me of the early days of Cloud, when infrastructure as code was talked about was, it was just code the infrastructure programmable. That's super important. That's what AI people wanted, first program AI. That's the new trend. And I want to understand, if you don't mind explaining, the relationship that Anyscale has to these foundational models and particular the large language models, also called LLMs, was seen with like OpenAI and ChatGPT. Before you get into the relationship that you have with them, can you explain why the hype around foundational models? Why are people going crazy over foundational models? What is it and why is it so important? >> Yeah, so foundational models and foundation models are incredibly important because they enable businesses and developers to get value out of machine learning, to use machine learning off the shelf with these large models that have been trained on tons of data and that are useful out of the box. And then, of course, you know, as a business or as a developer, you can take those foundational models and repurpose them or fine tune them or adapt them to your specific use case and what you want to achieve. But it's much easier to do that than to train them from scratch. And I think there are three, for people to actually use foundation models, there are three main types of workloads or problems that need to be solved. One is training these foundation models in the first place, like actually creating them. The second is fine tuning them and adapting them to your use case. And the third is serving them and actually deploying them. Okay, so Ray and Anyscale are used for all of these three different workloads. Companies like OpenAI or Cohere that train large language models. Or open source versions like GPTJ are done on top of Ray. There are many startups and other businesses that fine tune, that, you know, don't want to train the large underlying foundation models, but that do want to fine tune them, do want to adapt them to their purposes, and build products around them and serve them, those are also using Ray and Anyscale for that fine tuning and that serving. And so the reason that Ray and Anyscale are important here is that, you know, building and using foundation models requires a huge scale. It requires a lot of data. It requires a lot of compute, GPUs, TPUs, other resources. And to actually take advantage of that and actually build these scalable applications, there's a lot of infrastructure that needs to happen under the hood. And so you can either use Ray and Anyscale to take care of that and manage the infrastructure and solve those infrastructure problems. Or you can build the infrastructure and manage the infrastructure yourself, which you can do, but it's going to slow your team down. It's going to, you know, many of the businesses we work with simply don't want to be in the business of managing infrastructure and building infrastructure. They want to focus on product development and move faster. >> I know you got a keynote presentation we're going to go to in a second, but I think you hit on something I think is the real tipping point, doing it yourself, hard to do. These are things where opportunities are and the Cloud did that with data centers. Turned a data center and made it an API. The heavy lifting went away and went to the Cloud so people could be more creative and build their product. In this case, build their creativity. Is that kind of what's the big deal? Is that kind of a big deal happening that you guys are taking the learnings and making that available so people don't have to do that? >> That's exactly right. So today, if you want to succeed with AI, if you want to use AI in your business, infrastructure work is on the critical path for doing that. To do AI, you have to build infrastructure. You have to figure out how to scale your applications. That's going to change. We're going to get to the point, and you know, with Ray and Anyscale, we're going to remove the infrastructure from the critical path so that as a developer or as a business, all you need to focus on is your application logic, what you want the the program to do, what you want your application to do, how you want the AI to actually interface with the rest of your product. Now the way that will happen is that Ray and Anyscale will still, the infrastructure work will still happen. It'll just be under the hood and taken care of by Ray in Anyscale. And so I think something like this is really necessary for AI to reach its potential, for AI to have the impact and the reach that we think it will, you have to make it easier to do. >> And just for clarification to point out, if you don't mind explaining the relationship of Ray and Anyscale real quick just before we get into the presentation. >> So Ray is an open source project. We created it. We were at Berkeley doing machine learning. We started Ray so that, in order to provide an easy, a simple open source tool for building and running scalable applications. And Anyscale is the managed version of Ray, basically we will run Ray for you in the Cloud, provide a lot of tools around the developer experience and managing the infrastructure and providing more performance and superior infrastructure. >> Awesome. I know you got a presentation on Ray and Anyscale and you guys are positioning as the infrastructure for foundational models. So I'll let you take it away and then when you're done presenting, we'll come back, I'll probably grill you with a few questions and then we'll close it out so take it away. >> Robert: Sounds great. So I'll say a little bit about how companies are using Ray and Anyscale for foundation models. The first thing I want to mention is just why we're doing this in the first place. And the underlying observation, the underlying trend here, and this is a plot from OpenAI, is that the amount of compute needed to do machine learning has been exploding. It's been growing at something like 35 times every 18 months. This is absolutely enormous. And other people have written papers measuring this trend and you get different numbers. But the point is, no matter how you slice and dice it, it' a astronomical rate. Now if you compare that to something we're all familiar with, like Moore's Law, which says that, you know, the processor performance doubles every roughly 18 months, you can see that there's just a tremendous gap between the needs, the compute needs of machine learning applications, and what you can do with a single chip, right. So even if Moore's Law were continuing strong and you know, doing what it used to be doing, even if that were the case, there would still be a tremendous gap between what you can do with the chip and what you need in order to do machine learning. And so given this graph, what we've seen, and what has been clear to us since we started this company, is that doing AI requires scaling. There's no way around it. It's not a nice to have, it's really a requirement. And so that led us to start Ray, which is the open source project that we started to make it easy to build these scalable Python applications and scalable machine learning applications. And since we started the project, it's been adopted by a tremendous number of companies. Companies like OpenAI, which use Ray to train their large models like ChatGPT, companies like Uber, which run all of their deep learning and classical machine learning on top of Ray, companies like Shopify or Spotify or Instacart or Lyft or Netflix, ByteDance, which use Ray for their machine learning infrastructure. Companies like Ant Group, which makes Alipay, you know, they use Ray across the board for fraud detection, for online learning, for detecting money laundering, you know, for graph processing, stream processing. Companies like Amazon, you know, run Ray at a tremendous scale and just petabytes of data every single day. And so the project has seen just enormous adoption since, over the past few years. And one of the most exciting use cases is really providing the infrastructure for building training, fine tuning, and serving foundation models. So I'll say a little bit about, you know, here are some examples of companies using Ray for foundation models. Cohere trains large language models. OpenAI also trains large language models. You can think about the workloads required there are things like supervised pre-training, also reinforcement learning from human feedback. So this is not only the regular supervised learning, but actually more complex reinforcement learning workloads that take human input about what response to a particular question, you know is better than a certain other response. And incorporating that into the learning. There's open source versions as well, like GPTJ also built on top of Ray as well as projects like Alpa coming out of UC Berkeley. So these are some of the examples of exciting projects in organizations, training and creating these large language models and serving them using Ray. Okay, so what actually is Ray? Well, there are two layers to Ray. At the lowest level, there's the core Ray system. This is essentially low level primitives for building scalable Python applications. Things like taking a Python function or a Python class and executing them in the cluster setting. So Ray core is extremely flexible and you can build arbitrary scalable applications on top of Ray. So on top of Ray, on top of the core system, what really gives Ray a lot of its power is this ecosystem of scalable libraries. So on top of the core system you have libraries, scalable libraries for ingesting and pre-processing data, for training your models, for fine tuning those models, for hyper parameter tuning, for doing batch processing and batch inference, for doing model serving and deployment, right. And a lot of the Ray users, the reason they like Ray is that they want to run multiple workloads. They want to train and serve their models, right. They want to load their data and feed that into training. And Ray provides common infrastructure for all of these different workloads. So this is a little overview of what Ray, the different components of Ray. So why do people choose to go with Ray? I think there are three main reasons. The first is the unified nature. The fact that it is common infrastructure for scaling arbitrary workloads, from data ingest to pre-processing to training to inference and serving, right. This also includes the fact that it's future proof. AI is incredibly fast moving. And so many people, many companies that have built their own machine learning infrastructure and standardized on particular workflows for doing machine learning have found that their workflows are too rigid to enable new capabilities. If they want to do reinforcement learning, if they want to use graph neural networks, they don't have a way of doing that with their standard tooling. And so Ray, being future proof and being flexible and general gives them that ability. Another reason people choose Ray in Anyscale is the scalability. This is really our bread and butter. This is the reason, the whole point of Ray, you know, making it easy to go from your laptop to running on thousands of GPUs, making it easy to scale your development workloads and run them in production, making it easy to scale, you know, training to scale data ingest, pre-processing and so on. So scalability and performance, you know, are critical for doing machine learning and that is something that Ray provides out of the box. And lastly, Ray is an open ecosystem. You can run it anywhere. You can run it on any Cloud provider. Google, you know, Google Cloud, AWS, Asure. You can run it on your Kubernetes cluster. You can run it on your laptop. It's extremely portable. And not only that, it's framework agnostic. You can use Ray to scale arbitrary Python workloads. You can use it to scale and it integrates with libraries like TensorFlow or PyTorch or JAX or XG Boost or Hugging Face or PyTorch Lightning, right, or Scikit-learn or just your own arbitrary Python code. It's open source. And in addition to integrating with the rest of the machine learning ecosystem and these machine learning frameworks, you can use Ray along with all of the other tooling in the machine learning ecosystem. That's things like weights and biases or ML flow, right. Or you know, different data platforms like Databricks, you know, Delta Lake or Snowflake or tools for model monitoring for feature stores, all of these integrate with Ray. And that's, you know, Ray provides that kind of flexibility so that you can integrate it into the rest of your workflow. And then Anyscale is the scalable compute platform that's built on top, you know, that provides Ray. So Anyscale is a managed Ray service that runs in the Cloud. And what Anyscale does is it offers the best way to run Ray. And if you think about what you get with Anyscale, there are fundamentally two things. One is about moving faster, accelerating the time to market. And you get that by having the managed service so that as a developer you don't have to worry about managing infrastructure, you don't have to worry about configuring infrastructure. You also, it provides, you know, optimized developer workflows. Things like easily moving from development to production, things like having the observability tooling, the debug ability to actually easily diagnose what's going wrong in a distributed application. So things like the dashboards and the other other kinds of tooling for collaboration, for monitoring and so on. And then on top of that, so that's the first bucket, developer productivity, moving faster, faster experimentation and iteration. The second reason that people choose Anyscale is superior infrastructure. So this is things like, you know, cost deficiency, being able to easily take advantage of spot instances, being able to get higher GPU utilization, things like faster cluster startup times and auto scaling. Things like just overall better performance and faster scheduling. And so these are the kinds of things that Anyscale provides on top of Ray. It's the managed infrastructure. It's fast, it's like the developer productivity and velocity as well as performance. So this is what I wanted to share about Ray in Anyscale. >> John: Awesome. >> Provide that context. But John, I'm curious what you think. >> I love it. I love the, so first of all, it's a platform because that's the platform architecture right there. So just to clarify, this is an Anyscale platform, not- >> That's right. >> Tools. So you got tools in the platform. Okay, that's key. Love that managed service. Just curious, you mentioned Python multiple times, is that because of PyTorch and TensorFlow or Python's the most friendly with machine learning or it's because it's very common amongst all developers? >> That's a great question. Python is the language that people are using to do machine learning. So it's the natural starting point. Now, of course, Ray is actually designed in a language agnostic way and there are companies out there that use Ray to build scalable Java applications. But for the most part right now we're focused on Python and being the best way to build these scalable Python and machine learning applications. But, of course, down the road there always is that potential. >> So if you're slinging Python code out there and you're watching that, you're watching this video, get on Anyscale bus quickly. Also, I just, while you were giving the presentation, I couldn't help, since you mentioned OpenAI, which by the way, congratulations 'cause they've had great scale, I've noticed in their rapid growth 'cause they were the fastest company to the number of users than anyone in the history of the computer industry, so major successor, OpenAI and ChatGPT, huge fan. I'm not a skeptic at all. I think it's just the beginning, so congratulations. But I actually typed into ChatGPT, what are the top three benefits of Anyscale and came up with scalability, flexibility, and ease of use. Obviously, scalability is what you guys are called. >> That's pretty good. >> So that's what they came up with. So they nailed it. Did you have an inside prompt training, buy it there? Only kidding. (Robert laughs) >> Yeah, we hard coded that one. >> But that's the kind of thing that came up really, really quickly if I asked it to write a sales document, it probably will, but this is the future interface. This is why people are getting excited about the foundational models and the large language models because it's allowing the interface with the user, the consumer, to be more human, more natural. And this is clearly will be in every application in the future. >> Absolutely. This is how people are going to interface with software, how they're going to interface with products in the future. It's not just something, you know, not just a chat bot that you talk to. This is going to be how you get things done, right. How you use your web browser or how you use, you know, how you use Photoshop or how you use other products. Like you're not going to spend hours learning all the APIs and how to use them. You're going to talk to it and tell it what you want it to do. And of course, you know, if it doesn't understand it, it's going to ask clarifying questions. You're going to have a conversation and then it'll figure it out. >> This is going to be one of those things, we're going to look back at this time Robert and saying, "Yeah, from that company, that was the beginning of that wave." And just like AWS and Cloud Computing, the folks who got in early really were in position when say the pandemic came. So getting in early is a good thing and that's what everyone's talking about is getting in early and playing around, maybe replatforming or even picking one or few apps to refactor with some staff and managed services. So people are definitely jumping in. So I have to ask you the ROI cost question. You mentioned some of those, Moore's Law versus what's going on in the industry. When you look at that kind of scale, the first thing that jumps out at people is, "Okay, I love it. Let's go play around." But what's it going to cost me? Am I going to be tied to certain GPUs? What's the landscape look like from an operational standpoint, from the customer? Are they locked in and the benefit was flexibility, are you flexible to handle any Cloud? What is the customers, what are they looking at? Basically, that's my question. What's the customer looking at? >> Cost is super important here and many of the companies, I mean, companies are spending a huge amount on their Cloud computing, on AWS, and on doing AI, right. And I think a lot of the advantage of Anyscale, what we can provide here is not only better performance, but cost efficiency. Because if we can run something faster and more efficiently, it can also use less resources and you can lower your Cloud spending, right. We've seen companies go from, you know, 20% GPU utilization with their current setup and the current tools they're using to running on Anyscale and getting more like 95, you know, 100% GPU utilization. That's something like a five x improvement right there. So depending on the kind of application you're running, you know, it's a significant cost savings. We've seen companies that have, you know, processing petabytes of data every single day with Ray going from, you know, getting order of magnitude cost savings by switching from what they were previously doing to running their application on Ray. And when you have applications that are spending, you know, potentially $100 million a year and getting a 10 X cost savings is just absolutely enormous. So these are some of the kinds of- >> Data infrastructure is super important. Again, if the customer, if you're a prospect to this and thinking about going in here, just like the Cloud, you got infrastructure, you got the platform, you got SaaS, same kind of thing's going to go on in AI. So I want to get into that, you know, ROI discussion and some of the impact with your customers that are leveraging the platform. But first I hear you got a demo. >> Robert: Yeah, so let me show you, let me give you a quick run through here. So what I have open here is the Anyscale UI. I've started a little Anyscale Workspace. So Workspaces are the Anyscale concept for interactive developments, right. So here, imagine I'm just, you want to have a familiar experience like you're developing on your laptop. And here I have a terminal. It's not on my laptop. It's actually in the cloud running on Anyscale. And I'm just going to kick this off. This is going to train a large language model, so OPT. And it's doing this on 32 GPUs. We've got a cluster here with a bunch of CPU cores, bunch of memory. And as that's running, and by the way, if I wanted to run this on instead of 32 GPUs, 64, 128, this is just a one line change when I launch the Workspace. And what I can do is I can pull up VS code, right. Remember this is the interactive development experience. I can look at the actual code. Here it's using Ray train to train the torch model. We've got the training loop and we're saying that each worker gets access to one GPU and four CPU cores. And, of course, as I make the model larger, this is using deep speed, as I make the model larger, I could increase the number of GPUs that each worker gets access to, right. And how that is distributed across the cluster. And if I wanted to run on CPUs instead of GPUs or a different, you know, accelerator type, again, this is just a one line change. And here we're using Ray train to train the models, just taking my vanilla PyTorch model using Hugging Face and then scaling that across a bunch of GPUs. And, of course, if I want to look at the dashboard, I can go to the Ray dashboard. There are a bunch of different visualizations I can look at. I can look at the GPU utilization. I can look at, you know, the CPU utilization here where I think we're currently loading the model and running that actual application to start the training. And some of the things that are really convenient here about Anyscale, both I can get that interactive development experience with VS code. You know, I can look at the dashboards. I can monitor what's going on. It feels, I have a terminal, it feels like my laptop, but it's actually running on a large cluster. And I can, with however many GPUs or other resources that I want. And so it's really trying to combine the best of having the familiar experience of programming on your laptop, but with the benefits, you know, being able to take advantage of all the resources in the Cloud to scale. And it's like when, you know, you're talking about cost efficiency. One of the biggest reasons that people waste money, one of the silly reasons for wasting money is just forgetting to turn off your GPUs. And what you can do here is, of course, things will auto terminate if they're idle. But imagine you go to sleep, I have this big cluster. You can turn it off, shut off the cluster, come back tomorrow, restart the Workspace, and you know, your big cluster is back up and all of your code changes are still there. All of your local file edits. It's like you just closed your laptop and came back and opened it up again. And so this is the kind of experience we want to provide for our users. So that's what I wanted to share with you. >> Well, I think that whole, couple of things, lines of code change, single line of code change, that's game changing. And then the cost thing, I mean human error is a big deal. People pass out at their computer. They've been coding all night or they just forget about it. I mean, and then it's just like leaving the lights on or your water running in your house. It's just, at the scale that it is, the numbers will add up. That's a huge deal. So I think, you know, compute back in the old days, there's no compute. Okay, it's just compute sitting there idle. But you know, data cranking the models is doing, that's a big point. >> Another thing I want to add there about cost efficiency is that we make it really easy to use, if you're running on Anyscale, to use spot instances and these preemptable instances that can just be significantly cheaper than the on-demand instances. And so when we see our customers go from what they're doing before to using Anyscale and they go from not using these spot instances 'cause they don't have the infrastructure around it, the fault tolerance to handle the preemption and things like that, to being able to just check a box and use spot instances and save a bunch of money. >> You know, this was my whole, my feature article at Reinvent last year when I met with Adam Selipsky, this next gen Cloud is here. I mean, it's not auto scale, it's infrastructure scale. It's agility. It's flexibility. I think this is where the world needs to go. Almost what DevOps did for Cloud and what you were showing me that demo had this whole SRE vibe. And remember Google had site reliability engines to manage all those servers. This is kind of like an SRE vibe for data at scale. I mean, a similar kind of order of magnitude. I mean, I might be a little bit off base there, but how would you explain it? >> It's a nice analogy. I mean, what we are trying to do here is get to the point where developers don't think about infrastructure. Where developers only think about their application logic. And where businesses can do AI, can succeed with AI, and build these scalable applications, but they don't have to build, you know, an infrastructure team. They don't have to develop that expertise. They don't have to invest years in building their internal machine learning infrastructure. They can just focus on the Python code, on their application logic, and run the stuff out of the box. >> Awesome. Well, I appreciate the time. Before we wrap up here, give a plug for the company. I know you got a couple websites. Again, go, Ray's got its own website. You got Anyscale. You got an event coming up. Give a plug for the company looking to hire. Put a plug in for the company. >> Yeah, absolutely. Thank you. So first of all, you know, we think AI is really going to transform every industry and the opportunity is there, right. We can be the infrastructure that enables all of that to happen, that makes it easy for companies to succeed with AI, and get value out of AI. Now we have, if you're interested in learning more about Ray, Ray has been emerging as the standard way to build scalable applications. Our adoption has been exploding. I mentioned companies like OpenAI using Ray to train their models. But really across the board companies like Netflix and Cruise and Instacart and Lyft and Uber, you know, just among tech companies. It's across every industry. You know, gaming companies, agriculture, you know, farming, robotics, drug discovery, you know, FinTech, we see it across the board. And all of these companies can get value out of AI, can really use AI to improve their businesses. So if you're interested in learning more about Ray and Anyscale, we have our Ray Summit coming up in September. This is going to highlight a lot of the most impressive use cases and stories across the industry. And if your business, if you want to use LLMs, you want to train these LLMs, these large language models, you want to fine tune them with your data, you want to deploy them, serve them, and build applications and products around them, give us a call, talk to us. You know, we can really take the infrastructure piece, you know, off the critical path and make that easy for you. So that's what I would say. And, you know, like you mentioned, we're hiring across the board, you know, engineering, product, go-to-market, and it's an exciting time. >> Robert Nishihara, co-founder and CEO of Anyscale, congratulations on a great company you've built and continuing to iterate on and you got growth ahead of you, you got a tailwind. I mean, the AI wave is here. I think OpenAI and ChatGPT, a customer of yours, have really opened up the mainstream visibility into this new generation of applications, user interface, roll of data, large scale, how to make that programmable so we're going to need that infrastructure. So thanks for coming on this season three, episode one of the ongoing series of the hot startups. In this case, this episode is the top startups building foundational model infrastructure for AI and ML. I'm John Furrier, your host. Thanks for watching. (upbeat music)
SUMMARY :
episode one of the ongoing and you guys really had and other resources in the Cloud. and particular the large language and what you want to achieve. and the Cloud did that with data centers. the point, and you know, if you don't mind explaining and managing the infrastructure and you guys are positioning is that the amount of compute needed to do But John, I'm curious what you think. because that's the platform So you got tools in the platform. and being the best way to of the computer industry, Did you have an inside prompt and the large language models and tell it what you want it to do. So I have to ask you and you can lower your So I want to get into that, you know, and you know, your big cluster is back up So I think, you know, the on-demand instances. and what you were showing me that demo and run the stuff out of the box. I know you got a couple websites. and the opportunity is there, right. and you got growth ahead
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Robert Nishihara | PERSON | 0.99+ |
John | PERSON | 0.99+ |
Robert | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
Netflix | ORGANIZATION | 0.99+ |
35 times | QUANTITY | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
$100 million | QUANTITY | 0.99+ |
Uber | ORGANIZATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
100% | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
Ant Group | ORGANIZATION | 0.99+ |
first | QUANTITY | 0.99+ |
Python | TITLE | 0.99+ |
20% | QUANTITY | 0.99+ |
32 GPUs | QUANTITY | 0.99+ |
Lyft | ORGANIZATION | 0.99+ |
hundreds | QUANTITY | 0.99+ |
tomorrow | DATE | 0.99+ |
Anyscale | ORGANIZATION | 0.99+ |
three | QUANTITY | 0.99+ |
128 | QUANTITY | 0.99+ |
September | DATE | 0.99+ |
today | DATE | 0.99+ |
Moore's Law | TITLE | 0.99+ |
Adam Selipsky | PERSON | 0.99+ |
PyTorch | TITLE | 0.99+ |
Ray | ORGANIZATION | 0.99+ |
second reason | QUANTITY | 0.99+ |
64 | QUANTITY | 0.99+ |
each worker | QUANTITY | 0.99+ |
each worker | QUANTITY | 0.99+ |
Photoshop | TITLE | 0.99+ |
UC Berkeley | ORGANIZATION | 0.99+ |
Java | TITLE | 0.99+ |
Shopify | ORGANIZATION | 0.99+ |
OpenAI | ORGANIZATION | 0.99+ |
Anyscale | PERSON | 0.99+ |
third | QUANTITY | 0.99+ |
two things | QUANTITY | 0.99+ |
ByteDance | ORGANIZATION | 0.99+ |
Spotify | ORGANIZATION | 0.99+ |
One | QUANTITY | 0.99+ |
95 | QUANTITY | 0.99+ |
Asure | ORGANIZATION | 0.98+ |
one line | QUANTITY | 0.98+ |
one GPU | QUANTITY | 0.98+ |
ChatGPT | TITLE | 0.98+ |
TensorFlow | TITLE | 0.98+ |
last year | DATE | 0.98+ |
first bucket | QUANTITY | 0.98+ |
both | QUANTITY | 0.98+ |
two layers | QUANTITY | 0.98+ |
Cohere | ORGANIZATION | 0.98+ |
Alipay | ORGANIZATION | 0.98+ |
Ray | PERSON | 0.97+ |
one | QUANTITY | 0.97+ |
Instacart | ORGANIZATION | 0.97+ |
Sri Satish Ambati, H20.ai | CUBE Conversation, May 2020
>> connecting with thought leaders all around the world, this is a CUBE Conversation. Hi, everybody this is Dave Vellante of theCUBE, and welcome back to my CXO series. I've been running this through really since the start of the COVID-19 crisis to really understand how leaders are dealing with this pandemic. Sri Ambati is here, he's the CEO and founder of H20. Sri, it's great to see you again, thanks for coming on. >> Thank you for having us. >> Yeah, so this pandemic has obviously given people fits, no question, but it's also given opportunities for companies to kind of reassess where they are. Automation is a huge watchword, flexibility, business resiliency and people who maybe really hadn't fully leaned into things like the cloud and AI and automation are now realizing, wow, we have no choice, it's about survival. Your thought as to what you're seeing in the marketplace. >> Thanks for having us. I think first of all, kudos to the frontline health workers who have been ruthlessly saving lives across the country and the world, and what you're really doing is a fraction of what we could have done or should be doing to stay away the next big pandemic. But that apart I think, I usually tend to say BC is before COVID. So if the world was thinking about going digital after COVID-19, they have been forced to go digital and as a result, you're seeing tremendous transformation across our customers, and a lot of application to kind of go in and reinvent their business models that allow them to scale as effortlessly as they could using the digital means. >> So, think about, doctors and diagnosis machines, in some cases, are helping doctors make diagnoses, they're sometimes making even better diagnosis, (mumbles) is informing. There's been a lot of talk about the models, you know how... Yeah, I know you've been working with a lot of healthcare organizations, you may probably familiar with that, you know, the Medium post, The Hammer and the Dance, and if people criticize the models, of course, they're just models, right? And you iterate models and machine intelligence can help us improve. So, in this, you know, you talk about BC and post C, how have you seen the data and in machine intelligence informing the models and proving that what we know about this pandemic, I mean, it changed literally daily, what are you seeing? >> Yeah, and I think it started with Wuhan and we saw the best application of AI in trying to trace, literally from Alipay, to WeChat, track down the first folks who were spreading it across China and then eventually the rest of the world. I think contact tracing, for example, has become a really interesting problem. supply chain has been disrupted like never before. We're beginning to see customers trying to reinvent their distribution mechanisms in the second order effects of the COVID, and the the prime center is hospital staffing, how many ventilator, is the first few weeks so that after COVID crisis as it evolved in the US. We are busy predicting working with some of the local healthcare communities to predict how staffing in hospitals will work, how many PPE and ventilators will be needed and so henceforth, but that quickly and when the peak surge will be those with the beginning problems, and many of our customers have begin to do these models and iterate and improve and kind of educate the community to practice social distancing, and that led to a lot of flattening the curve and you're talking flattening the curve, you're really talking about data science and analytics in public speak. That led to kind of the next level, now that we have somewhat brought a semblance of order to the reaction to COVID, I think what we are beginning to figure out is, is there going to be a second surge, what elective procedures that were postponed, will be top of the mind for customers, and so this is the kind of things that hospitals are beginning to plan out for the second half of the year, and as businesses try to open up, certain things were highly correlated to surgeon cases, such as cleaning supplies, for example, the obvious one or pantry buying. So retailers are beginning to see what online stores are doing well, e-commerce, online purchases, electronic goods, and so everyone essentially started working from home, and so homes needed to have the same kind of bandwidth that offices and commercial enterprises needed to have, and so a lot of interesting, as one side you saw airlines go away, this side you saw the likes of Zoom and video take off. So you're kind of seeing a real divide in the digital divide and that's happening and AI is here to play a very good role to figure out how to enhance your profitability as you're looking about planning out the next two years. >> Yeah, you know, and obviously, these things they get, they get partisan, it gets political, I mean, our job as an industry is to report, your job is to help people understand, I mean, let the data inform and then let public policy you know, fight it out. So who are some of the people that you're working with that you know, as a result of COVID-19. What's some of the work that H2O has done, I want to better understand what role are you playing? >> So one of the things we're kind of privileged as a company to come into the crisis, with a strong balance and an ability to actually have the right kind of momentum behind the company in terms of great talent, and so we have 10% of the world's top data scientists in the in the form of Kaggle Grand Masters in the company. And so we put most of them to work, and they started collecting data sets, curating data sets and making them more qualitative, picking up public data sources, for example, there's a tremendous amount of job loss out there, figuring out which are the more difficult kind of sectors in the economy and then we started looking at exodus from the cities, we're looking at mobility data that's publicly available, mobility data through the data exchanges, you're able to find which cities which rural areas, did the New Yorkers as they left the city, which places did they go to, and what's to say, Californians when they left Los Angeles, which are the new places they have settled in? These are the places which are now busy places for the same kind of items that you need to sell if you're a retailer, but if you go one step further, we started engaging with FEMA, we start engaging with the universities, like Imperial College London or Berkeley, and started figuring out how best to improve the models and automate them. The SEER model, the most popular SEER model, we added that into our Driverless AI product as a recipe and made that accessible to our customers in testing, to customers in healthcare who are trying to predict where the surge is likely to come. But it's mostly about information right? So the AI at the end of it is all about intelligence and being prepared. Predictive is all about being prepared and that's kind of what we did with general, lots of blogs, typical blog articles and working with the largest health organizations and starting to kind of inform them on the most stable models. What we found to our not so much surprise, is that the simplest, very interpretable models are actually the most widely usable, because historical data is actually no longer as effective. You need to build a model that you can quickly understand and retry again to the feedback loop of back testing that model against what really happened. >> Yeah, so I want to double down on that. So really, two things I want to understand, if you have visibility on it, sounds like you do. Just in terms of the surge and the comeback, you know, kind of what those models say, based upon, you know, we have some advanced information coming from the global market, for sure, but it seems like every situation is different. What's the data telling you? Just in terms of, okay, we're coming into the spring and the summer months, maybe it'll come down a little bit. Everybody says it... We fully expect it to come back in the fall, go back to college, don't go back to college. What is the data telling you at this point in time with an understanding that, you know, we're still iterating every day? >> Well, I think I mean, we're not epidemiologists, but at the same time, the science of it is a highly local response, very hyper local response to COVID-19 is what we've seen. Santa Clara, which is just a county, I mean, is different from San Francisco, right, sort of. So you beginning to see, like we saw in Brooklyn, it's very different, and Bronx, very different from Manhattan. So you're seeing a very, very local response to this disease, and I'm talking about US. You see the likes of Brazil, which we're worried about, has picked up quite a bit of cases now. I think the silver lining I would say is that China is up and running to a large degree, a large number of our user base there are back active, you can see the traffic patterns there. So two months after their last research cases, the business and economic activity is back and thriving. And so, you can kind of estimate from that, that this can be done where you can actually contain the rise of active cases and it will take masking of the entire community, masking and the healthy dose of increase in testing. One of our offices is in Prague, and Czech Republic has done an incredible job in trying to contain this and they've done essentially, masked everybody and as a result they're back thinking about opening offices, schools later this month. So I think that's a very, very local response, hyper local response, no one country and no one community is symmetrical with other ones and I think we have a unique situation where in United States you have a very, very highly connected world, highly connected economy and I think we have quite a problem on our hands on how to safeguard our economy while also safeguarding life. >> Yeah, so you can't just, you can't just take Norway and apply it or South Korea and apply it, every situation is different. And then I want to ask you about, you know, the economy in terms of, you know, how much can AI actually, you know, how can it work in this situation where you have, you know, for example, okay, so the Fed, yes, it started doing asset buys back in 2008 but still, very hard to predict, I mean, at this time of this interview you know, Stock Market up 900 points, very difficult to predict that but some event happens in the morning, somebody, you know, Powell says something positive and it goes crazy but just sort of even modeling out the V recovery, the W recovery, deep recession, the comeback. You have to have enough data, do you not? In order for AI to be reasonably accurate? How does it work? And how does at what pace can you iterate and improve on the models? >> So I think that's exactly where I would say, continuous modeling, instead of continuously learning continuous, that's where the vision of the world is headed towards, where data is coming, you build a model, and then you iterate, try it out and come back. That kind of rapid, continuous learning would probably be needed for all our models as opposed to the typical, I'm pushing a model to production once a year, or once every quarter. I think what we're beginning to see is the kind of where companies are beginning to kind of plan out. A lot of people lost their jobs in the last couple of months, right, sort of. And so up scaling and trying to kind of bring back these jobs back both into kind of, both from the manufacturing side, but also lost a lot of jobs in the transportation and the kind of the airlines slash hotel industries, right, sort of. So it's trying to now bring back the sense of confidence and will take a lot more kind of testing, a lot more masking, a lot more social empathy, I think well, some of the things that we are missing while we are socially distant, we know that we are so connected as a species, we need to kind of start having that empathy for we need to wear a mask, not for ourselves, but for our neighbors and people we may run into. And I think that kind of, the same kind of thinking has to kind of parade, before we can open up the economy in a big way. The data, I mean, we can do a lot of transfer learning, right, sort of there are new methods, like try to model it, similar to the 1918, where we had a second bump, or a lot of little bumps, and that's kind of where your W shaped pieces, but governments are trying very well in seeing stimulus dollars being pumped through banks. So some of the US case we're looking for banks is, which small medium business in especially, in unsecured lending, which business to lend to, (mumbles) there's so many applications that have come to banks across the world, it's not just in the US, and banks are caught up with the problem of which and what's growing the concern for this business to kind of, are they really accurate about the number of employees they are saying they have? Do then the next level problem or on forbearance and mortgage, that side of the things are coming up at some of these banks as well. So they're looking at which, what's one of the problems that one of our customers Wells Fargo, they have a question which branch to open, right, sort of that itself, it needs a different kind of modeling. So everything has become a very highly good segmented models, and so AI is absolutely not just a good to have, it has become a must have for most of our customers in how to go about their business. (mumbles) >> I want to talk a little bit about your business, you have been on a mission to democratize AI since the beginning, open source. Explain your business model, how you guys make money and then I want to help people understand basic theoretical comparisons and current affairs. >> Yeah, that's great. I think the last time we spoke, probably about at the Spark Summit. I think Dave and we were talking about Sparkling Water and H2O our open source platforms, which are premium platforms for democratizing machine learning and math at scale, and that's been a tremendous brand for us. Over the last couple of years, we have essentially built a platform called Driverless AI, which is a license software and that automates machine learning models, we took the best practices of all these data scientists, and combined them to essentially build recipes that allow people to build the best forecasting models, best fraud prevention models or the best recommendation engines, and so we started augmenting traditional data scientists with this automatic machine learning called AutoML, that essentially allows them to build models without necessarily having the same level of talent as these great Kaggle Grand Masters. And so that has democratized, allowed ordinary companies to start producing models of high caliber and high quality that would otherwise have been the pedigree of Google, Microsoft or Amazon or some of these top tier AI houses like Netflix and others. So what we've done is democratize not just the algorithms at the open source level. Now, we've made it easy for kind of rapid adoption of AI across every branch inside a company, a large organization, also across smaller organizations which don't have the access to the same kind of talent. Now, third level, you know, what we've brought to market, is ability to augment data sets, especially public and private data sets that you can, the alternative data sets that can increase the signal. And that's where we've started working on a new platform called Q, again, more license software, and I mean, to give you an idea there from business models endpoint, now majority of our software sales is coming from closed source software. And sort of so, we've made that transition, we still make our open source widely accessible, we continue to improve it, a large chunk of the teams are improving and participating in building the communities but I think from a business model standpoint as of last year, 51% of our revenues are now coming from closed source software and that change is continuing to grow. >> And this is the point I wanted to get to, so you know, the open source model was you know, Red Hat the one company that, you know, succeeded wildly and it was, put it out there open source, come up with a service, maintain the software, you got to buy the subscription okay, fine. And everybody thought that you know, you were going to do that, they thought that Databricks was going to do and that changed. But I want to take two examples, Hortonworks which kind of took the Red Hat model and Cloudera which does IP. And neither really lived up to the expectation, but now there seems to be sort of a new breed I mentioned, you guys, Databricks, there are others, that seem to be working. You with your license software model, Databricks with a managed service and so there's, it's becoming clear that there's got to be some level of IP that can be licensed in order to really thrive in the open source community to be able to fund the committers that you have to put forth to open source. I wonder if you could give me your thoughts on that narrative. >> So on Driverless AI, which is the closest platform I mentioned, we opened up the layers in open source as recipes. So for example, different companies build their zip codes differently, right, the domain specific recipes, we put about 150 of them in open source again, on top of our Driverless AI platform, and the idea there is that, open source is about freedom, right? It is not necessarily about, it's not a philosophy, it's not a business model, it allows freedom for rapid adoption of a platform and complete democratization and commodification of a space. And that allows a small company like ours to compete at the level of an SaaS or a Google or a Microsoft because you have the same level of voice as a very large company and you're focused on using code as a community building exercise as opposed to a business model, right? So that's kind of the heart of open source, is allowing that freedom for our end users and the customers to kind of innovate at the same level of that a Silicon Valley company or one of these large tech giants are building software. So it's really about making, it's a maker culture, as opposed to a consumer culture around software. Now, if you look at how the the Red Hat model, and the others who have tried to replicate that, the difficult part there was, if the product is very good, customers are self sufficient and if it becomes a standard, then customers know how to use it. If the product is crippled or difficult to use, then you put a lot of services and that's where you saw the classic Hadoop companies, get pulled into a lot of services, which is a reasonably difficult business to scale. So I think what we chose was, instead, a great product that builds a fantastic brand, that makes AI, even when other first or second.ai domain, and for us to see thousands of companies which are not AI and AI first, and even more companies adopting AI and talking about AI as a major way that was possible because of open source. If you had chosen close source and many of your peers did, they all vanished. So that's kind of how the open source is really about building the ecosystem and having the patience to build a company that takes 10, 20 years to build. And what we are expecting unfortunately, is a first and fast rise up to become unicorns. In that race, you're essentially sacrifice, building a long ecosystem play, and that's kind of what we chose to do, and that took a little longer. Now, if you think about the, how do you truly monetize open source, it takes a little longer and is much more difficult sales machine to scale, right, sort of. Our open source business actually is reasonably positive EBITDA business because it makes more money than we spend on it. But trying to teach sales teams, how to sell open source, that's a much, that's a rate limiting step. And that's why we chose and also explaining to the investors, how open source is being invested in as you go closer to the IPO markets, that's where we chose, let's go into license software model and scale that as a regular business. >> So I've said a few times, it's kind of like ironic that, this pandemic is as we're entering a new decade, you know, we've kind of we're exiting the era, I mean, the many, many decades of Moore's law being the source of innovation and now it's a combination of data, applying machine intelligence and being able to scale and with cloud. Well, my question is, what did we expect out of AI this decade if those are sort of the three, the cocktail of innovation, if you will, what should we expect? Is it really just about, I suggest, is it really about automating, you know, businesses, giving them more agility, flexibility, you know, etc. Or should we should we expect more from AI this decade? >> Well, I mean, if you think about the decade of 2010 2011, that was defined by software is eating the world, right? And now you can say software is the world, right? I mean, pretty much almost all conditions are digital. And AI is eating software, right? (mumbling) A lot of cloud transitions are happening and are now happening much faster rate but cloud and AI are kind of the leading, AI is essentially one of the biggest driver for cloud adoption for many of our customers. So in the enterprise world, you're seeing rebuilding of a lot of data, fast data driven applications that use AI, instead of rule based software, you're beginning to see patterned, mission AI based software, and you're seeing that in spades. And, of course, that is just the tip of the iceberg, AI has been with us for 100 years, and it's going to be ahead of us another hundred years, right, sort of. So as you see the discovery rate at which, it is really a fundamentally a math, math movement and in that math movement at the beginning of every century, it leads to 100 years of phenomenal discovery. So AI is essentially making discoveries faster, AI is producing, entertainment, AI is producing music, AI is producing choreographing, you're seeing AI in every walk of life, AI summarization of Zoom meetings, right, you beginning to see a lot of the AI enabled ETF peaking of stocks, right, sort of. You're beginning to see, we repriced 20,000 bonds every 15 seconds using H2O AI, corporate bonds. And so you and one of our customers is on the fastest growing stock, mostly AI is powering a lot of these insights in a fast changing world which is globally connected. No one of us is able to combine all the multiple dimensions that are changing and AI has that incredible opportunity to be a partner for every... (mumbling) For a hospital looking at how the second half will look like for physicians looking at what is the sentiment of... What is the surge to expect? To kind of what is the market demand looking at the sentiment of the customers. AI is the ultimate money ball in business and then I think it's just showing its depth at this point. >> Yeah, I mean, I think you're right on, I mean, basically AI is going to convert every software, every application, or those tools aren't going to have much use, Sri we got to go but thanks so much for coming to theCUBE and the great work you guys are doing. Really appreciate your insights. stay safe, and best of luck to you guys. >> Likewise, thank you so much. >> Welcome, and thank you for watching everybody, this is Dave Vellante for the CXO series on theCUBE. We'll see you next time. All right, we're clear. All right.
SUMMARY :
Sri, it's great to see you Your thought as to what you're and a lot of application and if people criticize the models, and kind of educate the community and then let public policy you know, and starting to kind of inform them What is the data telling you of the entire community, and improve on the models? and the kind of the airlines and then I want to help people understand and I mean, to give you an idea there in the open source community to be able and the customers to kind of innovate and being able to scale and with cloud. What is the surge to expect? and the great work you guys are doing. Welcome, and thank you
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave | PERSON | 0.99+ |
2008 | DATE | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Wells Fargo | ORGANIZATION | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.99+ | |
San Francisco | LOCATION | 0.99+ |
Prague | LOCATION | 0.99+ |
Brooklyn | LOCATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
51% | QUANTITY | 0.99+ |
May 2020 | DATE | 0.99+ |
China | LOCATION | 0.99+ |
United States | LOCATION | 0.99+ |
100 years | QUANTITY | 0.99+ |
Bronx | LOCATION | 0.99+ |
Databricks | ORGANIZATION | 0.99+ |
Manhattan | LOCATION | 0.99+ |
US | LOCATION | 0.99+ |
Santa Clara | LOCATION | 0.99+ |
last year | DATE | 0.99+ |
10% | QUANTITY | 0.99+ |
20,000 bonds | QUANTITY | 0.99+ |
Imperial College London | ORGANIZATION | 0.99+ |
Hortonworks | ORGANIZATION | 0.99+ |
One | QUANTITY | 0.99+ |
COVID-19 | OTHER | 0.99+ |
Los Angeles | LOCATION | 0.99+ |
Netflix | ORGANIZATION | 0.99+ |
H20 | ORGANIZATION | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
South Korea | LOCATION | 0.99+ |
Sri Satish Ambati | PERSON | 0.99+ |
thousands | QUANTITY | 0.99+ |
FEMA | ORGANIZATION | 0.99+ |
Brazil | LOCATION | 0.99+ |
second half | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
second surge | QUANTITY | 0.99+ |
two months | QUANTITY | 0.99+ |
one | QUANTITY | 0.98+ |
second bump | QUANTITY | 0.98+ |
two things | QUANTITY | 0.98+ |
H2O | ORGANIZATION | 0.98+ |
both | QUANTITY | 0.98+ |
Czech Republic | LOCATION | 0.98+ |
Silicon Valley | LOCATION | 0.98+ |
TITLE | 0.98+ | |
three | QUANTITY | 0.98+ |
hundred years | QUANTITY | 0.98+ |
once a year | QUANTITY | 0.97+ |
Powell | PERSON | 0.97+ |
Sparkling Water | ORGANIZATION | 0.97+ |
Alipay | TITLE | 0.97+ |
Norway | LOCATION | 0.97+ |
pandemic | EVENT | 0.97+ |
second order | QUANTITY | 0.97+ |
third level | QUANTITY | 0.97+ |
first folks | QUANTITY | 0.97+ |
COVID-19 crisis | EVENT | 0.96+ |
Fed | ORGANIZATION | 0.95+ |
1918 | DATE | 0.95+ |
later this month | DATE | 0.95+ |
one side | QUANTITY | 0.94+ |
Sri Ambati | PERSON | 0.94+ |
two examples | QUANTITY | 0.93+ |
Moore | PERSON | 0.92+ |
Californians | PERSON | 0.92+ |
CXO | TITLE | 0.92+ |
last couple of months | DATE | 0.92+ |
COVID | OTHER | 0.91+ |
Spark Summit | EVENT | 0.91+ |
one step | QUANTITY | 0.91+ |
The Hammer | TITLE | 0.9+ |
COVID crisis | EVENT | 0.87+ |
every 15 seconds | QUANTITY | 0.86+ |
Sri Satish Ambati, H20.ai | CUBE Conversation, May 2020
>> Starting the record, Dave in five, four, three. Hi, everybody this is Dave Vellante, theCUBE, and welcome back to my CXO series. I've been running this through really since the start of the COVID-19 crisis to really understand how leaders are dealing with this pandemic. Sri Ambati is here, he's the CEO and founder of H20. Sri, it's great to see you again, thanks for coming on. >> Thank you for having us. >> Yeah, so this pandemic has obviously given people fits, no question, but it's also given opportunities for companies to kind of reassess where they are. Automation is a huge watchword, flexibility, business resiliency and people who maybe really hadn't fully leaned into things like the cloud and AI and automation are now realizing, wow, we have no choice, it's about survival. Your thought as to what you're seeing in the marketplace. >> Thanks for having us. I think first of all, kudos to the frontline health workers who have been ruthlessly saving lives across the country and the world, and what you're really doing is a fraction of what we could have done or should be doing to stay away the next big pandemic. But that apart I think, I usually tend to say BC is before COVID. So if the world was thinking about going digital after COVID-19, they have been forced to go digital and as a result, you're seeing tremendous transformation across our customers, and a lot of application to kind of go in and reinvent their business models that allow them to scale as effortlessly as they could using the digital means. >> So, think about, doctors and diagnosis machines, in some cases, are helping doctors make diagnoses, they're sometimes making even better diagnosis, (mumbles) is informing. There's been a lot of talk about the models, you know how... Yeah, I know you've been working with a lot of healthcare organizations, you may probably familiar with that, you know, the Medium post, The Hammer and the Dance, and if people criticize the models, of course, they're just models, right? And you iterate models and machine intelligence can help us improve. So, in this, you know, you talk about BC and post C, how have you seen the data and in machine intelligence informing the models and proving that what we know about this pandemic, I mean, it changed literally daily, what are you seeing? >> Yeah, and I think it started with Wuhan and we saw the best application of AI in trying to trace, literally from Alipay, to WeChat, track down the first folks who were spreading it across China and then eventually the rest of the world. I think contact tracing, for example, has become a really interesting problem. supply chain has been disrupted like never before. We're beginning to see customers trying to reinvent their distribution mechanisms in the second order effects of the COVID, and the the prime center is hospital staffing, how many ventilator, is the first few weeks so that after COVID crisis as it evolved in the US. We are busy predicting working with some of the local healthcare communities to predict how staffing in hospitals will work, how many PPE and ventilators will be needed and so henceforth, but that quickly and when the peak surge will be those with the beginning problems, and many of our customers have begin to do these models and iterate and improve and kind of educate the community to practice social distancing, and that led to a lot of flattening the curve and you're talking flattening the curve, you're really talking about data science and analytics in public speak. That led to kind of the next level, now that we have somewhat brought a semblance of order to the reaction to COVID, I think what we are beginning to figure out is, is there going to be a second surge, what elective procedures that were postponed, will be top of the mind for customers, and so this is the kind of things that hospitals are beginning to plan out for the second half of the year, and as businesses try to open up, certain things were highly correlated to surgeon cases, such as cleaning supplies, for example, the obvious one or pantry buying. So retailers are beginning to see what online stores are doing well, e-commerce, online purchases, electronic goods, and so everyone essentially started working from home, and so homes needed to have the same kind of bandwidth that offices and commercial enterprises needed to have, and so a lot of interesting, as one side you saw airlines go away, this side you saw the likes of Zoom and video take off. So you're kind of seeing a real divide in the digital divide and that's happening and AI is here to play a very good role to figure out how to enhance your profitability as you're looking about planning out the next two years. >> Yeah, you know, and obviously, these things they get, they get partisan, it gets political, I mean, our job as an industry is to report, your job is to help people understand, I mean, let the data inform and then let public policy you know, fight it out. So who are some of the people that you're working with that you know, as a result of COVID-19. What's some of the work that H2O has done, I want to better understand what role are you playing? >> So one of the things we're kind of privileged as a company to come into the crisis, with a strong balance and an ability to actually have the right kind of momentum behind the company in terms of great talent, and so we have 10% of the world's top data scientists in the in the form of Kaggle Grand Masters in the company. And so we put most of them to work, and they started collecting data sets, curating data sets and making them more qualitative, picking up public data sources, for example, there's a tremendous amount of job loss out there, figuring out which are the more difficult kind of sectors in the economy and then we started looking at exodus from the cities, we're looking at mobility data that's publicly available, mobility data through the data exchanges, you're able to find which cities which rural areas, did the New Yorkers as they left the city, which places did they go to, and what's to say, Californians when they left Los Angeles, which are the new places they have settled in? These are the places which are now busy places for the same kind of items that you need to sell if you're a retailer, but if you go one step further, we started engaging with FEMA, we start engaging with the universities, like Imperial College London or Berkeley, and started figuring out how best to improve the models and automate them. The SaaS model, the most popular SaaS model, we added that into our Driverless AI product as a recipe and made that accessible to our customers in testing, to customers in healthcare who are trying to predict where the surge is likely to come. But it's mostly about information right? So the AI at the end of it is all about intelligence and being prepared. Predictive is all about being prepared and that's kind of what we did with general, lots of blogs, typical blog articles and working with the largest health organizations and starting to kind of inform them on the most stable models. What we found to our not so much surprise, is that the simplest, very interpretable models are actually the most widely usable, because historical data is actually no longer as effective. You need to build a model that you can quickly understand and retry again to the feedback loop of back testing that model against what really happened. >> Yeah, so I want to double down on that. So really, two things I want to understand, if you have visibility on it, sounds like you do. Just in terms of the surge and the comeback, you know, kind of what those models say, based upon, you know, we have some advanced information coming from the global market, for sure, but it seems like every situation is different. What's the data telling you? Just in terms of, okay, we're coming into the spring and the summer months, maybe it'll come down a little bit. Everybody says it... We fully expect it to come back in the fall, go back to college, don't go back to college. What is the data telling you at this point in time with an understanding that, you know, we're still iterating every day? >> Well, I think I mean, we're not epidemiologists, but at the same time, the science of it is a highly local response, very hyper local response to COVID-19 is what we've seen. Santa Clara, which is just a county, I mean, is different from San Francisco, right, sort of. So you beginning to see, like we saw in Brooklyn, it's very different, and Bronx, very different from Manhattan. So you're seeing a very, very local response to this disease, and I'm talking about US. You see the likes of Brazil, which we're worried about, has picked up quite a bit of cases now. I think the silver lining I would say is that China is up and running to a large degree, a large number of our user base there are back active, you can see the traffic patterns there. So two months after their last research cases, the business and economic activity is back and thriving. And so, you can kind of estimate from that, that this can be done where you can actually contain the rise of active cases and it will take masking of the entire community, masking and the healthy dose of increase in testing. One of our offices is in Prague, and Czech Republic has done an incredible job in trying to contain this and they've done essentially, masked everybody and as a result they're back thinking about opening offices, schools later this month. So I think that's a very, very local response, hyper local response, no one country and no one community is symmetrical with other ones and I think we have a unique situation where in United States you have a very, very highly connected world, highly connected economy and I think we have quite a problem on our hands on how to safeguard our economy while also safeguarding life. >> Yeah, so you can't just, you can't just take Norway and apply it or South Korea and apply it, every situation is different. And then I want to ask you about, you know, the economy in terms of, you know, how much can AI actually, you know, how can it work in this situation where you have, you know, for example, okay, so the Fed, yes, it started doing asset buys back in 2008 but still, very hard to predict, I mean, at this time of this interview you know, Stock Market up 900 points, very difficult to predict that but some event happens in the morning, somebody, you know, Powell says something positive and it goes crazy but just sort of even modeling out the V recovery, the W recovery, deep recession, the comeback. You have to have enough data, do you not? In order for AI to be reasonably accurate? How does it work? And how does at what pace can you iterate and improve on the models? >> So I think that's exactly where I would say, continuous modeling, instead of continuously learning continuous, that's where the vision of the world is headed towards, where data is coming, you build a model, and then you iterate, try it out and come back. That kind of rapid, continuous learning would probably be needed for all our models as opposed to the typical, I'm pushing a model to production once a year, or once every quarter. I think what we're beginning to see is the kind of where companies are beginning to kind of plan out. A lot of people lost their jobs in the last couple of months, right, sort of. And so up scaling and trying to kind of bring back these jobs back both into kind of, both from the manufacturing side, but also lost a lot of jobs in the transportation and the kind of the airlines slash hotel industries, right, sort of. So it's trying to now bring back the sense of confidence and will take a lot more kind of testing, a lot more masking, a lot more social empathy, I think well, some of the things that we are missing while we are socially distant, we know that we are so connected as a species, we need to kind of start having that empathy for we need to wear a mask, not for ourselves, but for our neighbors and people we may run into. And I think that kind of, the same kind of thinking has to kind of parade, before we can open up the economy in a big way. The data, I mean, we can do a lot of transfer learning, right, sort of there are new methods, like try to model it, similar to the 1918, where we had a second bump, or a lot of little bumps, and that's kind of where your W shaped pieces, but governments are trying very well in seeing stimulus dollars being pumped through banks. So some of the US case we're looking for banks is, which small medium business in especially, in unsecured lending, which business to lend to, (mumbles) there's so many applications that have come to banks across the world, it's not just in the US, and banks are caught up with the problem of which and what's growing the concern for this business to kind of, are they really accurate about the number of employees they are saying they have? Do then the next level problem or on forbearance and mortgage, that side of the things are coming up at some of these banks as well. So they're looking at which, what's one of the problems that one of our customers Wells Fargo, they have a question which branch to open, right, sort of that itself, it needs a different kind of modeling. So everything has become a very highly good segmented models, and so AI is absolutely not just a good to have, it has become a must have for most of our customers in how to go about their business. (mumbles) >> I want to talk a little bit about your business, you have been on a mission to democratize AI since the beginning, open source. Explain your business model, how you guys make money and then I want to help people understand basic theoretical comparisons and current affairs. >> Yeah, that's great. I think the last time we spoke, probably about at the Spark Summit. I think Dave and we were talking about Sparkling Water and H2O or open source platforms, which are premium platforms for democratizing machine learning and math at scale, and that's been a tremendous brand for us. Over the last couple of years, we have essentially built a platform called Driverless AI, which is a license software and that automates machine learning models, we took the best practices of all these data scientists, and combined them to essentially build recipes that allow people to build the best forecasting models, best fraud prevention models or the best recommendation engines, and so we started augmenting traditional data scientists with this automatic machine learning called AutoML, that essentially allows them to build models without necessarily having the same level of talent as these Greek Kaggle Grand Masters. And so that has democratized, allowed ordinary companies to start producing models of high caliber and high quality that would otherwise have been the pedigree of Google, Microsoft or Amazon or some of these top tier AI houses like Netflix and others. So what we've done is democratize not just the algorithms at the open source level. Now, we've made it easy for kind of rapid adoption of AI across every branch inside a company, a large organization, also across smaller organizations which don't have the access to the same kind of talent. Now, third level, you know, what we've brought to market, is ability to augment data sets, especially public and private data sets that you can, the alternative data sets that can increase the signal. And that's where we've started working on a new platform called Q, again, more license software, and I mean, to give you an idea there from business models endpoint, now majority of our software sales is coming from closed source software. And sort of so, we've made that transition, we still make our open source widely accessible, we continue to improve it, a large chunk of the teams are improving and participating in building the communities but I think from a business model standpoint as of last year, 51% of our revenues are now coming from closed source software and that change is continuing to grow. >> And this is the point I wanted to get to, so you know, the open source model was you know, Red Hat the one company that, you know, succeeded wildly and it was, put it out there open source, come up with a service, maintain the software, you got to buy the subscription okay, fine. And everybody thought that you know, you were going to do that, they thought that Databricks was going to do and that changed. But I want to take two examples, Hortonworks which kind of took the Red Hat model and Cloudera which does IP. And neither really lived up to the expectation, but now there seems to be sort of a new breed I mentioned, you guys, Databricks, there are others, that seem to be working. You with your license software model, Databricks with a managed service and so there's, it's becoming clear that there's got to be some level of IP that can be licensed in order to really thrive in the open source community to be able to fund the committers that you have to put forth to open source. I wonder if you could give me your thoughts on that narrative. >> So on Driverless AI, which is the closest platform I mentioned, we opened up the layers in open source as recipes. So for example, different companies build their zip codes differently, right, the domain specific recipes, we put about 150 of them in open source again, on top of our Driverless AI platform, and the idea there is that, open source is about freedom, right? It is not necessarily about, it's not a philosophy, it's not a business model, it allows freedom for rapid adoption of a platform and complete democratization and commodification of a space. And that allows a small company like ours to compete at the level of an SaaS or a Google or a Microsoft because you have the same level of voice as a very large company and you're focused on using code as a community building exercise as opposed to a business model, right? So that's kind of the heart of open source, is allowing that freedom for our end users and the customers to kind of innovate at the same level of that a Silicon Valley company or one of these large tech giants are building software. So it's really about making, it's a maker culture, as opposed to a consumer culture around software. Now, if you look at how the the Red Hat model, and the others who have tried to replicate that, the difficult part there was, if the product is very good, customers are self sufficient and if it becomes a standard, then customers know how to use it. If the product is crippled or difficult to use, then you put a lot of services and that's where you saw the classic Hadoop companies, get pulled into a lot of services, which is a reasonably difficult business to scale. So I think what we chose was, instead, a great product that builds a fantastic brand, that makes AI, even when other first or second.ai domain, and for us to see thousands of companies which are not AI and AI first, and even more companies adopting AI and talking about AI as a major way that was possible because of open source. If you had chosen close source and many of your peers did, they all vanished. So that's kind of how the open source is really about building the ecosystem and having the patience to build a company that takes 10, 20 years to build. And what we are expecting unfortunately, is a first and fast rise up to become unicorns. In that race, you're essentially sacrifice, building a long ecosystem play, and that's kind of what we chose to do, and that took a little longer. Now, if you think about the, how do you truly monetize open source, it takes a little longer and is much more difficult sales machine to scale, right, sort of. Our open source business actually is reasonably positive EBITDA business because it makes more money than we spend on it. But trying to teach sales teams, how to sell open source, that's a much, that's a rate limiting step. And that's why we chose and also explaining to the investors, how open source is being invested in as you go closer to the IPO markets, that's where we chose, let's go into license software model and scale that as a regular business. >> So I've said a few times, it's kind of like ironic that, this pandemic is as we're entering a new decade, you know, we've kind of we're exiting the era, I mean, the many, many decades of Moore's law being the source of innovation and now it's a combination of data, applying machine intelligence and being able to scale and with cloud. Well, my question is, what did we expect out of AI this decade if those are sort of the three, the cocktail of innovation, if you will, what should we expect? Is it really just about, I suggest, is it really about automating, you know, businesses, giving them more agility, flexibility, you know, etc. Or should we should we expect more from AI this decade? >> Well, I mean, if you think about the decade of 2010 2011, that was defined by software is eating the world, right? And now you can say software is the world, right? I mean, pretty much almost all conditions are digital. And AI is eating software, right? (mumbling) A lot of cloud transitions are happening and are now happening much faster rate but cloud and AI are kind of the leading, AI is essentially one of the biggest driver for cloud adoption for many of our customers. So in the enterprise world, you're seeing rebuilding of a lot of data, fast data driven applications that use AI, instead of rule based software, you're beginning to see patterned, mission AI based software, and you're seeing that in spades. And, of course, that is just the tip of the iceberg, AI has been with us for 100 years, and it's going to be ahead of us another hundred years, right, sort of. So as you see the discovery rate at which, it is really a fundamentally a math, math movement and in that math movement at the beginning of every century, it leads to 100 years of phenomenal discovery. So AI is essentially making discoveries faster, AI is producing, entertainment, AI is producing music, AI is producing choreographing, you're seeing AI in every walk of life, AI summarization of Zoom meetings, right, you beginning to see a lot of the AI enabled ETF peaking of stocks, right, sort of. You're beginning to see, we repriced 20,000 bonds every 15 seconds using H2O AI, corporate bonds. And so you and one of our customers is on the fastest growing stock, mostly AI is powering a lot of these insights in a fast changing world which is globally connected. No one of us is able to combine all the multiple dimensions that are changing and AI has that incredible opportunity to be a partner for every... (mumbling) For a hospital looking at how the second half will look like for physicians looking at what is the sentiment of... What is the surge to expect? To kind of what is the market demand looking at the sentiment of the customers. AI is the ultimate money ball in business and then I think it's just showing its depth at this point. >> Yeah, I mean, I think you're right on, I mean, basically AI is going to convert every software, every application, or those tools aren't going to have much use, Sri we got to go but thanks so much for coming to theCUBE and the great work you guys are doing. Really appreciate your insights. stay safe, and best of luck to you guys. >> Likewise, thank you so much. >> Welcome, and thank you for watching everybody, this is Dave Vellante for the CXO series on theCUBE. We'll see you next time. All right, we're clear. All right.
SUMMARY :
Sri, it's great to see you Your thought as to what you're and a lot of application and if people criticize the models, and kind of educate the community and then let public policy you know, is that the simplest, What is the data telling you of the entire community, and improve on the models? and the kind of the airlines and then I want to help people understand and I mean, to give you an idea there in the open source community to be able and the customers to kind of innovate and being able to scale and with cloud. What is the surge to expect? and the great work you guys are doing. Welcome, and thank you
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Wells Fargo | ORGANIZATION | 0.99+ |
Dave | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
2008 | DATE | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
five | QUANTITY | 0.99+ |
San Francisco | LOCATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.99+ | |
Brooklyn | LOCATION | 0.99+ |
Prague | LOCATION | 0.99+ |
China | LOCATION | 0.99+ |
Bronx | LOCATION | 0.99+ |
100 years | QUANTITY | 0.99+ |
May 2020 | DATE | 0.99+ |
Manhattan | LOCATION | 0.99+ |
51% | QUANTITY | 0.99+ |
US | LOCATION | 0.99+ |
Brazil | LOCATION | 0.99+ |
Databricks | ORGANIZATION | 0.99+ |
United States | LOCATION | 0.99+ |
COVID-19 | OTHER | 0.99+ |
10% | QUANTITY | 0.99+ |
20,000 bonds | QUANTITY | 0.99+ |
Los Angeles | LOCATION | 0.99+ |
last year | DATE | 0.99+ |
H20 | ORGANIZATION | 0.99+ |
Imperial College London | ORGANIZATION | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
one | QUANTITY | 0.99+ |
four | QUANTITY | 0.99+ |
Santa Clara | LOCATION | 0.99+ |
One | QUANTITY | 0.99+ |
hundred years | QUANTITY | 0.99+ |
Hortonworks | ORGANIZATION | 0.99+ |
Netflix | ORGANIZATION | 0.99+ |
Sri Satish Ambati | PERSON | 0.99+ |
South Korea | LOCATION | 0.99+ |
three | QUANTITY | 0.99+ |
second half | QUANTITY | 0.99+ |
two things | QUANTITY | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
both | QUANTITY | 0.98+ |
second surge | QUANTITY | 0.98+ |
first | QUANTITY | 0.98+ |
H2O | ORGANIZATION | 0.98+ |
third level | QUANTITY | 0.98+ |
once a year | QUANTITY | 0.98+ |
Sparkling Water | ORGANIZATION | 0.98+ |
FEMA | ORGANIZATION | 0.98+ |
TITLE | 0.98+ | |
pandemic | EVENT | 0.98+ |
Powell | PERSON | 0.97+ |
COVID-19 crisis | EVENT | 0.97+ |
second bump | QUANTITY | 0.97+ |
Czech Republic | LOCATION | 0.96+ |
second order | QUANTITY | 0.96+ |
1918 | DATE | 0.96+ |
Norway | LOCATION | 0.96+ |
Fed | ORGANIZATION | 0.95+ |
first folks | QUANTITY | 0.94+ |
thousands of companies | QUANTITY | 0.94+ |
two examples | QUANTITY | 0.91+ |
10, 20 years | QUANTITY | 0.91+ |
COVID | OTHER | 0.91+ |
CXO | TITLE | 0.91+ |
two months | QUANTITY | 0.91+ |
last couple of months | DATE | 0.9+ |
Moore | PERSON | 0.9+ |
later this month | DATE | 0.9+ |
Alipay | TITLE | 0.89+ |
Sri Ambati | PERSON | 0.88+ |
every 15 seconds | QUANTITY | 0.88+ |
COVID crisis | EVENT | 0.86+ |
Californians | PERSON | 0.85+ |
Driverless | TITLE | 0.84+ |
Marshall Taplits, NYNJA Group | Blockchain Unbound 2018
>> Narrator: Live from San Juan, Puerto Rico It's theCUBE. Covering Blockchain Unbound. Brought to you by Blockchain Industries. (latin music) >> Hello and welcome back to theCUBE exclusive coverage in Puerto Rico for Blockchain Unbound I'm John Furrier, your host, here covering all the action in Puerto Rico as the global society and industry come together. Our next guest is Marhall Taplits he's the Chief Strategy Officer and Co Founder of Nynja.biz, check out their site, Nynja.biz. Marshall, thanks for joining me. >> Thank you. >> So tell about what you guys do. You guys are doing some disruptive stuff, tell us about what you guys do, then it will jam into a conversation. >> Sure, so are you familiar with WeChat in China, for example? >> Yeah. >> Okay great. So I've personally been living in China 15 years, so we've watched kind of the birth of the Chinese internet, which as we know, is a little different than the regular internet. >> A lot of mobile users. >> A lot of mobile users, 800 million China mobile subscribers alone. WeChat, basically, is a platform that started off as just a messenger but basically what it's done is it's integrated into every facet of Chinese society. To give you an example, you go to a restaurant, you scan the QR code, the menu comes up, you pick the food, you pay for the food, it comes, you walk out. Everything like that is in China. Everything like that is in Wuzhen China. So what we've done is we've kind of taken this concept, and we're working on a global version of it, that's cryptocurrency based, and we are working specifically with Chinese companies in order to help them go global as part of the China One Belt One Road program and working with companies like Alibaba, what have you, in order to help Chinese companies go overseas and take what they've built in China but operate globally with cryptocurrency. >> Are you guys in China? Cause it's been hard for companies to start companies in China. So you're living in China or you're working in China? >> Yeah so because we live in Shenzhen, right next to it is Hong Kong. Hong Kong is where our company is based. Hong Kong, as you know, previous British colony, the legal system, and the financial system-- >> And you domicile in Hong Kong, that's where you're based? >> Me personally in Shenzhen, but the company is in Hong Kong. So we also have a Wyoming corporation in the US. >> That's where all the action is. >> That's right >> That's where WeChat is >> That's right >> Alibaba's got Alipay and then there's more business to business with their app. So I get that WeChat's been highly successful. In fact we have a huge following on WeChat, Sou Kanai, Niki Bond, free content. But that brings up the question of Chinese kind of showing the way with mobile expansion, so their users are heavily mobile savy. >> Marshall: That's right. >> This is pretty obvious when you think about it, but in America and around the world, that's going to translate to the new user experience. So in your opinion, how would you describe the expectations that users have? Because you're living on the front end of the wave of what mobile's doing, I mean there's a lot of gamification going on, some if it's kind of creepy, but what is your view of the expectations that users have and what's different about what's currently available in the webstac, and the 20 year old e-commerce stacks, that are out there? >> Sure, I think the most important thing is reducing friction, all right. You don't want to be using platforms where you can not do it wherever you are whenever you are, you don't want to have to go through payment processes, you don't want to have to re-authenticate yourself across whatever platforms you use. And interestingly, when I first went to China, it was all about copying what was in the west over to there, but actually it's kind of the opposite now, right, so we basically want to take this concept of the frictionless digital life, and make it a global opportunity. And especially with BlockChain and cryptocurrency you have that really as an opportunity, because if you look at all the apps that are out there, and the platforms that are out there, the only ones that have gone past a billion users, WhatsApp, Instagram, whatever are the free ones. But as soon as you layer in payment, it becomes very locked. And as big as WeChat is, and as big as LINE is, but ultimately it's locked into the Rem and B system or Reo in Korea, what have you, so the cryptocurrency is really the first opportunity that the world's had to create platforms that can get up to a billion, two billion, three billion users that are able to pay. And we just think that's a once in a lifetime opportunity and we want to be part of it. >> So I got to ask you about the impact that cloud computing has had on this, obviously we've seen cloud computing destroy the data center model, allow people to get time to value faster, mobile on top, big data analytics using data, all this stuff's awesome stuff. So the question is, is that, that's kind of a horizontally disruptive view, so these stacks that are built old way where I got to own the stack end to end, yeah there's some standardization on the lower end of the stack. But now you're thinking about more of a horizontal, I got jurisdictions, I got regions, I got countries, I got sovereignty, all these things are in the melting pot of the cryptocurrency BlockChain, de-centralized applications, are major impacts to all those things. How do you see that playing out because, that's kind of what developers worry about, oh shit will this work on that chain? I got Neo I got this I got that, so the plumbing is totally a moving train right now. >> Marshall: That's right. >> But the business models are pretty obvious. So there's like a business ops thing going on. What Dev opts did for Cloud, you got this new abstraction thing going on with this world. What's your view on that, do you agree? Or what's your take? >> Yeah well you pretty much nailed it. I mean basically what's happening is over the last 10 or 15 years people have finally accepted that having your own server is kind of silly, you know, and most people now will just spin up whatever they need in terms of resources on TheCloud. But over the last couple years, you're really going more toward Edge Cloud, where the way the clouds work, is that basically it's pushing to get the least amount of latency and store the data as close to the user as possible. And then there's also regulatory in some countries now in terms of, if your users are from this country, you have to legally store the data in this area. So this is all kind of evolving. And if you look at the BlockChain technology, I think it's the payment version of that. So for example, everyone's always concerned about getting in and out of Fiat Currency, and how am I going to get back to dollars, and this and that, but I think what's going to wind up happening, is this is going to get pushed towards the edges and there will be opportunities and ways with exchanges and what have you to get in and out. But more importantly, it's going to be like, just other currencies, so for example, I live in China but I come to the US a few times a year, I also travel to Europe, I have some dollars, I have some Euros, I have some Rem and B, when I leave China, I don't immediately sell all of my Rem and B, I just keep it because at some point I'm going to need it. And I think what's going to happen in the cryptocurrency space is, especially on the larger BlockChains, like Ethereum and Neo and what have you, is people are just going to get used to keeping some of it and they're going to stop worrying about what the exact exchange rate is and how am I going to get in and out, and this and that, and they're just going to start treating it as part of their currency stack that they keep. >> Yeah as long as there's some level of stability. It's just like, I remember when I was growing up, there was no Euro, every country had their own currency. You had the French Franc, the Swiss Francs, the Deutsche Mark, Lira, etc, etc. But you're seeing that the viability of the money aspect, cause at the end of the day there's two things that we've identified in analysis, and I was talking about it last night, talked about it this morning on theCUBE, is the killer apps for BlockChain cryptocurrency, these sorts of apps is two things, money and marketplaces. >> Marshall: That's right. >> Everything else is just kind of circling around those two. >> Well there's more but certainly that's the main part of it >> Money, moving around. So the UK just announced with coin based, the Financial Conduct Authority, reading the news yesterday, has essentially said we're going to allow for the fast payment system to convert to Fiat. This is a government, the UK is a nation. This is the beginning, to your point, that if they don't get up to speed, the edge of the network will democratize them and kind of circle the wagons, if you will, so it's already happening. >> Yeah and I think what governments are starting to realize is hey guys this is just a technology and not only do you don't really have jurisdiction to control it, but also that you don't even have the technical means. So Wyoming is a good example of regulation coming into play, that just kind of accepts the presence that this now exists, right. And they're not going to try to make it something and fit it into the old way. So, and in terms of the stability of these coins, I think it is important because people want stability, but in other ways, if you don't look at the exchange rate, it's actually way more stable than the current system, and I'll give an example. In the last month or two, the prices of cryptocurrency have dropped almost 40%. Now if the stock markets and the global affects markets drop 40%, you'd have blood in the streets. But the crypto market is asset based instead of debt based and because it's so structurally sound it's able to handle these wild swings without actually collapsing the system, so in may ways, it's way more stable, and then as the market gaps and the buy in of these currencies get bigger and bigger, of course it's going to be more stable over time. >> Well I mean its stable from a fail standpoint, but a lot of emotional instability. People losing money for the first time. >> But that's just because they're-- >> That's a lot of speculation, right? >> There's a lot of speculating and then if they're down they feel like they lost but, that's life. >> People that are into the game, like you, were long on this. So what would you explain to someone, cause I have two, a lot of friends that have two schools of thought, that's a total scam, don't associate with that, to oh my god, that's the next biggest wave, lets get our surfboards out there and lets get on this, there's a multiple set coming in, it's the biggest thing we've seen, and everything in between. How do you explain it to people for the first time? >> It's just your traditional curve, there's early adopters and what have you, and if you were one of the guys buying up domaine names in the early 90s, you know some people would say I can't believe you're spending $100,000 buying up domaine names, but some of them now are worth, you know, tens of millions of dollars. But again, this is the speculatory piece of it. And there's no shortage of opportunities for speculation and I encourage everybody to speculate a little bit because what it does is it gets you a taste of the technology. And usually, when you have some money on the line, you pay more attention, so if speculation is what gets people interested, and it gets them watching it and understanding the technology and using it, then I'm all for it, but people shouldn't be speculating with money they don't have. Anything could happen in the short term. Nobody knows what's going to happen with any specific currency. But in terms of the technology itself, this is a revolution way bigger than the internet itself. This is where you're getting, not only, communications like the internet, but financing governance and all as one. Programmable money, programmable contracts, that wipes out finance, it wipes out legal, it whites out governance in many ways. So this is a huge evolution in human society, and we've termed this Open Unity actually. And so we believe that society has to reach a state of open unity in order to go into the singularity as we would envision it wanting to be, as something that's under our control. >> Yeah and I think one of the things, first of all that's a great statement, well said. I'll just kind of put some reality on that, connect the dots, is that if you look at the trajectory of cloud computing, Amazon Web Services was laughed at years ago. S3 came out, compute storage building, basic building blocks and a slew more services. What Cloud did for software developers, and what they've disrupted from a business standpoint, dev ops, it's proven. What open source has done, even going back to the old red-hat days and linux, is that now a tier one global citizen in software, you look at those two trends, you can connect that dots to what you just said. And what made Cloud great was they made application developers have access to programmable infrastructure. >> Marshall: Exactly. >> You're talking about a whole nother level of software programmability, money, marketplace, society, >> Yeah you hit it on the head. >> We're there right? >> That's exactly right, so when a programmer wants to start a business, instead of going to create an LLC, and getting their EIN Tax ID or whatever, and when they want to go into Europe, and dealing with that and then trying to open a bank account, which is almost impossible, internationally now, instead of that, you just have your SDKs and your APIs or whatever and you've got access to money, program adding, you can take money, you can move money around, globally, frictionless, permissionless, with governancy, smart contracts-- >> They might not not need an SDK dashboard, its a console, click, click, click, smart contracts, governance, turn key. >> And one of the things we're working on with Nynja in particular, is this kind of on-demand marketplace and putting together a de-centralized teams for work. And this is all driven by smart contracts. So one of the issues with the economy is the huge booms and busts that people have in the economy. And if you look at the root cause of that, my personal opinion, is that it's because of payment terms. So for example, if I do work for you, and then there's an invoice, but it's not due for 30 days, now your business may be structurally sound, but the truth is your cashflow is all over the place. With BlockChain technology, we can actually do real time payments. You could be paid minute by minute, hour by hour. Real time, program, contract. So we're going to create very flat even money flows through the entire economy globally, and we're going to just completely remove these booms and busts that are really nothing more than just cashflow issues that are compounded and compounded at a global level. >> I mean I lived through the dot com bubble, I was actually part of it on the front end, on the euphoria side, as well as on the crash. Part of the whole search paradigm, google right there. Key words, all that stuff happening, growth, massive growth. So I saw that, the scammers in there, or the bubble people, that's what we called them. But the reality is, everything happened. It was pet foods online, you could get shopping delivered to your house. So again, to your point, it's a little euphoric right now, but what's different is, is you have now, community data. See what I see happening is, it's not a major bubble crash, because self government, self governing, self governance, is a community dynamic. So I think there's going to be a lot of self healing, inside the networks themselves. You're already seeing it here, a lot of people, bad act is being identified, investors flight to quality, looking at quality deals. Interesting times, your thoughts? >> Well I mean you know, we've been through many evolutions of society, we've had surf-dom, we've had monarchies, we've had representative democracies, we have all these things, and I just think the next evolution is decentralized governance. And we don't even know what that means yet, because it's just starting, but I think we can all, if we can close our eyes and really think about it. I think it's pretty obvious what the issues are with our current system and not just the US, but globally, and I think we have an opportunity here to build in organic program governance. And what's really special about BoxChain technology is if I program it to do X, it's going to do X. So we don't need to, I don't need to know who you are to trust you. I don't need to worry about where we're going to sue each other, or we're going to have arbitration if things go wrong. We're just going to make an agreement, and we're going to program it that way, and that's it. And now the next phase is, I could build on top of that trusting that that's just going to happen. So you can create these chains of trust, and that can happen anywhere in the world. So I think this is a whole nother-- >> Sounds like a bunch of web services. >> Well in many ways, in terms of the architecture, sure you could absolutely think of it like that. >> The reusability, the leverage is amazing. All right, so I want to just end the segment Marshall, take a minute to end the segment, to talk about what you're working on, Nynja coin, Nynja, N-Y-N-J-A .biz, you guys have a product, you got a BlockChain enabled platform, you got a coin, take a minute to explain what you're working on. >> Basically we want to provide the tools and services to help people live in this new reality. So in order to basically function in the world that we're entering into, we're going to need tools that far surpass what's currently available in terms of the messengers, the web sites, all these things. We need to be operating at a level that takes communication completely frictionless, payment completely frictionless, and governance completely frictionless. And we have to put this all together, and that's what we're doing with Nynja. We're staring with a global communicator, which is basically, if you want to take WeChat, telegram, whatever, but we have about 50 additional features that really take communications to the next level. And then on top of it, creating the baseline with cryptocurrency payment, and also smart contract wizards and helping people kind of get these teams going and get paid and organize their financial life in a de-centralized way. So we're just basically going to be the next generation of these messenger type platforms with BlockChain integrated. And what you're going to see is that over the next couple years you're going to get to the first companies that are achieving not just a billion or two billion or three billion users, but paying users, and we're going to be one of the probably three to five platforms that are offering tools at the global level like this. >> And have you got an IC already or not? >> We've just started our private ICO about two weeks ago. We're getting tremendous support in Asia. Quite frankly, the US is not seeing it as much-- >> Is it a utility token or security? >> Utility Token, and I think it's really telling, interesting, coming here. It's the first time I've been doing the presenting. We spoke yesterday at the d10e and we also spoke at d10e in Korea a week or two ago, and the response is incredible. And I think the reason is because-- >> The Asian market gets it. >> Well they're already living in this world within their own confines in terms of the messenger with their payment and governance built in, so when I tell them that we're going to do this globally with crypto, immediately they get it. I'm having trouble here, especially in these five minute pitches which is ridiculous, it's like a chop shop, I don't know how to communicate the idea within this short time frame, so, what I'm looking for while we're here this week is just to find people who really want to take an hour or two or even people like yourself who want to do interviews and just kind of really talk to people and really explain-- >> Well platform is complex, a lot of pieces to it. It's a system, but the value you offer is essentially offering developers, who are building products, for tools that you've built so they can scale faster. That sounds like your value. >> That's right and although I can't say specifically, we're also working on a deal that's going to get us started with about 15 million active users on day one, so that's very exciting and we're really really excited about that. >> And the coins will be utility of measures, what? >> Sorry? >> Well your utility coins going to be measuring what, what's the main token economics that drives the-- >> For the ICO economics? >> Your Nynja Coin. >> So basically we're releasing 5 billion tokens, 45% of them will be sold. There's five cents a token, so the hard cap, by definition is about 112 million, actually we're planning to do the public sale in April, but we may cancel it or postpone it just because the private sale is going really well, but we'll see how that goes. But in terms of once it's live, this will basically be the utility token of the entire eco-system, so anybody, not just within our Nynja App or platform, but even people, I don't know if you know XMPP federation, like back in the day-- >> Yeah you know about real messaging >> If you could think of us as the next version of XMPP federation, but using cryptocurrency in order to avoid bad actors by making it very expensive to do bad things, and very cheap to do good things and globally. >> So it's like Twitter you can create a bot instantly, but if there's coins involved, you'd have to spend to get it. >> That's right and also people could spin up nodes that are basically their own Twitters and decide if those Twitters of their own, their Nynja boxes of their own, are either just internally, or you could specify specifically context or group of context-- >> We agree, that's a great way to get bad actors out because it costs them money. And it's de-centralized, there's no single spot. >> That's right, if email came out today, when cryptocurrency existed, there would be no spam. Because it would be expensive as hell to send more than a few a second, but it would still be free and for everybody generally, and you wouldn't even have spam. So we think we can do that for messaging globally. >> Great. Marshall, thanks so much for coming on theCUBE, really appreciate it, check out Nynja. Marshall Taplits is the Chief Strategy Officer and co-founder of Nynja.biz, check them out online. Check out the website, it's in Asia, bringing that culture of mobile and fast moving, real time apps, to the rest of the developers. This is theCUBE coverage in Puerto Rico for BlockChain Unbound exclusive two days of coverage. We'll be right back with more, after this short break, thanks for watching.
SUMMARY :
Brought to you by Blockchain Industries. as the global society and So tell about what you guys do. the Chinese internet, which as we know, go global as part of the to start companies in China. the legal system, and but the company is in Hong Kong. Chinese kind of showing the way of the wave of what mobile's doing, and the platforms that are out there, So I got to ask you about But the business and store the data as close of the money aspect, cause Everything else is just kind This is the beginning, to your point, So, and in terms of the People losing money for the first time. and then if they're down People that are into the game, in the early 90s, you connect the dots, is that if you look They might not not So one of the issues with the economy Part of the whole search and that can happen anywhere in the world. terms of the architecture, The reusability, the function in the world Quite frankly, the US is It's the first time I've the messenger with their payment It's a system, but the value you offer that's going to get us started like back in the day-- in order to avoid bad actors by making it So it's like Twitter you And it's de-centralized, and you wouldn't even have spam. Marshall Taplits is the
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Europe | LOCATION | 0.99+ |
America | LOCATION | 0.99+ |
Korea | LOCATION | 0.99+ |
Asia | LOCATION | 0.99+ |
Shenzhen | LOCATION | 0.99+ |
China | LOCATION | 0.99+ |
$100,000 | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
Alibaba | ORGANIZATION | 0.99+ |
Marshall | PERSON | 0.99+ |
April | DATE | 0.99+ |
US | LOCATION | 0.99+ |
Puerto Rico | LOCATION | 0.99+ |
Financial Conduct Authority | ORGANIZATION | 0.99+ |
Hong Kong | LOCATION | 0.99+ |
30 days | QUANTITY | 0.99+ |
John Furrier | PERSON | 0.99+ |
45% | QUANTITY | 0.99+ |
Hong Kong | LOCATION | 0.99+ |
20 year | QUANTITY | 0.99+ |
three | QUANTITY | 0.99+ |
15 years | QUANTITY | 0.99+ |
NYNJA Group | ORGANIZATION | 0.99+ |
two days | QUANTITY | 0.99+ |
Fiat | ORGANIZATION | 0.99+ |
Amazon Web Services | ORGANIZATION | 0.99+ |
yesterday | DATE | 0.99+ |
40% | QUANTITY | 0.99+ |
800 million | QUANTITY | 0.99+ |
two things | QUANTITY | 0.99+ |
an hour | QUANTITY | 0.99+ |
Marshall Taplits | PERSON | 0.99+ |
Niki Bond | PERSON | 0.99+ |
Wyoming | LOCATION | 0.99+ |
Marhall Taplits | PERSON | 0.99+ |
5 billion tokens | QUANTITY | 0.99+ |
Nynja.biz | ORGANIZATION | 0.99+ |
Sou Kanai | PERSON | 0.99+ |
five minute | QUANTITY | 0.99+ |
one | QUANTITY | 0.98+ |
a week | DATE | 0.98+ |
first time | QUANTITY | 0.98+ |
five platforms | QUANTITY | 0.98+ |
two billion | QUANTITY | 0.98+ |
first opportunity | QUANTITY | 0.98+ |
Lira | OTHER | 0.98+ |
today | DATE | 0.98+ |
two ago | DATE | 0.98+ |
Nynja | ORGANIZATION | 0.98+ |
UK | ORGANIZATION | 0.98+ |
two schools | QUANTITY | 0.98+ |
Chinese | OTHER | 0.97+ |
two trends | QUANTITY | 0.97+ |
last night | DATE | 0.97+ |
early 90s | DATE | 0.97+ |
Wuzhen China | LOCATION | 0.97+ |
Alipay | ORGANIZATION | 0.97+ |
about 50 additional features | QUANTITY | 0.97+ |
Deutsche Mark | OTHER | 0.97+ |
first companies | QUANTITY | 0.96+ |
two billion | QUANTITY | 0.96+ |
up to a billion | QUANTITY | 0.96+ |
S3 | TITLE | 0.96+ |
more than a few a second | QUANTITY | 0.95+ |