Image Title

Search Results for Forager Eatery:

Yuanhao Sun, Transwarp | Big Data SV 2018


 

>> Announcer: Live, from San Jose, it's The Cube (light music) Presenting Big Data Silicon Valley. Brought to you by Silicon Angle Media, and its ecosystem partners. >> Hi, I'm Peter Burris and welcome back to Big Data SV, The Cube's, again, annual broadcast of what's happening in the big data marketplace here at, or adjacent to Strada here in San Jose. We've been broadcasting all day. We're going to be here tomorrow as well, over at the Forager eatery and place to come meander. So come on over. Spend some time with us. Now, we've had a number of great guests. Many of the thought leaders that are visiting here in San Jose today were on the big data marketplace. But I don't think any has traveled as far as our next guest. Yuanhao Sun is the ceo of Transwarp. Come all the way from Shanghai Yuanhao. It's once again great to see you on The Cube. Thank you very much for being here. >> Good to see you again. >> So Yuanhao, the Transwarp as a company has become extremely well known for great technology. There's a lot of reasons why that's the case, but you have some interesting updates on how the technology's being applied. Why don't you tell us what's going on? >> Okay, so, recently we announced the first order to the TPC-DS benchmark result. Our product, calling scepter, that is, SQL engine on top of Hadoop. We already add quite a lot of features, like dissre transactions, like a full SQL support. So that it can mimic, like oracle or the mutual, and also traditional database features so that we can pass the whole test. This single is also scalable, because it's distributed, scalable. So the large benchmark, like TPC-DS. It starts from 10 terabytes. SQL engine can pester without much trouble. >> So I know that there have been other firms that have claimed to pass TPCC-DS, but they haven't been audited. What does it mean to say you're audited? I'd presume that as a result, you've gone through some extremely stringent and specific tests to demonstrate that you can actually pass the entire suite. >> Yes, actually, there is a third party auditor. They already audit our test process and it results for the passed six, uh, five months. So it is fully audited. The reason why we can pass the test is because, actually, there's two major reasons for traditional databases. They are not scalable to the process large dataset. So they could not pass the test. For (mumbles) vendors, because the SQL engine, the features to reach enough to pass all the test. You know, there several steps in the benchmark, and the SQL queries, there are 99 queries, the syntax is not supported by all howve vendors yet. And also, the benchmark required to upload the data, after the queries, and then we run the queries for multiple concurrent users. That means you have to support disputed transactions. You have to make the upload data consistent. For howve vendors, the SQL engine on Hadoop. They haven't implemented the de-switch transaction capabilities. So that's why they failed to pass the benchmark. >> So I had the honor of traveling to Shanghai last year and going and speaking at your user conference and was quite impressed with the energy that was in the room as you announced a large number of new products. You've been very focused on taking what open source has to offer but adding significant value to it. As you said, you've done a lot with the SQL interfaces and various capabilities of SQL on top of Hadoop. Where is Transwarp going with its products today? How is it expanding? How is it being organizing? How is it being used? >> We group these products into three catalog, including big data, cloud, AI and the machine learning. So there are three categories. The big data, we upgrade the SQL engine, the stream engine, and we have a set of tools called adjustable studio to help people to streamline the big data operations. And the second part I lie is data cloud. We call it transwarp data cloud. So this product is going to be raised in early in May this year. So this product we build this product on top of common idiots. We provide how to buy the service, get a sense as service, air as a service to customers. A lot of people took credit multiple tenets. And they turned as isolated by network, storage, cpu. They free to create a clusters and speeding up on turning it off. So it can also scale hundreds of cost. So this is the, I think this is the first we implement, like, a network isolation and sweaty percendency in cobinets. So that it can support each day affairs and all how to components. And because it is elastic, just like car computing, but we run on bare model, people can consult the data, consult the applications in one place. Because all application and Hadoop components are conternalized, that means, we are talking images. We can spend up a very quickly and scale through a larger cluster. So this data cloud product is very interesting for large company, because they usually have a small IT team. But they have to provide a (mumbles), and a machine only capability to larger groups, like one found the people. So they need a convenient way to manage all these bigger clusters. And they have to isolate the resources. Even they need a bidding system. So this product is, we already have few big names in China, like China Post, Picture Channel, and Secret of Source Channel. So they are already applying this data cloud for their internal customers. >> And China has a, has a few people, so I presume that, you know, China Post for example, is probably a pretty big implementation. >> Yes so, they have a, but the IT team is, like less than 100 people, but they have to support thousands of users. So that's why they, you usually would deploy 100 cluster for each application, right, but today, for large organization, they have lots of applications. They hope to leverage big data capability, but a very small team, IT team, can also part of so many applications. So they need a convenient the way like a, just like when you put Hadoop on public cloud. We provide a product that allows you to provide a hardware service in private cloud on bare model machines. So this is the second product category. And the third is the machine learning and artificial intelligence. We provide a data sales platform, a machine learning tool, that is, interactive tools that allows people to create the machine only pipelines and models. We even implemented some automatic modeling capability that allow you to, to fisher in youring automatically or seeming automatically and to select the best items for you so that the machine learning can be, so everyone can be at Los Angeles. So they can use our tool to quickly create a models. And we also have some probuter models for different industry, like financial service, like banks, security companies, even iot. So we have different probuter machine only models for them. We just need to modify the template, then apply the machine only models to the applications very quickly. So that probably like a lesson, for example, for a bank customer, they just use it to deploy a model in one week. This is very quick for them. Otherwise, in the past, they have a company to build that application, to develop much models. They usually takes several months. Today it is much faster. So today we have three categories, particularly like cloud and machine learning. >> Peter Burris: Machine learning and AI. >> And so three products. >> And you've got some very, very big implementations. So you were talking about a couple of banks, but we were talking, before we came on, about some of the smart cities. >> Yuanhao Sun: Right. Kinds of things that you guys are doing at enormous scale. >> Yes, so we deploy our streaming productor for more than 300 cities in China. So this cluster is like connected together. So we use streaming capability to monitor the traffic and send the information from city to the central government. So all the, the sort of essential repoetry. So whenever illegal behavior on the road is detected, that information will be sent to the policeman, or the central repoetry within two second. Whenever you are seen by the camera in any place in China, their loads where we send out within two seconds. >> So the bad behavior is detected. It's identified as the location. The system also knows where the nearest police person is. And it sends a message and says, this car has performed something bad. >> Yeah and you should stop that car in the next station or in the next crossroad. Today there are tens of thousands policeman. They depends on this system for their daily work. >> Peter Burris: Interesting. >> So, just a question on, it sounds like one of your, sort of nearest competitors, in terms of, let's take the open source community, at least the APIs, and in their case open source, Waway. Have their been customers that tried to do a POC with you and with Waway, and said, well it took four months using the pure open source stuff, and it took, say, two weeks with your stack having, being much broader and deeper? Are any examples like that? >> There are quite a lot. We have more macro-share, like in financial services, we have about 100 bank users. So if we take all banks into account, for them they already use Hadoop. So we, our macro-share is above 60%. >> George Gilbert: 60. >> Yeah, in financial services. We usually do POC and, like run benchmarks. They are real workloads and usually it takes us three days or one week. They can found, we can speed up their workload very quickly. For Bank of China, they might go to their oracle workload to our platform. And they test our platform and the huave platform too. So the first thing is they cannot marry the whole oracle workload to open source Hadoop, because the missing features. We are able to support all this workloads with very minor modifications. So the modification takes only several hours. And we can finish the whole workload within two hours, but originally they take, usually take oracle more than one day, >> George Gilbert: Wow. >> more than ten hours to finish the workload. So it is very easy to see the benefits quickly. >> Now the you have a streaming product also with that same SQL interface. Are you going to see a migration of applications that used to be batch to more near real time or continuous, or will you see a whole new set of applications that weren't done before, because the latency wasn't appropriate? >> For streaming applications, real time cases they are mostly new applications, but if we are using storm api or spark streaming api, it is not so easy to develop your applications. And another issue is once you detect one new rule, you had to add those rules dynamically to your cluster. So to add to your printer, they do not have so many knowledge of writing scholar codes. They only know how to configure. Probably they are familiar with c-code. They just need to add one SQL statement to add a new rule. So that they can. >> In your system. >> Yeah, in our system. So it is much easier for them to program streaming applications. And for those customers who they don't have real time equations, they hope to do, like a real time data warehousing. They collect all this data from websites from their censors, like Petrol Channel, an oil company, the large oil company. They collect all the (mumbles) information directly to our streaming product. In the past, they just accredit to oracle and around the dashboard. So it only takes hours to see the results. But today, the application can be moved through our streaming product with only a few modifications, because they are all SQL statements. And this application becomes the real time. They can see the real time dashboard results in several seconds. >> So Yuanhao, you're number one in China. You're moving more aggressively to participate in the US market. What's the, last question, what's the biggest difference between being number one in China, the way that big data is being done in China versus the way you're encountering big data being done here, certainly in the US, for example? Is there a difference? >> I think there are some difference. Some a seem, katsumoto usually request a POC. But in China, they usually, I think they focus more on the results. They focus on what benefit they can gain from your product. So we have to prove them. So we have to hip them to my great application to see the benefits. I think in US, they focus more on technology than Chinese customers. >> Interesting, so they're more on technology here in the US, more in the outcome in China. Once again, Yuanhao Sun, from, ceo of Transwarp, thank you very much for being on The Cube. >> Thank you. And I'm Peter Burris with George Gilbert, my co-host, and we'll be back with more from big data SV, in San Jose. Come on over to the Forager, and spend some time with us. And we'll be back in a second. (light music)

Published Date : Mar 8 2018

SUMMARY :

Brought to you by Silicon Angle Media, over at the Forager eatery and place to come meander. So Yuanhao, the Transwarp as a company has become So that it can mimic, like oracle or the mutual, to demonstrate that you can actually pass the entire suite. And also, the benchmark required to upload the data, So I had the honor of traveling to Shanghai last year So this product is going to be raised you know, China Post for example, and to select the best items for you So you were talking about a couple of banks, Kinds of things that you guys are doing at enormous scale. from city to the central government. So the bad behavior is detected. or in the next crossroad. and it took, say, two weeks with your stack having, So if we take all banks into account, So the first thing is they cannot more than ten hours to finish the workload. Now the you have a streaming product also So to add to your printer, So it only takes hours to see the results. to participate in the US market. So we have to prove them. in the US, more in the outcome in China. Come on over to the Forager, and spend some time with us.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Peter BurrisPERSON

0.99+

ShanghaiLOCATION

0.99+

George GilbertPERSON

0.99+

USLOCATION

0.99+

ChinaLOCATION

0.99+

99 queriesQUANTITY

0.99+

three daysQUANTITY

0.99+

two weeksQUANTITY

0.99+

Silicon Angle MediaORGANIZATION

0.99+

five monthsQUANTITY

0.99+

San JoseLOCATION

0.99+

China PostORGANIZATION

0.99+

Picture ChannelORGANIZATION

0.99+

one weekQUANTITY

0.99+

sixQUANTITY

0.99+

four monthsQUANTITY

0.99+

Los AngelesLOCATION

0.99+

10 terabytesQUANTITY

0.99+

last yearDATE

0.99+

todayDATE

0.99+

TodayDATE

0.99+

tomorrowDATE

0.99+

more than one dayQUANTITY

0.99+

more than 300 citiesQUANTITY

0.99+

second partQUANTITY

0.99+

two hoursQUANTITY

0.99+

less than 100 peopleQUANTITY

0.99+

more than ten hoursQUANTITY

0.99+

WawayORGANIZATION

0.99+

Bank of ChinaORGANIZATION

0.99+

thirdQUANTITY

0.99+

HadoopTITLE

0.99+

Petrol ChannelORGANIZATION

0.99+

three productsQUANTITY

0.98+

one new ruleQUANTITY

0.98+

hundredsQUANTITY

0.98+

three categoriesQUANTITY

0.98+

SQLTITLE

0.98+

singleQUANTITY

0.98+

TranswarpORGANIZATION

0.98+

firstQUANTITY

0.98+

tens of thousands policemanQUANTITY

0.98+

Yuanhao SunORGANIZATION

0.98+

each applicationQUANTITY

0.98+

two secondsQUANTITY

0.98+

100 clusterQUANTITY

0.97+

first thingQUANTITY

0.97+

about 100 bank usersQUANTITY

0.97+

two secondQUANTITY

0.97+

each dayQUANTITY

0.97+

Big Data SVORGANIZATION

0.97+

The CubeORGANIZATION

0.96+

two major reasonsQUANTITY

0.95+

oneQUANTITY

0.95+

above 60%QUANTITY

0.95+

early in May this yearDATE

0.94+

Source ChannelORGANIZATION

0.93+

Big DataORGANIZATION

0.92+

ChineseOTHER

0.9+

StradaLOCATION

0.89+

second product categoryQUANTITY

0.88+

Seth Dobrin, IBM | Big Data SV 2018


 

>> Announcer: Live from San Jose, it's theCUBE. Presenting Big Data Silicon Valley, brought to you by SiliconANGLE Media and it's ecosystem partners. >> Welcome back to theCUBE's continuing coverage of our own event, Big Data SV. I'm Lisa Martin, with my cohost Dave Vellante. We're in downtown San Jose at this really cool place, Forager Eatery. Come by, check us out. We're here tomorrow as well. We're joined by, next, one of our CUBE alumni, Seth Dobrin, the Vice President and Chief Data Officer at IBM Analytics. Hey, Seth, welcome back to theCUBE. >> Hey, thanks for having again. Always fun being with you guys. >> Good to see you, Seth. >> Good to see you. >> Yeah, so last time you were chatting with Dave and company was about in the fall at the Chief Data Officers Summit. What's kind of new with you in IBM Analytics since then? >> Yeah, so the Chief Data Officers Summit, I was talking with one of the data governance people from TD Bank and we spent a lot of time talking about governance. Still doing a lot with governance, especially with GDPR coming up. But really started to ramp up my team to focus on data science, machine learning. How do you do data science in the enterprise? How is it different from doing a Kaggle competition, or someone getting their PhD or Masters in Data Science? >> Just quickly, who is your team composed of in IBM Analytics? >> So IBM Analytics represents, think of it as our software umbrella, so it's everything that's not pure cloud or Watson or services. So it's all of our software franchise. >> But in terms of roles and responsibilities, data scientists, analysts. What's the mixture of-- >> Yeah. So on my team I have a small group of people that do governance, and so they're really managing our GDPR readiness inside of IBM in our business unit. And then the rest of my team is really focused on this data science space. And so this is set up from the perspective of we have machine-learning engineers, we have predictive-analytics engineers, we have data engineers, and we have data journalists. And that's really focus on helping IBM and other companies do data science in the enterprise. >> So what's the dynamic amongst those roles that you just mentioned? Is it really a team sport? I mean, initially it was the data science on a pedestal. Have you been able to attack that problem? >> So I know a total of two people that can do that all themselves. So I think it absolutely is a team sport. And it really takes a data engineer or someone with deep expertise in there, that also understands machine-learning, to really build out the data assets, engineer the features appropriately, provide access to the model, and ultimately to what you're going to deploy, right? Because the way you do it as a research project or an activity is different than using it in real life, right? And so you need to make sure the data pipes are there. And when I look for people, I actually look for a differentiation between machine-learning engineers and optimization. I don't even post for data scientists because then you get a lot of data scientists, right? People who aren't really data scientists, and so if you're specific and ask for machine-learning engineers or decision optimization, OR-type people, you really get a whole different crowd in. But the interplay is really important because most machine-learning use cases you want to be able to give information about what you should do next. What's the next best action? And to do that, you need decision optimization. >> So in the early days of when we, I mean, data science has been around forever, right? We always hear that. But in the, sort of, more modern use of the term, you never heard much about machine learning. It was more like stats, math, some programming, data hacking, creativity. And then now, machine learning sounds fundamental. Is that a new skillset that the data scientists had to learn? Did they get them from other parts of the organization? >> I mean, when we talk about math and stats, what we call machine learning today has been what we've been doing since the first statistics for years, right? I mean, a lot of the same things we apply in what we call machine learning today I did during my PhD 20 years ago, right? It was just with a different perspective. And you applied those types of, they were more static, right? So I would build a model to predict something, and it was only for that. It really didn't apply it beyond, so it was very static. Now, when we're talking about machine learning, I want to understand Dave, right? And I want to be able to predict Dave's behavior in the future, and learn how you're changing your behavior over time, right? So one of the things that a lot of people don't realize, especially senior executives, is that machine learning creates a self-fulfilling prophecy. You're going to drive a behavior so your data is going to change, right? So your model needs to change. And so that's really the difference between what you think of as stats and what we think of as machine learning today. So what we were looking for years ago is all the same we just described it a little differently. >> So how fine is the line between a statistician and a data scientist? >> I think any good statistician can really become a data scientist. There's some issues around data engineering and things like that but if it's a team sport, I think any really good, pure mathematician or statistician could certainly become a data scientist. Or machine-learning engineer. Sorry. >> I'm interested in it from a skillset standpoint. You were saying how you're advertising to bring on these roles. I was at the Women in Data Science Conference with theCUBE just a couple of days ago, and we hear so much excitement about the role of data scientists. It's so horizontal. People have the opportunity to make impact in policy change, healthcare, etc. So the hard skills, the soft skills, mathematician, what are some of the other elements that you would look for or that companies, enterprises that need to learn how to embrace data science, should look for? Someone that's not just a mathematician but someone that has communication skills, collaboration, empathy, what are some of those, openness, to not lead data down a certain, what do you see as the right mix there of a data scientist? >> Yeah, so I think that's a really good point, right? It's not just the hard skills. When my team goes out, because part of what we do is we go out and sit with clients and teach them our philosophy on how you should integrate data science in the enterprise. A good part of that is sitting down and understanding the use case. And working with people to tease out, how do you get to this ultimate use case because any problem worth solving is not one model, any use case is not one model, it's many models. How do you work with the people in the business to understand, okay, what's the most important thing for us to deliver first? And it's almost a negotiation, right? Talking them back. Okay, we can't solve the whole problem. We need to break it down in discreet pieces. Even when we break it down into discreet pieces, there's going to be a series of sprints to deliver that. Right? And so having these soft skills to be able to tease that in a way, and really help people understand that their way of thinking about this may or may not be right. And doing that in a way that's not offensive. And there's a lot of really smart people that can say that, but they can come across at being offensive, so those soft skills are really important. >> I'm going to talk about GDPR in the time we have remaining. We talked about in the past, the clocks ticking, May the fines go into effect. The relationship between data science, machine learning, GDPR, is it going to help us solve this problem? This is a nightmare for people. And many organizations aren't ready. Your thoughts. >> Yeah, so I think there's some aspects that we've talked about before. How important it's going to be to apply machine learning to your data to get ready for GDPR. But I think there's some aspects that we haven't talked about before here, and that's around what impact does GDPR have on being able to do data science, and being able to implement data science. So one of the aspects of the GDPR is this concept of consent, right? So it really requires consent to be understandable and very explicit. And it allows people to be able to retract that consent at any time. And so what does that mean when you build a model that's trained on someone's data? If you haven't anonymized it properly, do I have to rebuild the model without their data? And then it also brings up some points around explainability. So you need to be able to explain your decision, how you used analytics, how you got to that decision, to someone if they request it. To an auditor if they request it. Traditional machine learning, that's not too much of a problem. You can look at the features and say these features, this contributed 20%, this contributed 50%. But as you get into things like deep learning, this concept of explainable or XAI becomes really, really important. And there were some talks earlier today at Strata about how you apply machine learning, traditional machine learning to interpret your deep learning or black box AI. So that's really going to be important, those two things, in terms of how they effect data science. >> Well, you mentioned the black box. I mean, do you think we'll ever resolve the black box challenge? Or is it really that people are just going to be comfortable that what happens inside the box, how you got to that decision is okay? >> So I'm inherently both cynical and optimistic. (chuckles) But I think there's a lot of things we looked at five years ago and we said there's no way we'll ever be able to do them that we can do today. And so while I don't know how we're going to get to be able to explain this black box as a XAI, I'm fairly confident that in five years, this won't even be a conversation anymore. >> Yeah, I kind of agree. I mean, somebody said to me the other day, well, it's really hard to explain how you know it's a dog. >> Seth: Right (chuckles). But you know it's a dog. >> But you know it's a dog. And so, we'll get over this. >> Yeah. >> I love that you just brought up dogs as we're ending. That's my favorite thing in the world, thank you. Yes, you knew that. Well, Seth, I wish we had more time, and thanks so much for stopping by theCUBE and sharing some of your insights. Look forward to the next update in the next few months from you. >> Yeah, thanks for having me. Good seeing you again. >> Pleasure. >> Nice meeting you. >> Likewise. We want to thank you for watching theCUBE live from our event Big Data SV down the street from the Strata Data Conference. I'm Lisa Martin, for Dave Vellante. Thanks for watching, stick around, we'll be rick back after a short break.

Published Date : Mar 8 2018

SUMMARY :

brought to you by SiliconANGLE Media Welcome back to theCUBE's continuing coverage Always fun being with you guys. Yeah, so last time you were chatting But really started to ramp up my team So it's all of our software franchise. What's the mixture of-- and other companies do data science in the enterprise. that you just mentioned? And to do that, you need decision optimization. So in the early days of when we, And so that's really the difference I think any good statistician People have the opportunity to make impact there's going to be a series of sprints to deliver that. in the time we have remaining. And so what does that mean when you build a model Or is it really that people are just going to be comfortable ever be able to do them that we can do today. I mean, somebody said to me the other day, But you know it's a dog. But you know it's a dog. I love that you just brought up dogs as we're ending. Good seeing you again. We want to thank you for watching theCUBE

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Dave VellantePERSON

0.99+

Lisa MartinPERSON

0.99+

SethPERSON

0.99+

DavePERSON

0.99+

IBMORGANIZATION

0.99+

Seth DobrinPERSON

0.99+

20%QUANTITY

0.99+

50%QUANTITY

0.99+

TD BankORGANIZATION

0.99+

San JoseLOCATION

0.99+

two peopleQUANTITY

0.99+

tomorrowDATE

0.99+

IBM AnalyticsORGANIZATION

0.99+

two thingsQUANTITY

0.99+

SiliconANGLE MediaORGANIZATION

0.99+

one modelQUANTITY

0.99+

five yearsQUANTITY

0.98+

20 years agoDATE

0.98+

Big Data SVEVENT

0.98+

five years agoDATE

0.98+

GDPRTITLE

0.98+

theCUBEORGANIZATION

0.98+

oneQUANTITY

0.98+

Strata Data ConferenceEVENT

0.97+

todayDATE

0.97+

first statisticsQUANTITY

0.95+

CUBEORGANIZATION

0.94+

Women in Data Science ConferenceEVENT

0.94+

bothQUANTITY

0.94+

Chief Data Officers SummitEVENT

0.93+

Big Data SV 2018EVENT

0.93+

couple of days agoDATE

0.93+

yearsDATE

0.9+

Forager EateryORGANIZATION

0.9+

firstQUANTITY

0.86+

WatsonTITLE

0.86+

Officers SummitEVENT

0.74+

Data OfficerPERSON

0.73+

SVEVENT

0.71+

PresidentPERSON

0.68+

StrataTITLE

0.67+

Big DataORGANIZATION

0.66+

earlier todayDATE

0.65+

Silicon ValleyLOCATION

0.64+

yearsQUANTITY

0.6+

ChiefEVENT

0.44+

KaggleORGANIZATION

0.43+

Dr. Tendu Yogurtcu, Syncsort | Big Data SV 2018


 

>> Announcer: Live from San Jose, it's theCUBE. Presenting data, Silicon Valley brought to you by Silicon Angle Media and it's ecosystem partners. >> Welcome back to theCUBE. We are live in San Jose at our event, Big Data SV. I'm Lisa Martin, my co-host is George Gilbert and we are down the street from the Strata Data Conference. We are at a really cool venue: Forager Eatery Tasting Room. Come down and join us, hang out with us, we've got a cocktail par-tay tonight. We also have an interesting briefing from our analysts on big data trends tomorrow morning. I want to welcome back to theCUBE now one of our CUBE VIP's and alumna Tendu Yogurtcu, the CTO at Syncsort, welcome back. >> Thank you. Hello Lisa, hi George, pleasure to be here. >> Yeah, it's our pleasure to have you back. So, what's going on at Syncsort, what are some of the big trends as CTO that you're seeing? >> In terms of the big trends that we are seeing, and Syncsort has grown a lot in the last 12 months, we actually doubled our revenue, it has been really an successful and organic growth path, and we have more than 7,000 customers now, so it's a great pool of customers that we are able to talk and see the trends and how they are trying to adapt to the digital disruption and make data as part of their core strategy. So data is no longer an enabler, and in all of the enterprise we are seeing data becoming the core strategy. This reflects in the four mega trends, they are all connected to enable business as well as operational analytics. Cloud is one, definitely. We are seeing more and more cloud adoption, even our financial services healthcare and banking customers are now, they have a couple of clusters running in the cloud, in public cloud, multiple workloads, hybrid seems to be the new standard, and it comes with also challenges. IT governance as well as date governance is a major challenge, and also scoping and planning for the workloads in the cloud continues to be a challenge, as well. Our general strategy for all of the product portfolio is to have our products following design wants and deploy any of our strategy. So whether it's a standalone environment on Linux or running on Hadoop or Spark, or running on Premise or in the Cloud, regardless of the Cloud provider, we are enabling the same education with no changes to run all of these environments, including hybrid. Then we are seeing the streaming trend, with the connected devices with the digital disruption and so much data being generated, being able to stream and process data on the age, with the Internet of things, and in order to address the use cases that Syncsort is focused on, we are really providing more on the Change Data Capture and near real-time and real-time data replication to the next generation analytics environments and big data environments. We launched last year our Change Data Capture, CDC, product offering with data integration, and we continue to strengthen that vision merger we had data replication, real-time data replication capabilities, and we are now seeing even Kafka database becoming a consumer of this data. Not just keeping the data lane fresh, but really publishing the changes from multiple, diverse set of sources and publishing into a Kafka database and making it available for applications and analytics in the data pipeline. So the third trend we are seeing is around data science, and if you noticed this morning's keynote was all about machine learning, artificial intelligence, deep learning, how to we make use of data science. And it was very interesting for me because we see everyone talking about the challenge of how do you prepare the data and how do you deliver the the trusted data for machine learning and artificial intelligence use and deep learning. Because if you are using bad data, and creating your models based on bad data, then the insights you get are also impacted. We definitely offer our products, both on the data integration and data quality side, to prepare the data, cleanse, match, and deliver the trusted data set for data scientists and make their life easier. Another area of focus for 2018 is can we also add supervised learning to this, because with the premium quality domain experts that we have now in Syncsort, we have a lot of domain experts in the field, we can infuse the machine learning algorithms and connect data profiling capabilities we have with the data quality capabilities recommending business rules for data scientists and helping them automate the mandate tasks with recommendations. And the last but not least trend is data governance, and data governance is almost a umbrella focus for everything we are doing at Syncsort because everything about the Cloud trend, the streaming, and the data science, and developing that next generation analytics environment for our customers depends on the data governance. It is, in fact, a business imperative, and the regulatory compliance use cases drives more importance today than governance. For example, General Data Protection Regulation in Europe, GDPR. >> Lisa: Just a few months away. >> Just a few months, May 2018, it is in the mind of every C-level executive. It's not just for European companies, but every enterprise has European data sourced in their environments. So compliance is a big driver of governance, and we look at governance in multiple aspects. Security and issuing data is available in a secure way is one aspect, and delivering the high quality data, cleansing, matching, the example Hilary Mason this morning gave in the keynote about half of what the context matters in terms of searches of her name was very interesting because you really want to deliver that high quality data in the enterprise, trust of data set, preparing that. Our Trillium Quality for big data, we launched Q4, that product is generally available now, and actually we are in production with very large deployment. So that's one area of focus. And the third area is how do you create visibility, the farm-to-table view of your data? >> Lisa: Yeah, that's the name of your talk! I love that. >> Yes, yes, thank you. So tomorrow I have a talk at 2:40, March 8th also, I'm so happy it's on the Women's Day that I'm talking-- >> Lisa: That's right, that's right! Get a farm-to-table view of your data is the name of your talk, track data lineage from source to analytics. Tell us a little bit more about that. >> It's all about creating more visibility, because for audit reasons, for understanding how many copies of my data is created, valued my data had been, and who accessed it, creating that visibility is very important. And the last couple of years, we saw everyone was focused on how do I create a data lake and make my data accessible, break the data silos, and liberate my data from multiple platforms, legacy platforms that the enterprise might have. Once that happened, everybody started worrying about how do I create consumable data set and how do I manage this data because data has been on the legacy platforms like Mainframe, IMBI series has been on relational data stores, it is in the Cloud, gravity of data originating in the Cloud is increasing, it's originating from mobile. Hadoop vendors like Hortonworks and Cloudera, they are creating visibility to what happens within the Hadoop framework. So we are deepening our integration with the Cloud Navigator, that was our announcement last week. We already have integration both with Hortonworks and Cloudera Navigator, this is one step further where we actually publish what happened to every single granular level of data at the field level with all of the transformations that data have been through outside of the cluster. So that visibility is now published to Navigator itself, we also publish it through the RESTful API, so governance is a very strong and critical initiative for all of the businesses. And we are playing into security aspect as well as data lineage and tracking aspect and the quality aspect. >> So this sounds like an extremely capable infrastructure service, so that it's trusted data. But can you sell that to an economic buyer alone, or do you go in in conjunction with anther solution like anti-money laundering for banks or, you know, what are the key things that they place enough value on that they would spend, you know, budget on it? >> Yes, absolutely. Usually the use cases might originate like anti-money laundering, which is very common, fraud detection, and it ties to getting a single view of an entity. Because in anti-money laundering, you want to understand the single view of your customer ultimately. So there is usually another solution that might be in the picture. We are providing the visibility of the data, as well as that single view of the entity, whether it's the customer view in this case or the product view in some of the use cases by delivering the matching capabilities and the cleansing capabilities, the duplication capabilities in addition to the accessing and integrating the data. >> When you go into a customer and, you know, recognizing that we still have tons of silos and we're realizing it's a lot harder to put everything in one repository, how do customers tell you they want to prioritize what they're bringing into the repository or even what do they want to work on that's continuously flowing in? >> So it depends on the business use case. And usually at the time that we are working with the customer, they selected that top priority use case. The risk here, and the anti-money laundering, or for insurance companies, we are seeing a trend, for example, building the data marketplace, as that tantalize data marketplace concept. So depending on the business case, many of our insurance customers in US, for example, they are creating the data marketplace and they are working with near real-time and microbatches. In Europe, Europe seems to be a bit ahead of the game in some cases, like Hadoop production was slow but certainly they went right into the streaming use cases. We are seeing more directly streaming and keeping it fresh and more utilization of the Kafka and messaging frameworks and database. >> And in that case, where they're sort of skipping the batch-oriented approach, how do they keep track of history? >> It's still, in most of the cases, microbatches, and the metadata is still associated with the data. So there is an analysis of the historical what happened to that data. The tools, like ours and the vendors coming to picture, to keep track, of that basically. >> So, in other words, by knowing what happened operationally to the data, that paints a picture of a history. >> Exactly, exactly. >> Interesting. >> And for the governance we usually also partner, for example, we partner with Collibra data platform, we partnered with ASG for creating that business rules and technical metadata and providing to the business users, not just to the IT data infrastructure, and on the Hadoop side we partner with Cloudera and Hortonworks very closely to complete that picture for the customer, because nobody is just interested in what happened to the data in Hadoop or in Mainframe or in my relational data warehouse, they are really trying to see what's happening on Premise, in the Cloud, multiple clusters, traditional environments, legacy systems, and trying to get that big picture view. >> So on that, enabling a business to have that, we'll say in marketing, 360 degree view of data, knowing that there's so much potential for data to be analyzed to drive business decisions that might open up new business models, new revenue streams, increase profit, what are you seeing as a CTO of Syncsort when you go in to meet with a customer, data silos, when you're talking to a Chief Data Officer, what's the cultural, I guess, not shift but really journey that they have to go on to start opening up other organizations of the business, to have access to data so they really have that broader, 360 degree view? What's that cultural challenge that they have to, journey that they have to go on? >> Yes, Chief Data Officers are actually very good partners for us, because usually Chief Data Officers are trying to break the silos of data and make sure that the data is liberated for the business use cases. Still most of the time the infrastructure and the cluster, whether it's the deployment in the Cloud versus on Premise, it's owned by the IT infrastructure. And the lines of business are really the consumers and the clients of that. CDO, in that sense, almost mitigates and connects to those line of businesses with the IT infrastructure with the same goals for the business, right? They have to worry about the compliance, they have to worry about creating multiple copies of data, they have to worry about the security of the data and availability of the data, so CDOs actually help. So we are actually very good partners with the CDOs in that sense, and we also usually have IT infrastructure owner in the room when we are talking with our customers because they have a big stake. They are like the gatekeepers of the data to make sure that it is accessed by the right... By the right folks in the business. >> Sounds like maybe they're in the role of like, good cop bad cop or maybe mediator. Well Tendu, I wish we had more time. Thanks so much for coming back to theCUBE and, like you said, you're speaking tomorrow at Strata Conference on International Women's Day: Get a farm-to-table view of your data. Love the title. >> Thank you. >> Good luck tomorrow, and we look forward to seeing you back on theCUBE. >> Thank you, I look forward to coming back and letting you know about more exciting both organic innovations and acquisitions. >> Alright, we look forward to that. We want to thank you for watching theCUBE, I'm Lisa Martin with my co-host George Gilbert. We are live at our event Big Data SV in San Jose. Come down and visit us, stick around, and we will be right back with our next guest after a short break. >> Tendu: Thank you. (upbeat music)

Published Date : Mar 7 2018

SUMMARY :

brought to you by Silicon Angle Media and we are down the street from the Strata Data Conference. Hello Lisa, hi George, pleasure to be here. Yeah, it's our pleasure to have you back. and in all of the enterprise we are seeing data and delivering the high quality data, Lisa: Yeah, that's the name of your talk! it's on the Women's Day that I'm talking-- is the name of your talk, track data lineage and make my data accessible, break the data silos, that they place enough value on that they would and the cleansing capabilities, the duplication So it depends on the business use case. It's still, in most of the cases, operationally to the data, that paints a picture And for the governance we usually also partner, and the cluster, whether it's the deployment Love the title. to seeing you back on theCUBE. and letting you know about more exciting and we will be right back with our next guest Tendu: Thank you.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Lisa MartinPERSON

0.99+

GeorgePERSON

0.99+

May 2018DATE

0.99+

George GilbertPERSON

0.99+

SyncsortORGANIZATION

0.99+

LisaPERSON

0.99+

EuropeLOCATION

0.99+

HortonworksORGANIZATION

0.99+

USLOCATION

0.99+

Hilary MasonPERSON

0.99+

San JoseLOCATION

0.99+

ASGORGANIZATION

0.99+

2018DATE

0.99+

TenduPERSON

0.99+

Silicon Angle MediaORGANIZATION

0.99+

ClouderaORGANIZATION

0.99+

360 degreeQUANTITY

0.99+

tomorrowDATE

0.99+

CollibraORGANIZATION

0.99+

more than 7,000 customersQUANTITY

0.99+

last weekDATE

0.99+

last yearDATE

0.99+

tomorrow morningDATE

0.99+

one aspectQUANTITY

0.99+

third areaQUANTITY

0.99+

LinuxTITLE

0.99+

Cloud NavigatorTITLE

0.99+

2:40DATE

0.98+

Women's DayEVENT

0.98+

Tendu YogurtcuPERSON

0.98+

GDPRTITLE

0.98+

SparkTITLE

0.97+

tonightDATE

0.97+

Big Data SVEVENT

0.97+

KafkaTITLE

0.97+

International Women's DayEVENT

0.97+

bothQUANTITY

0.97+

CDCORGANIZATION

0.96+

NavigatorTITLE

0.96+

Strata Data ConferenceEVENT

0.96+

single viewQUANTITY

0.96+

HadoopTITLE

0.95+

third trendQUANTITY

0.95+

one stepQUANTITY

0.95+

single viewQUANTITY

0.95+

Dr.PERSON

0.94+

theCUBEORGANIZATION

0.94+

CUBEORGANIZATION

0.94+

this morningDATE

0.94+

CloudTITLE

0.92+

last 12 monthsDATE

0.91+

Change Data CaptureORGANIZATION

0.9+

todayDATE

0.9+

EuropeanOTHER

0.88+

last couple of yearsDATE

0.88+

General Data Protection Regulation in EuropeTITLE

0.86+

Strata ConferenceEVENT

0.84+

oneQUANTITY

0.83+

one repositoryQUANTITY

0.83+

tons of silosQUANTITY

0.82+

one areaQUANTITY

0.82+

Q4DATE

0.82+

Big Data SV 2018EVENT

0.81+

four mega trendsQUANTITY

0.76+

March 8thDATE

0.76+

Jacque Istok, Pivotal | Big Data SV 2018


 

>> Announcer: Live from San Jose, it's The Cube. Presenting Big Data, Silicon Valley. Brought to you by SiliconANGLE Media and its ecosystem partners. >> Welcome back to The Cube, we are live in San Jose at Forager Eatery, a really cool place down the street from the Strata Data Conference. This is our 10th big data event, we call this BigData SV, we've done five here, five in New York, and this is our day one of coverage, I'm Lisa Martin with my co-host George Gilbert, and we're joined by a Cube alumni, Jacque Istok, the head of data from Pivotal. Welcome back to the cube, Jacque. >> Thank you, it's great to be here. >> So, just recently you guys announced, Pivotal announced, the GA of your Kubernetes-based Pivotal container service, PKS following this initial beta that you guys released last year, tell us about that, what's the main idea behind PKS? >> So, as we were talking about earlier, we've had this opinionated platform as a service for the last couple of years, it's taken off, but it really requires a very specific methodology for deploying microservices and kind of next gen applications, and what we've seen with the ground swell behind Kubernetes is a very seamless way where we can not just do our opinionated applications, we can do any applications leveraging Kubernetes. In addition, it actually allows us to again, kind of have an opinionated way to work with stateful, stateful data, if you will. And so, what you'll see is two of the main things we have going on, again, if you look at both of those products they're all managed by a thing we call Bosch and Bosch allows for not just the ease of installation, but also the actual operation of the entire platform. And so, what we're seeing is the ability to do day two operations not just around just the apps, not just the platform, but also the data products that run within it. And you'll see later this year as we continue to evolve our data products running on top of either the PKS product or the PCF product. >> Quick question before you jump in George, so you talk about some of the technology benefits and reasoning for that, from a customer perspective, what are some of the key benefits that you've designed this for, or challenges to solve? >> I'd say the key benefits, one is convenience and ease of installation, and operationalization. Kubernetes seems to have basically become the standard for being able to deploy containers, whether its on Pram or off Pram, and having an enterprise solution to do that is something that customers are actually really looking towards, in fact, we had sold about a dozen of these products even before it was GA there was so much excitement around it. But, beyond that, I think we've been really focused on this idea of digital transformation. So Pivotal's whole talk track really is changing how companies build software. And I think the introduction of PKS really takes us to the next level, which is that there's no digital transformation without data, and basically Kubernetes and PKS allow us to implement that and perform for our customers. >> This is really a facilitator of a company's digital transformation journey. >> Correct. In a very easy and convenient way, and I think, you know, whether it's our generation, or, you know, what's going on in just technology, but everybody is so focused on convenience, push button, I just want it to work. I don't want to have to dig into the details. >> So this picks up on a theme we've been pounding on for a couple of years on our side, which is the infrastructure was too hard to stand up and operate >> Male Speaker: Yeah. >> But now that we're beginning to solve some of those problems, talk about some of the use case. Let's pick GE because that's a flagship customer, start with some of the big outcomes, some of the big business outcomes they're shooting for and then how some of the pivotal products map into that. >> Sure, so there's a lot of use cases. Obviously, GE is both a large organization, as well as an investor inside of Pivotal. A lot of different things we can talk about one that comes to mind out of the gate is we've got a data suite we sell in addition to PKS and PCF, and within that data suite there are a couple of products, green plum being one of them. Green plum is this open source MPP data platform. Probably one of the most successful implementations within GE is this ability to actually consolidate a bunch of different ERP data and have people be able to querey it, again, cheaply, easily, effectively and there are a lot of different ways you can implement a solution like that. I think what's attractive to these guys specifically around green plum is that it leverages, you know, standard ANSI SQL, it scales to pedobytes of data, we have this ability to do on pram and off pram I was actually at the Gartner Conference earlier this week and walking around the show it was actually somewhat eye opening to me to be able to see that if you look at just that one product, there really isn't a competitive product that was being showcased that was open source, multi cloud, analytical in nature, et cetera. And so I think, again, to get back to the GE scenario, what was attractive to them was everything they're doing on pram can move to the cloud, whether it's Google, Azure, Amazon they can literally run the exact same product and the exact same queries. If you extend it beyond that particular use case, there are other use cases that are more real time, and again, inside of the data suite, we've got another product called gem fire, which is an in-memory data grid that allows for this rapid ingest, so you can kind of think and imagine whether it's jet engines, or whether it's wind turbines data is constantly being generated, and our ability to take that data in real time, ingest it, actually perform analytics on it as it comes in, so, again, kind of a loose example would be if you know the heat tolerance of a wind turbine is between this temperature and this temperature, do something: send an alarm, shut it down, et cetera. If you can do that in real time, you can actually save millions of dollars by not letting that turbine fail. >> Okay, it sounds here like the gem fire product and the green plum DBMS are very complimentary. You know, one is speed, and one is sort of throughput. And we've seen almost like with Hadupen overreaction in turning a coherent platform into a bunch of building blocks. >> Male Speaker: Yes. >> And with green plum you have everything packaged together. Would it be proper to think of green plum as combining the best of the data link and the data warehouse where you've got the data scientists and data engineers with what would have been another product and the business analysts and the BI crowd satisfied with the same product, but what would have been another? >> Male Speaker: So, I'd say you're spot on. What is super interesting to me is, one, I've been doing data warehousing now for, I don't know, 20 years, and for the last five, I've kind of felt like data warehouse, just the term, was equivalent to the mainframe. So, I actually kind of relegated it the I'm not going to use that term anymore, but with the advent of the cloud and with other products that are out there we're seeing this resurgence where the data warehouse is cool again, and I think part of it is because we had this shift where we had really expensive products doing the classic EDW and it was too rigid, and it was too expensive, and Haduke sort of came on and everyone was like hey this is really easy, this is really cheap, we can store whatever we want, we can do any kind of analytics, and I think, I was saying before, the love affair with piecing all of that together is kind of over and I also think, it's funny, it was really hard for organizations to successfully stand up a Haduke platform, and I think the metric we hear is fifty percent of them fail, right, so part of that, I believe is because there just aren't enough people to be able to do what needed to be done. So, interestingly enough, because of those failures, because the Haduke ecosystem didn't quite integrate into the classic enterprise, products like green plum are suddenly very popular. I was just seeing our downloads for the open source part of green plum, and we're literally, at this juncture seeing 1500 distinct customers leveraging the open source product, so I feel like we're on kind of this upswing of getting everybody to understand that you don't have to go to Haduke to be able to do structured to unstructured data at scale. You can actually use some of these other products. >> Female Speaker: Sorry George, quickly, being in the industry for 20 years, we talk about, you know, culture a lot, and we say cultural shift. People started embracing Haduke, we can dump everything that data lake turned into swamps. I'm curious though, what is that, maybe it's not a cultural shift, maybe it's a cultural roller coaster, like, mainframes are cool again. Give us your perspective on how you've helped companies like GE sort of as technology waves come really kind of help design and maybe drive a culture that embraces the velocity of this change. >> Sure, so one of the things we do a lot is help our customers better leverage technology, and really kind of train it. So, we have a couple different aspects to pivotal. One of them is our labs aspect, and effectively that is our ability to teach people how to better build applications, how to better do data science, how to better do data engineering. Now, when we come in, we have an opinionated way to do all those things, and when a customer embraces it it actually opens up a lot of doors. So we're somewhat technology agnostic, which aids in your question, right, so we can come in, we're not trying to push a specific technology, we're trying to push a methodology and an end goal and solution. And I think, you know, often times of course that end goal and solution is best met by our products, but to your point about the roller coaster, it seems as though as we have evolved there is a notion that data will, from an organization, will all come together in a common object store, and then the ability to quickly be able to spin up an analytical or a programmmatic interface within that data is super important and that's where we're kind of leaning, and that's where I think this idea of convenience being able to push button instantiate a green plum cluster, push button instantiate a gem fire grid so that you can do analytics or you can take actions on it is so super important. >> Male Speaker: You said something that sounds really important which is we want to get it sounded like you were alluding to a single source of truth, and then you spin up whatever compute, you bring it to the data. But there's an emerging, still early school of thought which is maybe the single source of truth should be a hub centered around real time streams. >> Male Speaker: Sure. Yeah. >> How does Pivotal play in that role? >> So, there are a lot of products that can help facilitate that including our own. I would say that there is a broad ecosystem that kind of says, if I was going to start an organization today there are a number of vertical products I would need in order to be successful with data. One of the would be just a standard relational database. And if I pause there for a second, if you look at it, there is definitely a move toward building microservices so that you can glue all those pieces together. Those microservices require smaller, simpler relational type databases, or you know, SQL type databases on the front end, but they become simpler and simpler where I think if I was Oracle or some of the more stalwart on the relational side, it's not about how many widgets you can put into the database, it's really about it's simplicity and performance. From there, having some kind of message queue or system to be able to take the changes and the updates of the data down the line so that, not so much IT providing it to an end user, but more self service, being able to subscribe to the data that I care about. And again, going back to the simplicity, me as an end user being able to take control of my destiny and use whatever product or technology makes the most sense to me and if I sort of dovetail on the side of that, we've focused so much this year on convenience and flexibility that I think it is now at a spot where all of the innovations that we're doing in the Amazon marketplace on green plum, all of those innovations are actually leading us to the same types of innovations in data deployments on top of Kubernetes. And so two of them that come to mind, I felt like, I was in front of a group last week and we were presenting some of the things we had done, and one of them was self-healing of green plum and so it's often been said that these big analytical solutions are really hard to operate and through our innovations we're able to have, if a segment goes down or a host goes down, or network problems, through the implementation the system will actually self heal itself, so all of a sudden the operational needs become quite a bit less. In addition, we've also created this automatic snapshotting capability which allows, I think our last benchmark we did about a pedobyte of data in less than three minutes, so suddenly you've got this operational stalwart, almost a database as a service without really being a service really just this living breathing thing. And that kind of dovetails back to where we're trying to make all of our products perform in a way that customers can just use them and not worry about the nuts and bolts of it. >> Female Speaker: So last question, we've got about 30 seconds left. You mentioned a lot of technologies but you mentioned methodology. Is that approach from Pivotal one of the defining competitive advantages that you deliver to the market? >> Male Speaker: It is 100 per cent one of our defining our defining things. Our methodology is what is enabling our customers to be successful and it actually allows me to say we've partnered with postcrestkampf and green plum summit this year is next month in April and the theme of that is hashtag data tells the story. And so, from our standpoint, green plum is continuing to take off, gem fire is continuing to take off, Kubernetes is continuing to take off, PCF is continuing to take off, but we believe that digital transformation doesn't happen without data. We think data tells a story. I'm here to encourage everyone to come to green plum summit, I'm also here to encourage everyone to share their stories with us on twitter, hashtag data tells a story, so that we can continue to broaden this ecosystem. >> Female Speaker: Hahtag data tells a story. Jacque, thanks so much for carving out some time this week to come back to the cube and share what's new and differentiating at Pivotal. >> Thank you. >> We want to thank you for watching The Cube. I'm Lisa Martin with my co-host George Gilbert. We are live at Big Data SV, our tenth big data event come down here, see us, we're in San Jose at Forrager eatery, we've got a great party tonight and also tomorrow morning at eight am we've got a breakfast briefing you wont' want to miss. Stick around, we'll be back with our next guest after a short break.

Published Date : Mar 7 2018

SUMMARY :

Brought to you by SiliconANGLE Media Welcome back to The Cube, we are live in San Jose and Bosch allows for not just the ease of installation, and having an enterprise solution to do that This is really a facilitator of a company's you know, whether it's our generation, But now that we're beginning to solve and again, inside of the data suite, we've got and the green plum DBMS are very complimentary. and the business analysts and the BI crowd of getting everybody to understand a culture that embraces the velocity of this change. and then the ability to quickly be able to Male Speaker: You said something that And that kind of dovetails back to where we're competitive advantages that you deliver to the market? and it actually allows me to say and share what's new and differentiating at Pivotal. we've got a breakfast briefing you wont' want to miss.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
George GilbertPERSON

0.99+

Lisa MartinPERSON

0.99+

BoschORGANIZATION

0.99+

GeorgePERSON

0.99+

twoQUANTITY

0.99+

GEORGANIZATION

0.99+

San JoseLOCATION

0.99+

fifty percentQUANTITY

0.99+

AmazonORGANIZATION

0.99+

JacquePERSON

0.99+

New YorkLOCATION

0.99+

Jacque IstokPERSON

0.99+

20 yearsQUANTITY

0.99+

last yearDATE

0.99+

PivotalORGANIZATION

0.99+

oneQUANTITY

0.99+

100 per centQUANTITY

0.99+

GoogleORGANIZATION

0.99+

last weekDATE

0.99+

Silicon ValleyLOCATION

0.99+

less than three minutesQUANTITY

0.99+

fiveQUANTITY

0.99+

The CubeTITLE

0.99+

HadukeORGANIZATION

0.99+

next monthDATE

0.99+

SiliconANGLE MediaORGANIZATION

0.99+

OracleORGANIZATION

0.99+

bothQUANTITY

0.99+

PKSORGANIZATION

0.98+

CubeORGANIZATION

0.98+

Strata Data ConferenceEVENT

0.98+

OneQUANTITY

0.97+

tonightDATE

0.97+

millions of dollarsQUANTITY

0.97+

this weekDATE

0.97+

tomorrow morning atDATE

0.97+

about a dozenQUANTITY

0.96+

AzureORGANIZATION

0.96+

earlier this weekDATE

0.96+

todayDATE

0.96+

SQLTITLE

0.96+

1500 distinct customersQUANTITY

0.96+

this yearDATE

0.95+

later this yearDATE

0.95+

last couple of yearsDATE

0.95+

Gartner ConferenceEVENT

0.95+

single sourceQUANTITY

0.94+

KubernetesORGANIZATION

0.94+

one productQUANTITY

0.93+

PramORGANIZATION

0.93+

10thQUANTITY

0.93+

PCFORGANIZATION

0.92+

tenthQUANTITY

0.92+

KubernetesTITLE

0.88+

about 30 secondsQUANTITY

0.86+

AprilDATE

0.86+

HahtagORGANIZATION

0.85+

about a pedobyte of dataQUANTITY

0.84+

2018DATE

0.84+

twitterORGANIZATION

0.84+

One of themQUANTITY

0.84+

gem fireORGANIZATION

0.83+

The CubeORGANIZATION

0.82+