Image Title

Search Results for Javi:

Nevash Pillay & Javier Castellanos | UiPath FORWARD 5


 

The Cube presents UI Path Forward five. Brought to you by UI Path. >>We're back at forward five UI Paths, Big customer event. We're here in the Venetian, formerly the Sands Convention Center, Dave Ante and David Nicholson. Javier Castanos is here. He's the Robot Factory director. How's that for a title for Orange ESP Spania. And he's joined by Niva Pillow, who is Senior Director of Telecommunications Industry at UiPath. Folks, welcome to the Cube. Thank you. Thanks for coming on, Javier. Just off the keynote, it was really amazing to see what you were doing with your dashboard, how much you've operationalized automation, you really far down the journey. But I wanna start with your title. I've never seen this before. Robot Factory director, that's unique. What is that all about? >>Yeah, the Robot Factory is our brand to create the RPA journey to involve all the company in this amazing story regarding automation, because for us, automation is only a piece of the digital transformation and the culture transformation for the employees. >>Your robot factory obviously builds robots. Yeah. For employees and employees build them as well. >>Yeah, both. We have two different ways to, to build robots. We have a citizen developer program with more than 500 and employees certified in UiPath technology, and they build a small robot for the daily task for avoid repetitive task, very board. And in the other hand we have the robot factory team automating the business. The core business processes very complex in the telco industry, you know, and both teams working together, the community of employees, the best ambassadors for to find new opportunities and for discovery for robots and the robot factory are automating real complex processes to impacting our customer satisfaction. >>So if a, if a, if a citizen developer develops a robot, does the factory then have to audit it and make sure it's governed? Or do you add a, maybe I'm not such a good developer. Do you make it better? How does that collaboration work? >>The good thing is with you at Pat, you don't need to be a tech guy. You, you can be a finance guy and every morning you need a report, create an Excel, create a graph, put in a power point and send to your box. And you can create by your own a robot doing that task and going to the bending to take a coffee in, in the meantime that the robot is working. And as soon as you discover in your domain a complex tax, you can call us and say, Hey guys, I need your job because we need to ize this process. You need traceability. And we have a big savings below the desk. It's not only my health, it's the area work. >>Now, Navage, you specialize in the telecommunications industry. Now of course, the telcos are going through a massive transformation. It's almost, I call it revenge. The, the telcos now they're coming back with 5g. It's gonna be a great new future. But what kind of patterns are you seeing in the industry for automation? >>Sure. Look, as you said, telecoms going through quite a transformational era. There's this huge demand for connectivity around the whole world, and that presents opportunities and some challenges. But the key areas of focus right now is really helping the telecom achieve their strategic goals. And they include the customer experience at the most significant point, and thereafter driving a few more efficiencies and improving the employee experience. But organizations like Orange, you know, they start with the customer experience. These are large areas, but they tend to be the patterns where we are really helping telecoms transform and deliver better outcomes. >>Javi, I'm I'm curious about the concept of the citizen developer. Now you said that they don't have to have a deep technical background and they may come from finance or other places, but how do you, how do you recruit these people? What's in it for them? I, I can understand automating a process that is repetitive, mundane, something they don't want to do. But is there ever a concern that they might be automating themselves out of a job? >>Yeah, the, the people use Dex Excel and 30 years ago, Dex Excel does not assist and change our work. Your iPad technology is more or less the same. It's changing the way that you are working with your desktop every morning. You can create for your daily task a robot by yourself and executing your corporate desktop. And then you can save this time or use to improve your satisfaction as employee. Because sometimes in, in, in this kind of companies, we have a telecommunications engineering with a lot of talent making repetitive task. And with this technology, you can use your talent only to improve the processes. So we train these people in Miami, the training is very easy. A robot enter on the web searching, Google make different search regarding prices on, on device creates an Excel and only in a few hours that kind of people that we have in all companies that very easy excel some macros and these kind of things is the people prepared to jump to the next step to the robotization. So in all areas, in all departments, there are people prepared. In our company, 500 people. >>I, I'd like to get into a little mini case study if we could, and understand orange esp Spania is way deep. You should see this dashboard that Javier showed. I mean it's amazing, I think you said 7 million euro business benefit so far to date. But you can slice it and dice it and look at a lot of different angles. But where did you get started? Did you get started? Was it a bottoms up? In other words, an individual started to automate on their desktop. Was it a top down? The, the, the CEO said this is, we're gonna automate. How did it, I mean I'm sure you get this question a lot nivo, but where did it start at Orange? >>Yeah. Our story is very linked with the finance department because the citizen developer are saving internal hours and transforming the employee satisfaction and improving the talent and the reskilling of the people. But in the other hand, from the efficiency point of view, if you look for, for the finance approach, what happened, we, we take one profit and now domain perhaps 80% of the process. And next month the invoice reduce because your external cost disappear because the robot is making the task is improving the satisfaction of the customers. Because sometimes we have a, a human back office or another kind of task. And the compliance, the, the SLAs, the, the, the delay on time with all the people disappear with the robots because the robots are working at night. We can and repeating the job, 1, 1 1. And every tracking of that task are controlled by finance. Because if you save in a transaction three minutes, when you multiply for a thousand, a thousand, thousand tasks, you save on real time, you can see how much money you are saving and making the the things better. Not only a question of money is a question of money, but a attempt below that the customer is, is taking better experience for us. >>Robots don't sleep Nova. >>I never, >>So you started in finance and how much have you gone permeated other parts of the organization? What other parts of the organization are adopting RPA and automation? Where are you on that journey? >>More or less? Our eight, nine hundred and fifty three FTS equivalent robots working okay's like a contact center. It's robots navigating through the user interface applications, making transactions for our customers. So when you put in the middle of your customer relation, you can transform all because if a human agent is making a very complex process for, because telco is a complex market and very fast, perhaps the robot can help the human agent saving time and taking advantage of that part of, of the operations. And at the end, the operation is short and the customer satisfaction is better. And we measure the MPAs, the net, the net promoter score. And when you combine human agents with robots, the satisfaction improve because the transaction is made on real time very fast and doesn't fail. >>Is this a common story nivas that you're seeing in Telco in terms of the, the starting points? Does it tend to be bottoms up? Does it more top down? What are you seeing in >>Look, it actually varies by telecom. You know, Orange started their journey with us four years ago. So companies that have started while they tend to start in finance or IT or, or hr, but the customer experience I think is the ultimate area where many telecoms focus and what Harvey Edge just shared is it doesn't matter if a customer's calling you through a contact center or reaching you through a chatbot. They want their issue resolved at the first point. And what the robots do is they integrate information from multiple sources and provide that data to the agent so you can actually resolve the issue. And that is the beautiful example of humans and robots working together. Because if you know what the data's telling you, if it's a billing issue and a customer's been been billed because they have gone overseas and used international roaming and they weren't aware that the contract had that as a leader or a person in a contact center, you can make the right decision quite often. It takes a long time to find the data, but in this way you can actually address the issue real time, first point of resolution. And we're seeing up to 60% increase in first time resolutions across telecoms, irrespective of whether it's a chat bot or a contact center or a service desk. >>That's key. I mean, that's as a, that's consumer, that's what you just want to get off the phone or you want to get off the chat notice. So I have to ask you, what would you say is your secret to success? >>The secret is to be transparent with the organization, serve the savings and put on the table. We put on the table to the finance guys every month, all the robots that we put in production the month before and it's finance will declare officially the savings for each robot. As soon as you reach this, the credibility appear because it's not the robot factory team telling Aren, saving a lot of money of the company. No, no. It's the finance guys that trust on you. And as soon as you ask more money to buy more license or to improve the processes on whatever finance say, okay, these guys, as soon as we invest money in robots, we obtain twice or three times more by savings and they are improving not only for the quantity point of view, the quality is improving too. Because when you, a brief example, when you have a wifi problem connection and you call to our contact center, there is an ecosystem for more than 25 robots working from the beginning of your call, testing your line and making decisions. If we are going to send you a new router or you have a connectivity problem or, and the robot decide of, we are going to send to you a new install at your home and then the human manage you and take the conversation. But all the decisions are made by robots. So it's very powerful from the point of view of customer satisfaction. >>So what I'm hearing is you started four years ago. Yeah. And it, it, the ROI for your first instantiation was very fast, I presume inside of 12 months or what was the, how fast did you get a return with >>In the first three months we developed 25 robots and we saved more than 1 million to the company in three >>Months. In three months. Okay. So it was self-funding. >>Yeah. >>Right. You took that million dollars and you said, Okay, let's double down on that. Let's do it again. Do it again. Do it >>Again. It's only a question of resources and budget and only companies wants to create robots, but sometimes big companies only put on that one people to people. From the beginning of our story, we put 13 people and a budget. So if you have resources, the things happen be because the process are very accomplished. Sometimes you start one process. Sometimes our block, and we started at the beginning, a lot of process and imagine in telco we developed 900 processes, but every day we have a new opportunity for discovery. So I, I think the scalability is, is, is a challenge, but it's very, is possible if you put people and money >>And we, we focused on, we talk a lot in, in, in the broader IT world about the edge. And so I sort of think of these citizen developers as living at the edge. Part of your robot factory is at the core of the enterprise also. Is that, is that correct? Yes. >>Yes. >>Now what, what is, what has that looked like in terms of ROI cycles and development cycles? What kinds of projects do you work on at the core that are, that are different than what citizen soldiers are doing at the edge? >>Yeah. When, when we need to apply a discount or change your taif or switch on your bonus or your voicemail, that kind of transactions with impacting customers are made by the robot factory with robots made by the robot factory team. With a big traceability. With a big security because okay, with, with human awake the robot, we need to, to make a traceability because we have thousand of agents in the contact center working with robots and we have a lot of security disability and these kind of things. But in the other hand, internally we have a lot of task and a lot of processes for the citizen developers. There are very important tasks for the employee, perhaps not impacting in, in final customers, but we combine both. Because if you only work in one way, the citizen developer are making a lot of savings in terms of internal hours, but it's not real money. But in the other hand, you have the robot factory business processes impacting the money, combining both, you obtain the most powerful tool because the ambassadors, the, the, the employees are discovering you new opportunities. >>Last question, Javier, Why did you choose UiPath? What were the determining factors four years ago? >>Yeah, we, we were researching a lot in the market, but UiPath is pretty easy. You don't need to be an IT guy. People from, from customer care, people from finance in every areas. We have a lot of people learning this, this technology because it's easy, intuitive and very nice from the point of view of look and field. >>This a common story. This is really, we've reported on this a lot. This is how you UiPath really was able to get its foothold in the marketplace because of the simplicity. If you look at the legacy tools and even some of the modern tools, they were a lot more complicated. Now of course, UiPath is expanding its platform. So thank you very much. Don't welcome. Thank, thanks for coming. Thank you very much. Appreciate it. All right, you, you're gonna hear a lot of customer stories cuz that's what UI path brings in the cube. Proof is in the pudding. We right back at forward five from Las Vegas. Keep it right there.

Published Date : Sep 29 2022

SUMMARY :

Brought to you by UI Just off the keynote, it was really amazing to see what you were doing with Yeah, the Robot Factory is our brand to create the RPA journey to involve all Yeah. And in the other hand we have the robot factory team automating does the factory then have to audit it and make sure it's governed? And you can create by your own a robot doing that task and going to But what kind of patterns are you seeing in the industry for automation? But organizations like Orange, you know, Javi, I'm I'm curious about the concept of the citizen developer. It's changing the way that you are working with your desktop every morning. But you can slice it and dice it and look at a lot of different angles. But in the other hand, from the efficiency point So when you put in the middle of your customer but in this way you can actually address the issue real time, what would you say is your secret to success? We put on the table to the finance guys every So what I'm hearing is you started four years ago. You took that million dollars and you said, Okay, let's double down on that. So if you have resources, the things happen be because the at the edge. But in the other hand, you have the robot factory business processes You don't need to be an IT guy. If you look at the legacy tools and even some of the modern tools, they were a lot more complicated.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
JavierPERSON

0.99+

Javier CastanosPERSON

0.99+

OrangeORGANIZATION

0.99+

David NicholsonPERSON

0.99+

MiamiLOCATION

0.99+

Las VegasLOCATION

0.99+

80%QUANTITY

0.99+

twiceQUANTITY

0.99+

Dave AntePERSON

0.99+

Niva PillowPERSON

0.99+

Javier CastellanosPERSON

0.99+

13 peopleQUANTITY

0.99+

ExcelTITLE

0.99+

25 robotsQUANTITY

0.99+

GoogleORGANIZATION

0.99+

UiPathORGANIZATION

0.99+

iPadCOMMERCIAL_ITEM

0.99+

900 processesQUANTITY

0.99+

Robot FactoryORGANIZATION

0.99+

500 peopleQUANTITY

0.99+

more than 500QUANTITY

0.99+

threeQUANTITY

0.99+

one peopleQUANTITY

0.99+

bothQUANTITY

0.99+

million dollarsQUANTITY

0.99+

three monthsQUANTITY

0.99+

7 million euroQUANTITY

0.99+

first instantiationQUANTITY

0.99+

Nevash PillayPERSON

0.99+

more than 25 robotsQUANTITY

0.99+

more than 1 millionQUANTITY

0.98+

one wayQUANTITY

0.98+

three minutesQUANTITY

0.98+

JaviPERSON

0.98+

TelcoORGANIZATION

0.98+

next monthDATE

0.98+

each robotQUANTITY

0.98+

four years agoDATE

0.98+

one processQUANTITY

0.98+

telcoORGANIZATION

0.98+

three timesQUANTITY

0.98+

VenetianLOCATION

0.98+

four years agoDATE

0.98+

up to 60%QUANTITY

0.97+

both teamsQUANTITY

0.97+

first pointQUANTITY

0.97+

excelTITLE

0.97+

first three monthsQUANTITY

0.97+

four years agoDATE

0.97+

NavageORGANIZATION

0.96+

first timeQUANTITY

0.96+

fiveQUANTITY

0.95+

two different waysQUANTITY

0.95+

UI PathTITLE

0.94+

PathTITLE

0.94+

one profitQUANTITY

0.93+

thousand of agentsQUANTITY

0.92+

Dex ExcelTITLE

0.91+

30 years agoDATE

0.91+

ArenORGANIZATION

0.9+

SpaniaLOCATION

0.88+

PatORGANIZATION

0.86+

a thousandQUANTITY

0.84+

thousand tasksQUANTITY

0.83+

UIORGANIZATION

0.8+

nine hundred and fifty three FTSQUANTITY

0.77+

CubeORGANIZATION

0.75+

Sands Convention CenterLOCATION

0.75+

eight,QUANTITY

0.74+

fiveTITLE

0.73+

12 monthsQUANTITY

0.73+

Harvey EdgePERSON

0.71+

UI PathsTITLE

0.69+

UITITLE

0.6+

UiPathTITLE

0.54+

LastQUANTITY

0.5+

5QUANTITY

0.49+

5gORGANIZATION

0.34+

Javier de la Torre, Carto | AWS Startup Showcase S2 E2


 

(upbeat music) >> Hello, and welcome to theCUBE's presentation of the a AWS startup showcase, data as code is the theme. This is season two episode two of the ongoing series covering the exciting startups from the AWS ecosystem and we talk about data analytics. I'm your old John Furrier with the cube, and we have Javier De La Torre. who's the founder and chief strategy officer of Carto, which is doing some amazing innovation around geographic information systems or GIS. Javier welcome to the cube for this showcase. >> Thank you. Thank you for having me. >> So, you know, one of the things that you guys are bringing to the table is spatial analytic data that now moves into spatial relations, which is, you know, we know about geofencing. You're seeing more data coming from satellites, ground stations, you name it. Things are coming into the market from a data perspective, that's across the board and geo's one of them GIS systems. This is what you guys are doing in the rise of SQL in particular with spatial. This is a huge new benefit to the world. Can you take a minute to explain what Carto's doing and what spatial SQL is? >> Sure. Yeah. So like you said, like data, obviously we know is growing very fast and as you know now, being leveraged by many organizations in many different ways. There's one part of data, one dimension that is location. We like to say that everything happens somewhere. So therefore everything can be analyzed and understood based on the location. So we like to put an example, if all your neighbors get an alarm in their homes, the likelihood that you will get an alarm increases, right? So that's obvious we are all affected by our surroundings. What is spatial analytics, this type of analytics does is try to uncover those spacial relations so that you can model, you can predict where something is going to happen, or, you know, like, or optimize it, you know, like where else you want it to happen, right? So that's at the core of it. Now, this is something that as an industry has been done for many years, like the GIS or geographic information systems have existed for a long time. But now, and this is what Carto really brings to the table. We're looking at really the marketizing it, so that it's in the hands of any analyst, our vision is that you need to go five years, to a geography school to be able to do this type of spatial analysis. And the way that we want to make that happen is what we call with the rise of a spatial SQL. We add these capabilities around spatial analytics based on the language that is very, very popular for a analysts, which is SQL. So what we do is enables you to do this spatial analysis on top of the well known and well used SQL methods. >> It's interesting the cloud native and the cloud scale wave and now data as code has shown that the old school, the old guard, the old way of doing things, you mentioned data warehousing, okay, as one. BI tools in particular have always been limited. And the scope of the limitation was the environment was different. You have to have domain expertise, rich knowledge of the syntax. Usually it's for an application developer, not for like real time and building it into the CICD pipeline, or just from a workflow standpoint, making it available. The so-called democratization, this is where this connects. And so I got to ask you, what are you most excited about in the innovations at Carto? Can you share some of the things that people might know about or might not know about that's happening at Carto, that takes advantage of this cloud native wave because companies are now on this bandwagon. >> Yeah, no, it is. And cloud native analytics is probably the most disruptive kind of like trend that we've seen over the few years, in our particular space on the spatial it has tremendous effects on the way that we provide our service. So I'd like to kind of highlight four main reasons why cloud analytics, cloud native is super important to us. So the first one is obviously is a scalability, the working with the sizes of data that we work now in terms of location was just not possible or before. So for someone that is performing now analysis on autonomous car, or you're like that has any sensorized GPS on a device and is collecting hundreds of billions of points. If you want to do analysis on that type of data, cloud native allows you to do that in a scalable way, but it also is very cost effective. That is something that you'll see very quickly when your data grows a lot, which is that this computing storage separation, the idea that is store your data at cloud prices, but then use them with these data warehouses that we work in this private, makes for a very, very cost effective solution. But then, you know, there is other two, obviously one of them being SQL and spatial SQL that like means we like to say that SQL is becoming the lingua franca for analytics. So it's used by many products that you can connect through the usage of SQL, but I think like you coming towards why I think it's even more interesting it's like, in the cloud the concept like we all are serving, we are all living in the same infrastructure enables us that we can distribute a spatial data sets to a customer that they can join it on their database on SQL without having to move the data from one another, like in the case of Redshift or Amazon Redshift car connects and you using something called a spectrum, we can connect live to data that is stored on S3. And I think that is going to disrupt a lot the way that we think about data distributions and how cost effective it is. I think, it has a lot of your like potential on it. And in that sense what Carto is providing on top of it in the format of formats like parquet, which is a very popular with big data format. We adding geo parquet, we are specializing this big data technology for doing the spatial analysis. And that to me it is very exciting because it's putting some of the best tools at the hands of doing the space analytics for something that we're not able to do before. So to me, this is one area that I'm very, very excited. >> Well, I want to back up for a second. So you mentioned parquet and the standards around that format. And also you mentioned Redshift, so let me get this right. So you just saying that you can connect into Redshift. So I'm a customer and I have Redshift I'm using, I got my S3, I'm using Redshift for analysis. You're saying you can plug right into Redshift. >> Yes. And this is a very, very, very important part because what Carto does is leverage Redshift computing infrastructure to essentially kind of like do all the analysis. So what we do is we bring a spatial analysis where the data is, where Redshift is versus in the past, what we will do is take the data where the analysis was and that sense, it's at the core of cloud native. >> Okay. This is really where I see the exciting shift where data as code now becomes a reality is that you bring the... It redefines architecture, the script is flipped. The architecture has been redefined. You're making the data move to the environments that needs to move when it has to, if it doesn't have to move you bring compute to it. So you're seeing new kinds of use cases. So I have to ask you on the use cases and examples for Carto AWS customers with spatial analytics, what are some of the examples on how your clients are using cloud native spatial analytics or Carto? >> Yeah. So one, for example, that we've seen a lot, on the AWS ecosystem, obviously because of its suites and its position. We work together with another service in the AWS ecosystem called Amazon Location. So that actually provides you access to maps and SDKs for navigation. So it means that you are like a company that is delivering food or any other goods in the city. We have like hundreds or thousands of drivers around the city moving, doing all these deliveries. And each of these drivers they have an app and they're collecting actively their location, their position, right? So you get all the data and then it gets stored on something like a Redshift data cluster on S3 as well. There's different architectures in there, but now you essentially have like a full log of the activity that is happening on the ground from your business. So what Carto does on top of that data is you connect your data into Carto. And now you can do analysis, for example, for finding out where you user may be placed, another distribution center, you know, for optimizing your delivering routes, or like if you're in the restaurant business where you might want to have a new dark kitchen, right? So all this type of analysis based on, since I know where you're doing your operations, I can post analyze the data and then provide you a different way that you can think about solving your operation. So that's an example of a great use case that we're seeing right now. >> Talk to me about about the traditional BI tools out there, because you mentioned earlier, they lack the specific capabilities. You guys bring that to the table. What about the scalability limitations? Can you talk about where that is? Is there limitations there, obviously, if they don't have the capabilities, you can't scale that's one, but you know, as you start plugging into Redshift, scale and performance matters, what's the issue there? Can you unpack that a little bit real quick? >> Yeah. It goes back to the particulars of the spacial data, location data, like in the use case, like I was describing you very quickly are going to end up with really a lot of your like terabytes, if not petabytes of data very quickly, if you're start aggregating all this data, because it gets created by sensors. So volumes in our world kind of tends to grow a lot now. So when you work with BI tools, there's two things that you have to take in consideration. BI tools are great for seeing things like for example, if all you want to see is where your customers are, a BI tool is great. Seeing, creating a map and seeing your customers. That's totally in the world of BI. But if you want to understand why your customers are there, or where else could they be, you're going to need to perform what we call a spatial analysis. You're going to have to create a spatial model. You're going to have to, and for that BI tools will not give you that that's one side, the other it talks about the volumes that I was describing. Most of these BI tools can handle certain aggregations. Like, for example, if you are reading, if you're connecting your, let's say 10 billion data set to a BI tool, the BI tool will do some aggregations because you cannot display 10,000 rows on a BI tool and that's okay, you get aggregations and that works. But when it comes to a map, you cannot aggregate the data on the map. You actually want to see all the data on the map, and that's what Carto provides you. It allows you to make maps that sees all the data, not just aggregated by county or aggregated by other kind of like area, you see all your data on the map. >> You know, what's interesting is that location based service has been around for a long time. You know, when mobile started even hitting the scene, you saw it get better mashups, Google Maps, all this Google API mashups, things like that. You know, developers are used to it, but they could never get to the promised land on the big data side, because they just didn't have the compute. But now you add in geofencing, geo information, you now have access to this new edge like data, right? So I have to ask you on the mobile side, are you guys working with any 5G or edge providers? Because I can almost imagine that the spatial equation gets more complicated and more data full when you start blowing out edge data, like with 5G, you got more, more things happening at the edge. It's only going to fill in more data points. Can you share that's how that use case is going with mobile, mobile carriers or 5G? >> Yeah, that's totally, yeah. It's totally the case. Well, first, even before, you know, like we are there, we actually helping a lot of telcos on actually planning the 5G deployment. Where do you place your antennas is a very, very important topic when you're like talking about 5G. Because you know, like 5G networks require a lot of density. So it's a lot about like, okay, where do I start deploying my infrastructure to ensure the customers like meet, like have the best service and the places where I want to kind of like go first So like... >> You mean like the RF maps, like understanding how RF propagates. >> Well, that's one signal, but the other is like, imagine that your telco is more interested on, you know, let's say on a certain kind of like consumer profile, like young people that are using the one type of service. Well, we know where these demographics kind of lives. So you might want to start kind of like deploying your 5G in those areas, right. Versus if you go to more commercial and more kind of like residential areas, there might be other demographics. So that's one part around market analysis. Then the second part is once these 5G networks are in place, you're right. I mean, one of the premises that kind of like these news technologies give us is because the network is much smarter. You can have all these edge cases, there's much more location data that can be collected. So what we see now is a rise on the amount of what we call telemetry. That for example, the IOT space can make around location. And that's now enabled because of 5G. So I think 5G is going to be one of those trends that are going to make like more and more data coming into, I mean, more location, data available for analysis. >> So how does that, I mean, this is a great conversation because everyone can realize they're at a stadium and they see multiple bars but they can't get bandwidth. So they got a back haul problem or not enough signal. Everyone knows when they're driving their car, they know they can relate to the consumer side of it. So I get how the spatial data grows. What's the impact to Carto and specifically the cloud, because if you have more data coming in, you need the actionable insight. So I can see the use case, oh, put the antenna here. That's an actionable business decision, more content, more revenue, more happy customers, but where else is the impact to you guys and the spatial piece of it? >> Yeah. Well, I mean like there's many, many factors, right? So one of them, for example, on the telco, one of the things where we realize impact is that it gives the visibility to the operator, for example, around the quality of service. Like, okay, are my customers getting the quality of services where I want? Or like you said, like if there sitting outside a concert the quality of service in one particular area is dropping very fast. So the idea of like being able to now in real time, kind of like detect location issues, like I'm having an issue in this place. That means that then now I can act, I can drive up bandwidth, put more capacity et cetera right. So I think the biggest impact that we are seeing we are going to see on the upcoming years is that like more and more use cases going towards real time. So where, like before it was like, well, now that it has happened, I'm going to analyze it. I'm going to look at, you know, like how I could do better next time towards a more of like an industry where Carto ourselves, we are embedded in more real time type of, you know, like analytics where it's okay, if this happens, then do that, right. So it's going to be more personalized at the level that like in the code environment, it has to be art of a full kind of like pipeline kind of like type of analysis. That's already programmatically prepared to act on real time. >> That's great and it's a good segue. My next question, as more and more companies adopt cloud native analytics, what trends are you seeing out of the key to watch? Obviously you're seeing more developers coming on site, on the scene, open sources growing, what's the big cloud native analytics trends for Carto and geographic information. >> Yeah. So I think you know like the, we were talking before the cloud native now is unstoppable, but one of the things that we are seeing that is still needs to be developed and we are seeing progress is around a standardization, for example, around like data sets that are provided by different providers. What I mean with that is like, you as an organization, you're going to be responsible for your data like that you create on your cloud, right. On S3, or, you know and then you going to have a competing engine, like Redshift and you're going to have all that set up, but then you also going to have to think about like, okay, how do I ingest data from third party providers that are important for my analysis? So for example, Carto provides a lot of demographics, human mobility. we aggregate and clean up and prepare lot of spacial data so that we can then enrich your business. So for us, how we deliver that into your cloud native solution is a very important factor. And we haven't seen yet enough standardization around that. And that's one of the things, what we are pushing, you know, with the concept of geo Parquet of standardizing that body. That's one, then there is another, this is more what I like to say that you know, we are helping companies figure out their own geographies. What we mean by that is like most companies, when they start thinking about like how they interact, on the space, on the location, some of them will work like by zip codes and other by cities, they organize their operations based on a geography in a way, or technically what we call a geographic support system. Well, nowadays, like the most advance companies are defining their geographies in a continuous spectrum in what we call global grid system or spatial indexes that allows them to understand the business, not just as a set of regions, but as a continuous space. And that is now possible because of the technologies that we are introducing around spatial indexes at the cloud native infrastructure. And it provides a great a way to match data with resources and operate at scale. To me those two trends are going to be like very, very important because of the capabilities that cloud native brings to our spatial industry. >> So it changes the operation. So it's data as ops, data as code, is data ops, like infrastructures code means cloud DevOps. So I got to ask you because that's cool. Spatial index is a whole another way to think of it, rather than you go hyper local, super local, you get local zones for AWS and regions. Things are getting down to the granular levels I see that. So I have to ask you, what does data as code mean to you and what does it mean to Carto? Because you're kind of teasing at this new way because it's redefining the operation, the data operations, data engineering. So data as code is real. What does that mean to you? >> No, I think we already seeing it happening to me and to Carto what I will describe data as code is when an organization has moved from doing an analysis after the fact, like where they're like post kind of like analysis in a way to where they're actually kind of like putting analytics on their operational cycle. So then they need to really code it. They need to make these analysis, put them and insert them into the architecture bus, if you want to say of the organization. So if I get a customer, happens to be in this location, I'm going to trigger that and then this is going to do that. Or if this happens, I'm need to open up. And this is where if an organization is going to react in more real time, and we know that organizations need to drive in that direction, the only way that they can make that happen is if they operationalize analytics on their daily operations. And that can only happen with data as code. >> Yeah. And that's interesting. Look at ML ops, AI ops, people talk about that. This is data, so developers meets operations, that's the cloud, data meets code that's operations, that's data business. >> You got it. And add to that, the spacial with Carto and we go it. >> Yeah, because every piece of data now is important. And the spatial's key real quick before we close out, what is the index thing? Explain the benefit real quick of a spatial index. >> Yes. So the spatial index is well everybody can understand how we organize societies politically, right? Our countries, you have like states and then you have like counties and you have all these different kind, what we call administrative boundaries, right? That's a way that we organize information too, right? A spatial index is when you divide the world, not in administrative boundaries, but you actually make a grid. Imagine that you just essentially make a grid of the world. right? And you make that grid so that in every cell you can then split it into, let's say for example, four more cells. So you now have like an organization. You split the world in a grid that you can have multiple resolutions think like Google maps when you see the entire world, but you can zoom in and you end up seeing, you know, like one particular place, so that's one thing. So what a spatial indexes allows you is to technically put, you know like your location, not based coordinate, but actually on one grid place on an index. And we use that then later to correlate, let's say your data with someone else data, as we can use what we call this spatial indexes to do joints very, very fast and we can do a lot of operations with it. So it is a new way to do spatial computing based on this type of indexes, but for more than anything for an organization, what spatial index allows is that you don't need to work on zip codes or in boundaries on artificial boundaries. I mean, your customer doesn't change because he goes from this place to the road, to the other side of the road, this is the same place. It's an arbitrary in location. It's a spatial index break out all of that. You're like you break with your zip codes, you break. And you essentially have a continuous geography, that actually is a much closer look up to the reality. >> It's like the forest and the trees and the bark of the tree. (Javier laughing) You can see everything. >> That's it, you can get a look at everything. >> Javi, great to have you on. In real quick closing give a quick plug for the company, summarize what you do, what you're looking into, how many people you got, when you're hiring, what's the key goals for the company? >> Yeah, sure. So Carto is a company, now we are around 200 people. Our vision is that spatial analytics is something that every organization should do. So we really try to enable organizations with the best data and analysis around spatial. And we do all that cloud native on top of your data warehouse. So what we are really in enabling these organizations is to take that cloud native approach that they're already embracing it also to spatial analysis. >> Javi, founder, chief strategy officer for Carto. Great to have you on data as code, all data's real, all data has impact, operational impact with data is the new big trend. Thanks for coming on and sharing the company story and all your key innovations. Thank you. >> Thanks to you. >> Okay. This is the startup showcase. Data as code, season two episode two of the ongoing series. Every episode will explore new topics and new exciting companies pioneering this next cloud native wave of innovation. I'm John Furrier, your host of theCUBE. Thanks for watching. (upbeat music)

Published Date : Apr 26 2022

SUMMARY :

data as code is the theme. Thank you for having me. one of the things that you guys the likelihood that you will shown that the old school, products that you can connect So you just saying that you like do all the analysis. So I have to ask you on the use cases So it means that you are like a company You guys bring that to the table. So when you work with BI tools, So I have to ask you on the mobile side, and the places where I want You mean like the RF maps, on the amount of what we call telemetry. So I can see the use case, I'm going to look at, you know, out of the key to watch? that you create on your cloud, right. So I got to ask you because that's cool. and to Carto what I will operations, that's the cloud, And add to that, the spacial And the spatial's key real is to technically put, you and the bark of the tree. That's it, you can Javi, great to have you on. is to take that cloud native approach Great to have you on data and new exciting companies pioneering

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Javier De La TorrePERSON

0.99+

JaviPERSON

0.99+

John FurrierPERSON

0.99+

CartoORGANIZATION

0.99+

10,000 rowsQUANTITY

0.99+

hundredsQUANTITY

0.99+

JavierPERSON

0.99+

five yearsQUANTITY

0.99+

AWSORGANIZATION

0.99+

one partQUANTITY

0.99+

RedshiftTITLE

0.99+

SQLTITLE

0.99+

second partQUANTITY

0.99+

oneQUANTITY

0.99+

eachQUANTITY

0.99+

AmazonORGANIZATION

0.99+

two thingsQUANTITY

0.99+

twoQUANTITY

0.98+

one sideQUANTITY

0.98+

firstQUANTITY

0.98+

Google mapsTITLE

0.97+

5GORGANIZATION

0.97+

around 200 peopleQUANTITY

0.97+

one dimensionQUANTITY

0.97+

Google MapsTITLE

0.97+

one signalQUANTITY

0.97+

CartoPERSON

0.96+

two trendsQUANTITY

0.96+

telcoORGANIZATION

0.95+

one areaQUANTITY

0.95+

Javier de la Torre, CartoPERSON

0.94+

four more cellsQUANTITY

0.92+

10 billion dataQUANTITY

0.92+

first oneQUANTITY

0.91+

hundreds ofQUANTITY

0.9+

one thingQUANTITY

0.9+

S3TITLE

0.89+

parquetTITLE

0.84+

CartoTITLE

0.83+

season twoQUANTITY

0.82+

petabytesQUANTITY

0.77+

billions of pointsQUANTITY

0.76+

RedshiftORGANIZATION

0.76+

one gridQUANTITY

0.75+

episodeQUANTITY

0.75+

Brett Rudenstein - Hadoop Summit 2014 - theCUBE - #HadoopSummit


 

the cube and hadoop summit 2014 is brought to you by anchor sponsor Hortonworks we do have do and headline sponsor when disco we make hadoop invincible okay welcome back and when we're here at the dupe summit live is looking valance the cube our flagship program we go out to the events expect a signal from noise i'm john per year but Jeff Rick drilling down on the topics we're here with wind disco welcome welcome Brett room Stein about senior director tell us what's going on for you guys I'll see you at big presence here so all the guys last night you guys have a great great booth so causing and the crew what's happening yeah I mean the show is going is going very well what's really interesting is we have a lot of very very technical individuals approaching us they're asking us you know some of the tougher more technical in-depth questions about how our consensus algorithm is able to do all this distributor replication which is really great because there's a little bit of disbelief and then of course we get to do the demonstration for them and then suspend disbelief if you will and and I think the the attendance has been great for our brief and okay I always get that you always we always have the geek conversations you guys are a very technical company Jeff and I always comment certainly de volada and Jeff Kelly that you know when disco doesn't has has their share pair of geeks and that dudes who know they're talking about so I'm sure you get that but now them in the business side you talk to customers I want to get into more the outcome that seems to be the show focused this year is a dupe of serious what are some of the outcomes then your customers are talking about when they get you guys in there what are their business issues what are they tore what are they working on to solve yeah I mean I think the first thing is to look at you know why they're looking at us and then and then with the particular business issues that we solve and the first thing and sort of the trend that we're starting to see is the prospects and the customers that we have are looking at us because of the data that they have and its data that matters so it's important data and that's when people start to come to is that's when they look to us as they have data that's very important to them in some cases if you saw some of the UCI stuff you see that the data is you know doing live monitoring of various you know patient activity where it's not just about about about a life and monitoring a life but potentially about saving the life and systems that go down not only can't save lives but they can potentially lose them so you have a demos you want to jump into this demo here what is this all about you know the demo that the demonstration that I'm going to do for you today is I want to show you our non-stop a new product i'm going to show you how we can basically stand up a single HDFS or a single Hadoop cluster across multiple data centers and I think that's one of the tough things that people are really having trouble getting their heads wrapped around because most people when they do multi data center Hadoop they tend to do two different clusters and then synchronize the data between the two of them the way they do that is they'll use you know flume or they'll use some form of parallel ingest they'll use technologies like dis CP to copy data between the data centers and each one of those has sort of an administrative burden on them and then some various flaws in their and their underlying architecture that don't allow them to do a really really detailed job as ensuring that all blocks are replicated properly that no mistakes are ever made and again there's the administrative burden you know somebody who always has to have eyes in the system we alleviate all those things so I think the first thing I want to start off with we had somebody come to our booth and we were talking about this consensus algorithm that we that we perform and the way we synchronize multiple name nodes across multiple geographies and and again and that sort of spirit of disbelief I said you know one of the key tenants of our application is it doesn't underlie it doesn't change the behavior of the application when you go from land scope to win scope and so I said for example if you create a file in one data center and 3,000 miles apart or 7,000 miles apart from that you were to hit the same create file operation you would expect that the right thing happens what somebody gets the file created and somebody gets file already exists even if at 7,000 miles distance they both hit this button at the exact same time I'm going to do a very quick demonstration of that for you here I'm going to put a file into HDFS the my top right-hand window is in Northern Virginia and then 3,000 miles distance from that my bottom right-hand window is in Oregon I'm going to put the etsy hosts file into a temp directory in Hadoop at the exact same time 3,000 miles distance apart and you'll see that exact behavior so I've just launched them both and again if you look at the top window the file is created if you look at the bottom window it says file already exists it's exactly what you'd expect a land scope up a landscape application and the way you'd expect it to behave so that is how we are ensure consistency and that was the question that the prospect has at that distance even the speed of light takes a little time right so what are some of the tips and tricks you can share this that enable you guys to do this well one of the things that we're doing is where our consensus algorithm is a majority quorum based algorithm it's based off of a well-known consensus algorithm called paxos we have a number of significant enhancements innovations beyond that dynamic memberships you know automatic scale and things of that nature but in this particular case every transaction that goes into our system gets a global sequence number and what we're able to do is ensure that those sequence numbers are executed in the correct order so you can't create you know you can't put a delete before a create you know everything has to happen in the order that it actually happened occurred in regardless of the UN distance between data centers so what is the biggest aha moment you get from customer you show them the demo is it is that the replication is availability what is the big big feature focus that they jump on yeah I think I think the biggest ones are basically when we start crashing nodes well we're running jobs we separate the the link between the win and maybe maybe I'll just do that for you now so let's maybe kick into the demonstration here what I have here is a single HDFS cluster it is spanning two geographic territory so it's one cluster in Northern Virginia part of it and the other part is in Oregon I'm going to drill down into the graphing application here and inside you see all of the name notes so you see I have three name nodes running in Virginia three name nodes running in Oregon and the demonstration is as follows I'm going to I'm going to run Terrigen and Terra sort so in other words i'm going to create some data in the cluster I'm then going to go to sort it into a total order and then I'm going to run Tara validate in the alternate data center and prove that all the blocks replicated from one side to the other however along the way I'm going to create some failures I am going to kill some of that active name nodes during this replication process i am going to shut down the when link between the two data centers during the replication paris's and then show you how we heal from from those kinds of conditions because our algorithm treats failure is a first class citizen so there's really no way to deal in the system if you will so let's start unplug John I'm active the local fails so let's go ahead and run the Terrigen in the terrorists or I'm going to put it in the directory called cube one so we're creating about 400 megabytes of data so a fairly small set that we're going to replicate between the two data centers now the first thing that you see over here on the right-hand side is that all of these name nodes kind of sprung to life that is because in an active active configuration with multiple name nodes clients actually load balance their requests across all of them also it's a synchronous namespace so any change that I make to one immediately Curzon immediately occurs on all of them the next thing you might notice in the graphing application is these blue lines over and only in the Oregon data center the blue lines essentially represent what we call a foreign block a block that is not yet made its way across the wide area network from the site of ingest now we move these blocks asynchronously from the site of in jeff's oh that I have land speed performance in fact you can see I just finished the Terrigen part of the application all at the same time pushing data across the wide area network as fast as possible now as we start to get into the next phase of the application here which is going to run terrace sort i'm going to start creating some failures in the environment so the first thing I'm going to do is want to pick two named nodes I'm going to fail a local named node and then we're also going to fail a remote name node so let's pick one of these i'm going to pick HD p 2 is the name of the machine so want to do ssh hd2 and i'm just going to reboot that machine so as I hit the reboot button the next time the graphing application updates what you'll notice here in the monitor is that a flat line so it's no longer taking any data in but if you're watching the application on the right hand side there's no interruption of the service the application is going to continue to run and you'd expect that to happen maybe in land scope cluster but remember this is a single cluster a twin scope with 3,000 miles between the two of them so I've killed one of the six active named nodes the next thing I'm going to do is kill one of the name nodes over in the Oregon data center so I'm going to go ahead and ssh into i don't know let's pick the let's pick the bottom one HTTP nine in this case and then again another reboot operation so I've just rebooted two of the six name nose while running the job but if again if you look in the upper right-hand corner the job running in Oregon kajabi running in North Virginia continues without any interruption and see we just went from 84 to eighty eight percent MapReduce and so forth so again uninterruptedly like to call continuous availability at when distances you are playing that what does continuous availability and wins because that's really important drill down on yeah I mean I think if you look at the difference between what people traditionally call high availability that means that generally speaking the system is there there is a very short time that the system will be unavailable and then it will then we come available again a continuously available system ensures that regardless of the failures that happen around it the system is always up and running something is able to take the request and in a leaderless system like ours where no one single node actually it actually creates a leadership role we're able to continue replication we're and we're also able to continue the coordinator that's two distinct is high availability which everyone kind of know was in loves expensive and then continues availability which is a little bit kind of a the Sun or cousin I guess you know saying can you put in context and cost implementation you know from a from a from a from a perspective of a when disco deployment it's kind of a continuously available system even though people look at us as somewhat traditional disaster recovery because we are replicating data to another data center but remember it's active active that means both data centers are able to write at the same time you have you get to maximize your cluster resources and again if we go back to one of the first questions you asked what are what a customer's doing this with this what a prospects want to do they want to maximize their resource investment if they have half a million dollars sitting in another data center that only is able to perform an emergency recovery situation that means they either have to a scale the primary data center or be what they want to do is utilize existing resource in an active active configuration which is why i say continuous availability they're able to do that in both data centers maximizing all their resource so you versus the consequences of not having that would be the consequences of not being able to do that is you have a one-way synchronization a disaster occurs you then have to bring that data center online you have to make sure that all the appropriate resources are there you have to you have an administrative burden that means a lot of people have to go into action very quickly with the win disco systems right what that would look like I mean with time effort cost and you have any kind of order of magnitude spec like a gay week called some guy upside dude get in the office login you have to look at individual customer service level agreements a number that i hear thrown out very very often is about 16 hours we can be back online within 16 hours really RTO 44 when disco deployment is essentially zero because both sites are active you're able to essentially continue without without any doubt some would say some would say that's contingent availability is high available because essentially zero 16 that's 16 hours I mean any any time down bad but 16 hours is huge yeah that's the service of level agreement then everyone says but we know we can do it in five hours the other of course the other part of that is of course ensuring that once a year somebody runs through the emergency configure / it you know procedure to know that they truly can be back up in line in the service level agreement timeframe so again there's a tremendous amount of effort that goes into the ongoing administrating some great comments here on our crowd chatter out chat dot net / hadoop summit joined the conversation i'll see ya we have one says nice he's talking about how the system has latency a demo is pretty cool the map was excellent excellent visual dave vellante just weighed in and said he did a survey with Jeff Kelly said large portion twenty-seven percent of respondents said lack of enterprises great availability was the biggest barriers to adoption is this what you're referring to yeah this is this is exactly what we're seeing you know people are not able to meet the uptime requirements and therefore applications stay in proof-of-concept mode or those that make it out of proof of concept are heavily burdened by administrators and a large team to ensure that same level of uptime that can be handled without error through software configuration like Linda scope so another comment from Burt thanks Burt for watching there's availability how about security yeah so security is a good one of course we are you know we run on standard dupe distributions and as such you know if you want to run your cluster with on wire encryption that's okay if you want to run your cluster with kerberos authentication that's fine we we fully support those environments got a new use case for crowd chapel in the questions got more more coming in so send them in we're watching the crowd chat slep net / hadoop summit great questions and a lot of people aren't i think people have a hard time partial eh eh versus continues availability because you can get confused between the two is it semantics or is it infrastructure concerns what is what is the how do you differentiate between those two definitions me not I think you know part of it is semantics but but but also from a win disco perspective we like to differentiate because there really isn't that that moment of downtime there is there really isn't that switch over moment where something has to fail over and then go somewhere else that's why I use that word continuous availability the system is able to simply continue operating by clients load balancing their requests to available nodes in a similar fashion when you have multiple data centers as I do here I'm able to continue operations simply by running the jobs in the alternate data center remember that it's active active so any data ingest on one side immediately transfers to the other so maybe let me do the the next part I showed you one failure scenario you've seen all the nodes have actually come back online and self healed the next part of this I want to do an separation I want to run it again so let me kick up kick that off when I would create another directory structure here only this time I'm going to actually chop the the network link between the two data centers and then after I do that I'm going to show you some some of our new products in the works give you a demonstration of that as well well that's far enough Britain what are some of the applications that that this enables people to use the do for that they were afraid to before well I think it allows you know when we look at our you know our customer base and our prospects who are evaluating our technologies it opens up all the all the regulated industries you know things like pharmaceutical companies financial services companies healthcare companies all these people who have strict regulations auditing requirements and now have a very clear concise way to not only prove that they're replicating data that data has actually made its way it can prove that it's in both locations that it's not just in both locations that it's the correct data sometimes we see in the cases of like dis CP copying files between data centers where the file isn't actually copied because it thinks it's the same but there is a slight difference between the two when the cluster diverges like that it's days of administration hour depending on the size of the cluster to actually to put the cluster you know to figure out what went wrong what went different and then of course you have to involve multiple users to figure out which one of the two files that you have is the correct one to keep so let me go ahead and stop the van link here of course with LuAnn disco technology there's nothing to keep track of you simply allow the system to do HDFS replication because it is essentially native HDFS so I've stopped the tunnel between the two datacenters while running this job one of the things that you're going to see on the left-hand size it looks like all the notes no longer respond of course that's just I have no visibility to those nodes there's no longer replicating any data because the the tunnel between the two has been shut down but if you look on the right hand side of the application the upper right-hand window of course you see that the MapReduce job is still running it's unaffected and what's interesting is once I start replicating the data again or once i should say once i start the tunnel up again between the two data centers i'll immediately start replicating data this is at the block level so again when we look at other copy technologies they are doing things of the file level so if you had a large file and it was 10 gigabytes in size and for some reason you know your your file crash but in that in that time you and you were seventy percent through your starting that whole transfer again because we're doing block replication if you had seventy percent of your box that had already gone through like perhaps what I've done here when i start the tunnel backup which i'm going to do now what's going to happen of course is we just continue from those blocks that simply haven't made their way across the net so i've started the tunnel back up the monitor you'll see springs back to life all the name nodes will have to resync that they've been out of sync for some period of time they'll learn any transactions that they missed they'll be they'll heal themselves into the cluster and we immediately start replicating blocks and then to kind of show you the bi-directional nature of this I'm going to run Tara validate in the opposite data center over in Oregon and I'll just do it on that first directory that we created and in what you'll see is that we now wind up with foreign blocks in both sides I'm running applications at the same time across datacenters fully active active configuration in a single Hadoop cluster okay so the question is on that one what is the net net summarized that demo reel quick bottom line in two sentences is that important bottom line is if name notes fail if the wind fails you are still continuously operational okay so we have questions from the commentary here from the crowd chat does this eliminate the need for backup and what is actually transferring certainly not petabytes of data ? I mean you somewhat have to transfer what what's important so if it's important for you to I suppose if it was important for you to transfer a petabyte of data then you would need the bandwidth that support I transfer of a petabyte of data but we are to a lot of Hollywood studios we were at OpenStack summit that was a big concern a lot of people are moving to the cloud for you know for workflow and for optimization Star Wars guys were telling us off the record that no the new film is in remote locations they set up data centers basically in the desert and they got actually provisioned infrastructure so huge issues yeah absolutely so what we're replicating of course is HDFS in this particular case I'm replicating all the data in this fairly small cluster between the two sites or in this case this demo is only between two sites I could add a third site and then a failure between any two would actually still allow complete you know complete availability of all the other sites that still participate in the algorithm Brent great to have you on I want to get the perspective from you in the trenches out in customers what's going on and win disco tell us what the culture there what's going on the company what's it like to work there what's the guys like I mean we we know some of the dudes there cause we always drink some vodka with him because you know likes to tip back a little bit once in a while but like great guy great geeks but like what's what's it like it when disco I think the first you know you touched on a little piece of it at first is there are a lot of smart people at windows go in fact I know when I first came on board I was like wow I'm probably the most unsmoked person at this company but culturally this is a great group of guys they like to work very hard but equally they like to play very hard and as you said you know I've been out with cause several times myself these are all great guys to be out with the culture is great it's a it's a great place to work and you know so you know people who are who are interested should certainly yeah great culture and it fits in we were talking last night very social crowd here you know something with a Hortonworks guide so javi medicate fortress ada just saw him walk up ibm's here people are really sociable this event is really has a camaraderie feel to it but yet it's serious business and you didn't the days they're all a bunch of geeks building in industry and now it's got everyone's attention Cisco's here in Intel's here IBM's here I mean what's your take on the big guys coming in I mean I think the big guys realize that that Hadoop is is is the elephant is as large as it appears elephant is in the room and exciting and it's and everybody wants a little piece of it as well they should want a piece of it Brett thanks for coming on the cube really appreciate when discs are you guys a great great company we love to have them your support thanks for supporting the cube we appreciate it we right back after this short break with our next guest thank you

Published Date : Jun 4 2014

**Summary and Sentiment Analysis are not been shown because of improper transcript**

ENTITIES

EntityCategoryConfidence
two sitesQUANTITY

0.99+

Jeff KellyPERSON

0.99+

seventy percentQUANTITY

0.99+

OregonLOCATION

0.99+

two sitesQUANTITY

0.99+

Jeff KellyPERSON

0.99+

3,000 milesQUANTITY

0.99+

VirginiaLOCATION

0.99+

Jeff RickPERSON

0.99+

BurtPERSON

0.99+

84QUANTITY

0.99+

Northern VirginiaLOCATION

0.99+

North VirginiaLOCATION

0.99+

twoQUANTITY

0.99+

five hoursQUANTITY

0.99+

3,000 milesQUANTITY

0.99+

7,000 milesQUANTITY

0.99+

two data centersQUANTITY

0.99+

BrettPERSON

0.99+

Star WarsTITLE

0.99+

10 gigabytesQUANTITY

0.99+

half a million dollarsQUANTITY

0.99+

16 hoursQUANTITY

0.99+

Brett RudensteinPERSON

0.99+

JeffPERSON

0.99+

both locationsQUANTITY

0.99+

two sentencesQUANTITY

0.99+

two filesQUANTITY

0.99+

IBMORGANIZATION

0.99+

two datacentersQUANTITY

0.99+

two data centersQUANTITY

0.99+

oneQUANTITY

0.99+

two different clustersQUANTITY

0.99+

both sidesQUANTITY

0.99+

both sitesQUANTITY

0.99+

first directoryQUANTITY

0.98+

third siteQUANTITY

0.98+

first thingQUANTITY

0.98+

firstQUANTITY

0.98+

CiscoORGANIZATION

0.98+

twenty-seven percentQUANTITY

0.98+

JohnPERSON

0.98+

first thingQUANTITY

0.98+

one sideQUANTITY

0.97+

BritainLOCATION

0.97+

todayDATE

0.97+

two definitionsQUANTITY

0.97+

OpenStackEVENT

0.96+

HortonworksORGANIZATION

0.96+

eighty eight percentQUANTITY

0.96+

last nightDATE

0.96+

both data centersQUANTITY

0.94+

each oneQUANTITY

0.94+

zeroQUANTITY

0.94+

once a yearQUANTITY

0.94+

one failureQUANTITY

0.93+

the cube and hadoop summit 2014EVENT

0.93+

two geographic territoryQUANTITY

0.93+

IntelORGANIZATION

0.92+

bothQUANTITY

0.92+

singleQUANTITY

0.92+

this yearDATE

0.91+

one data centerQUANTITY

0.91+

dupe summitEVENT

0.9+

Brett room SteinPERSON

0.9+