Thierry Pellegrino, Dell EMC | Dell EMC: Get Ready For AI
[Music] and welcome back everybody Jeff Rick here at the cube we're in Austin Texas at the deli MC high performance computing and artificial intelligence labs last been here for a long time as you can see behind us and probably here racks and racks and racks of some of the biggest baddest computers on the planet in fact I think number 256 we were told earlier it's just behind us we're excited to be here really as Dell and EMC puts together you know pre-configured solutions for artificial intelligence machine learning deep learning applications because that's a growing growing concern and growing growing importance to all the business people out there so we're excited to have the guy running the show he's Terry Pellegrino the VP of HPC and business strategy had a whole bunch of stuff you're a pretty busy guy I'm busy but you can see all those servers they're very busy too they're humming so just your perspective so the HPC part of this has been around for a while the rise of kind of machine learning and artificial intelligence as a business priority is relatively recent but you guys are jumping in with both feet oh absolutely I mean HPC is not new to us AI machine learning deep learning is happening that's the buzzword but we've been working on HPC clusters since back in the 90s and it's it's great to see this technology or this best practice getting into the enterprise space where data scientists need help and instead of looking for a one processor that will solve it all they look for the knowledge of HPC and what we've been able to put together and applying into their field right so how do you kind of delineate between HPC and say the AI portion of the lab or is it just kind of on a on a continuum how do you kind of slice and dice absolutely it's it's all in one place and you see it all behind us this area in front of us we try to get all those those those servers put together and add the value for all the different workloads right so you get HPC a piece equal a IML deal all in one lab right and they're all here they're all here the old the legacy application only be called legacy applications all the way to the to the meanest and the newest and greatest exactly the old stuff the new stuff and and actually you know what some things you don't see is we're also looking at where the technology is going to take all those workloads AI m LD L is the buzzword today but down the road you're gonna see more applications and we're already starting to test those technologies in this lab so it's past present and future right so one of the specific solutions you guys have put together is the DL using the new Nvidia technology what if you could talk we hear about a media all the time obviously they're in really well position in autonomous vehicles and and their GPUs are taking data centers by storm how's that going where do you see some of the applications outside of autonomous vehicles for the the Nvidia base oh there are many applications I think the technology itself is is proving to solve a lot of customer problems and you can apply it in many different verticals many workloads again and you can see it in autonomous vehicles you can see it in healthcare live science in financial services risk management it's it's really everywhere you need to solve a problem and you need dense compute solutions and NVIDIA has one of technologies that a lot of our customers leverage to solve their problems right and you're also launching a machine learning solution based on Hadoop which we we've been going to Hadoop summit Hadoop world and strata for eight nine years I guess since 2010 eight years and you know it's kind of funny because the knock on Hadoop is always there aren't enough people it's too hard you know it's just a really difficult technology so you guys are really taken again a solutions approach with a dupe for machine learning to basically deliver either a whole rack full of stuff or that spec that you can build at your own place no absolutely that's one of the three major tenants that we have for those solutions that we're launching we really want it to be a solution that's faster so performance is key when you're trying to extract data and insights from from your data set you really need to be fast you don't want it to take months it has to be within countable measures so it's one of them we want to make it simple a data scientist is never going to be a PhD in HPC or any kind of computer technologies so making it simple it's critical and the last one is we want to have this proven trusted adviser feel for our customers you see it around you this HPC lab was not built yesterday it's been here showcasing our capabilities in HPC world our ability to combine the Hadoop environment with other environments to solve enterprise class problems and bring business value to our customers and that's really what we we think are our differentiation comes from right and it's really a lab I mean you and I are both wearing court coats right now but there's a gear stack following really heights of every shape and size and I think what's interesting is while we talk about the sexy stuff the GPUs and the CPUs and the do there's a lot of details that make one of these racks actually work and it's probably integrating some of those things as lower tier things and making sure they all work seamlessly together so you don't get some nasty bottleneck on an inexpensive part that's holding back all that capacity oh absolutely you know it's funny you mentioned that we're talking to customers about the technologies we're assembling and contrary to some web tech type companies that just look for any compute at all costs and they'll just stack up a lot of technologies because they want the compute in in HPC type environments or when you try to solve problems with deep learning and machine learning you're only as strong as your weakest link and if you have a a server or a storage unit or a an interconnect between all those that is really weak you're gonna see your performance go way down and we watch out for that and you know the one thing that you alluded to which I just wanted to point out what you see behind us is the hardware the the secret sauce is really in the aggregation of all the components and all the software stacks because AI M LDL great easy acronyms but when you start peeling the layers you realize it's layers and layers of software which are moving very fast where you don't want to be spending your life understanding the inter up requirements between those layers and and worrying about whether your your compute and your storage solution is gonna work right you want to solve problems a scientist and that's what we're trying to do give you a solution which is an infrastructure plus a stack that's been validated proven and you can really get to work right and even within that validated design for a particular workload customers have an opportunity maybe needs a little bit more IO as a relative scale these a little bit more storage needs a little bit more compute so even within a basic structured system that you guys have SPECT and certified still customers can come in and make little mods based on what their specific workload you've got it this is not we're not in the phase in the acceptance of a I am LDL where things are cookie cutter it's still going to be a collaboration that's what we have a really strong team working with our customers directly and trying to solution for their problem right if you need a little bit more storage if you need faster storage for your scratch if you need a little bit more i/o bandwidth because you're in a remote environment I mean all those characteristics are gonna be critical and the solutions we're launching are not rigid they're they're perfect starting point for customers I want to get something to run directly they feel like it but if you if you have a solution that's more pointed we can definitely iterate and that's what our team in the field and all the engineers that you have seen today walk through the lab that's what their role is we want to be as a consultant as a partner designing the right solution for the customer right so Terry before I let you guys just kind of one question from your perspective of customers and you're out talking to customers and how the conversation around artificial intelligence and machine learning has evolved over the last several years from you know kind of a cool science experiment or it's all the HPC stuff with the government or whether or heavy lifting really moving from that actually into a boardroom conversation as a priority and a strategic imperative going forward how's that conversation evolving when you're out talking to customers well you know it has changed you're right back in the 60s the science was there the technology wasn't there today we have the science we have the technology and we're seeing all the C Class decision makers really want to find value out of the data that we've collected and that that's where the discussion takes place this is not a CIO discussion most of the time and in what's really fantastic in mama contrary to a lot of the the technologies I have grown on like big data cloud and all those buzzwords here we're looking at something that's tangible we have real-life examples of companies that are using deep learning and machine learning to solve problems save lives and get our technology in the hands of the right folks so they can impact the community it's really really fantastic and that growth is set for success and we want to be part of that right it's just a minute just you know the continuation of this democratization trend you know get more people more data give more people more tools get more people more power and you're gonna get innovation you're gonna solve more problems and it's so exciting absolutely totally agree with you all right teri well thanks for taking a few minutes out of your busy day and congrats on the Innovation Lab here thank you so much all righty teri I'm Jeff Rick we're at the Dell EMC HPC and AI innovation labs in Austin Texas thanks for watching
**Summary and Sentiment Analysis are not been shown because of improper transcript**
ENTITIES
Entity | Category | Confidence |
---|---|---|
Terry Pellegrino | PERSON | 0.99+ |
Jeff Rick | PERSON | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
EMC | ORGANIZATION | 0.99+ |
NVIDIA | ORGANIZATION | 0.99+ |
Terry | PERSON | 0.99+ |
Austin Texas | LOCATION | 0.99+ |
Austin Texas | LOCATION | 0.99+ |
2010 | DATE | 0.99+ |
today | DATE | 0.99+ |
Nvidia | ORGANIZATION | 0.98+ |
yesterday | DATE | 0.98+ |
both feet | QUANTITY | 0.98+ |
one question | QUANTITY | 0.98+ |
Hadoop | TITLE | 0.97+ |
one | QUANTITY | 0.97+ |
Thierry Pellegrino | PERSON | 0.96+ |
eight years | QUANTITY | 0.96+ |
three major tenants | QUANTITY | 0.95+ |
Innovation Lab | ORGANIZATION | 0.94+ |
both | QUANTITY | 0.93+ |
Dell EMC | ORGANIZATION | 0.93+ |
eight nine years | QUANTITY | 0.92+ |
deli MC | ORGANIZATION | 0.92+ |
Dell EMC HPC | ORGANIZATION | 0.9+ |
one thing | QUANTITY | 0.88+ |
one processor | QUANTITY | 0.87+ |
HPC | ORGANIZATION | 0.87+ |
one place | QUANTITY | 0.84+ |
60s | DATE | 0.71+ |
SPECT | ORGANIZATION | 0.62+ |
years | DATE | 0.58+ |
few minutes | QUANTITY | 0.55+ |
90s | DATE | 0.52+ |
256 | OTHER | 0.51+ |
last | DATE | 0.49+ |
strata | LOCATION | 0.44+ |
Wikibon Research Meeting | Systems at the Edge
>> Hi I'm Peter Burris and welcome once again to Wikibons's weekly research meeting on theCUBE. (funky electronic music) This week we're going to discuss something that we actually believe is extremely important. And if you listen to the recent press announcements this week from Deli MC, the industry increasingly is starting to believe is important. And that is, how are we going to build systems that are dependent upon what happens at the edge? The past 10 years have been dominated about the cloud. How are we going to build things in the cloud? How are we going to get data to the cloud? How are we going to integrate things in the cloud? While all those questions remain very relevant, increasingly, the technology's becoming available, the systems and the design elements are becoming available, and the expertise is now more easily bought together so that we can start attacking some extremely complex problems at the edge. A great example of that is the popular notion of what's happening with automated driving. That is a clear example of huge design requirements at the edge. Now to understand these issues, we have to be able to generalize certain attributes of the differences in the resources, whether they be hardware or software, but increasingly, especially from a digital business transformation standpoint, the differences in the characteristics of the data. And that's what we're going to talk about this week. How do different types of data, data that's generated at the edge, data that's generated elsewhere, going to inform decisions about the classes of infrastructure that we're going to have to build and support as we move forward with this transformation that's taking place in the industry. So to kick it off, Neil Raden I want to turn to you. What are some of those key data differences and what taxonomically do we regard as what we call primary, secondary, and tertiary data? Neil. >> Well, primary data come in from sensors. It's a little bit different than anything we've ever seen in terms of doing analytics. Now I know that operational systems do pick up primary data, credit card transactions, something like that. But, scanner data, not scanner data, I mean sensor data is really designed for analysis. It's not designed for record keeping. And because it's designed for analysis, we have to have a different way of treating it than we do other things. If you think about a data lake, everything that falls into that data lake has come from somewhere else, it's been used for something else. But this data is fresh, and that requires that we really have to treat it carefully. Now, the retention and stewardship of that requires a lot of thought. And I don't think industry has really thought of that through a great deal. But look, sensor data is not new, it's been around for a long time. But what's different now is the volume and the lack of latency in it. But any organization that wants to get involved in it really needs to be thinking about what's the business purpose of it. If you're just going into, IOT as we call it generically, to save a few bucks you might as well not bother. It really is something that will change your organization. Now, what do we do with this data is a real problem because for the most part, these senses are going to be remote, and there's going to be a lot of, that means they're going to generate a lot of data. So what do we do with it? Do we reduce it at the sight? That's been one suggestion. There's an issue that any model for reduction could conceivably lose data that may be important somewhere down the line. Can the data be reconstituted through metadata or some sort of reverse algorithms? You know, perhaps. Those are the things we really need to think about. My humble opinion is the software and the devices need to be a single unit. And for the most part, they need to be designed by vendors, not by individual ITs. >> So David Floyer, let's pick up on that. Software and devices as single unit, designed more by vendors who have specific demand expertise, turn into solutions and present it to business. What do you think? >> Absolutely, I completely concur with that. The initial attempts to using the sensors and connecting to the sensors were very simple things like for example, the nest, the thermostats. And that's worked very well. But if you look at it over time, the processing for that has gone into the home, into your Apple TV device or your Alexa or whatever it is. So, that's coming down and now it's getting even closer to the edge. In the future, our proposition is that it will get even closer and then those will put together solutions, all types of solutions that are appropriate to the edge that will be taking not just one sensor but multiple sensors, collecting that data together, just like in the autonomous car for example where you take the lidars and the radars and the cameras etcetera. We'll be taking that data, we'll be analyzing it, and we'll be making decisions based on that data at the edge. And vendors are going to play a crucial role in providing these solutions to IT and to the OT and to many other parts. And a large value will be in their expertise that they will develop in this area. >> So as a rule of thumb, when I was growing up and learned to drive, I was told always keep five car lengths between you and whatever's in front of you at whatever speed you're traveling. What you just described David is that there will be sensors and there will be processing that takes place in that automated car that isn't using that type of rule of thumb but know something about tire temperature, and therefore the coefficient of friction on the tires, know something about the brakes, knows what the stopping power needs to be at the speed and therefore what buffer needs to be between it and whatever else is around it. >> Absolutely. >> This is no longer a rule of thumb, this is physics and deep understanding of what it's going to require to stop that car. >> And on top of that, what you'll also want to know, outside from your car is, what type of car is in front of you? Is that an autonomous car, or is that somebody being driven bye Peter? In which case, you have 10 lengths behind you. >> But that's not going to be primary data. Is that what we mean by secondary data? >> No, that's still primary because you're going to set up a connection between you and that other car. That car is going to tell you I'm primary to you, that's primary data. >> Here's what I mean, correct use primary data but, from a standpoint of that the car in that case is submitting a signal, right? So even though to your car it's primary data, but one of the things from a design standpoint that's interesting, is that car is now transmitting a digital signal about it's state that's relevant to you so that you can combine that >> Correct. inside effectively, a gateway inside your car. >> Yes. >> So there's external information that is in fact digital coming in, combining with the sensors about what's happening in your car. Have I got that right? >> Absolutely. That to me is a sort of sengrey one, then you've got the tertiary data which is the big picture about the traffic conditions >> Routes. and the weather and the routes and that sort of thing which is at that much higher cloud level, yes. So David Vellante, we always have to make sure as we have these conversations. We've talked a bit about this data, we've talked a little bit about the classes of work that's going to be performed at the different levels. How do we ensure that we sustain the business problem in this conversation? >> So, I mean I think Wikibon's done some really good work on describing what this sort of data model looks like from edge devices where you have primary data, the gateways where you're doing aggregated data in the cloud where maybe the serious modeling occurs. And my assertion would be is that the technology to support that elongating and increasingly distributed data model has been maturing for a decade and the real customer challenge is not just technical, it's really understanding a number of factors and I'll name some. Where in the distributed data value chain are you going to differentiate? And how does the data that you're capturing in that data pipeline contribute to monetization? What are the data sources, who has access to that data, how do you trust that data, and interpret it, and act on it with confidence? There are significant IP ownership in data protection issues. Who owns the data? Is it the device manufacturer, is it the factory, etcetera. What's the business model that's going to allow you to succeed? What skill sets are required to win? And really importantly, what's the shape of the ecosystem that needs to form to go to market and succeed? These are the things that I think customers are really struggling with that I talk to. >> Now, the one thing I'd add to that and I want to come back to it is the idea that, and who is ultimately bonding the solution because this is going to end up in a court of law. But let's come to this IP issue, George. Let's talk about how local data is going to be, is going to enter into the flow of analytics, and that question of who owns data, because that's important and then have the question about some of the ramifications and liabilities associated with this. >> Okay well, just on the IP protection and the idea that a vendor has to take sort of whole product responsibility for the solution. That vendor is probably going to be dealing with multiple competitors when they're sort of enabling say, self-driving car or other, you know edge, or smaller devices. The key thing is that, a vendor will say, you know, the customer keeps their data and the customer gets the insights from that data. But that data is informing in the middle a black box, an analytic black box. It's flowing through it, that's where the insights come out, on the other side. But the data changes that black box as it flows through it. So, that is something where, you know, when the vendor provides a whole solution to Mercedes, that solution will be better when they come around to BMW. And the customers should make sure that what BMW gets the benefit of, goes back to Mercedes. That's on the IP thing. I want to add one more thing on the tertiary side which is, when you're close to the edge, it's much more data intensive. When we've talked about the reduction in data and the real-time analytics, at the tertiary level it's going to be more where time is a bigger factor and you're essentially running a simulation, it's more compute intensive. And so you're doing optimizations of the model and those flow back as context to inform both the gateway and the edge. >> David Floyer I want to turn it to you. So we've talked a little bit about the characteristics of the data, great list of Dave Vellante about some of the business considerations, we will get very quickly in a second to some of the liability issues cause that's going to be important. But take us through how, which George just said about the tertiary elements. Now we've got all the data laid out, how is that going to map to the classes of devices? And we'll then talk a bit about some of the impacts on the industry. What's it going to look like? >> So if we take the primary edge first, and you take that as a unit, you'll have a number of senses within that. >> So just released, this is data about the real world that's coming into the system to be processed? >> Yes. So it'll have, for example, cameras. If we take a simple example of making sure that bad people don't get into your site. You'll have a camera there which will be facial recognition. They'll have a badge of some sort, so you'll read that badge, you may want to take their weight, you may want to have a infrared sensor on them so that you can tell their exact distance. So, a whole set of sensors that the vendor will put together for the job of insuring you don't get bad guys in there. And what you're insuring is that bad guys don't get in there, that's obviously one, very important, and also, that you don't go and- >> Stop good guys from going in. stop good guys from going in there. So those are the two characteristics >> The false-positive problem. the false-positives. Those are the two things you're trying to design that- >> At the primary edge. at the primary edge. And there's a mass amount of data going into that, which is only going to be reduced to very, very little data coming up to the next level which is this guy came here, this was his characteristics, he didn't look well today, maybe you should see a nurse, or whatever other information you can gather from that will go up to that secondary level, and then that'll also be a record of to HR maybe, about who has arrived there or what time they arrived, to the manufacturing systems about who is there and who has those skills to do a particular job. There are multiple uses of that data which can then be used for differentiation for whatever else from that secondary layer into local systems and then equally they can be pushed up to the higher level which is, how much power should be generating today, what are the higher levels. >> We now have 4,000 people in the building, air condition therefore is going to look like this, or, it could be combined with other types of data like over time we're going to need new capacity, or payroll, or whatever else it might be. >> And each level will have its own type of AI. So you've got AI at the edge, which is to produce a specific result, and then there's AI to optimize at the secondary level and then the AI optimize bigger things at the tertiary level. >> So we're going to talk more about some of the AI next week, but for right now we're talking about classes of devices that are high performance, high bandwidth, cheap, constrained, proximate to the event. >> Yep. >> Gateways that are capable of taking that information and start to synthesize it for the business, for other business types of things, and then tertiary systems, true private cloud for example, although we may have very sizable things at the gateway as well, >> There will be true private clouds. that are capable of integrating data in a more broad way. What's the impact in the industry? Are we going to see IT firms roll in and control this sweeping, (man chuckles) as Neil said, trillions of new devices. Is this all going to be intel? Is it all going to be, you know, looking like clients and PCs? >> My strong advice is, that the devices themselves will be done by extreme specialists in those areas that they will need a set of very deep technology understanding of the devices themselves, the senses themselves, the AI software relevant to that. Those are the people that are going to make money in that area. And you're much better off partnering with those people and letting them solve the problems, and you solve, as Dave said earlier, the ones that can differentiate you within your processes, within your business. So yes, leave that to other people is my strong advice. And from an IT's point of view, just don't do it yourself. >> Well the gateway's, sound like you're suggesting, the gateway is where that boundary's going to be. >> Yes. That's where the boundary is. >> And the IT technologies may increasingly go down to the edge, but it's not clear that the IT vendor expertise goes down to the edge >> Correct. at the same degree. >> Correct. >> So, Neil let's come back to you. When we think about this arrangement of data, you know, how the use cases are going to play out, and where the vendors are, we still have to address this fundamental challenge that Dave Vellante bought up. Who's going to end up being responsible for this? Now you've worked in insurance, what does that mean from an overall business standpoint? What kinds of failure weights are we going to accommodate? How is this going to play out? What do you think? >> Well, I'd like to point out that I worked in insurance 30 years ago. (men chuckling) >> Male Voice: I didn't want to date ya Neil. (men chuckling) >> Yeah the old reliable life insurance company. Anyway, one of the things David was just discussing sounded a lot to me like complex event processing. And I'm wondering where the logical location event needs to be, because it needs some prior data to do CEP, you have to have something to compare it against. But if you're pushing it all back to the tertiary level, there's going to be a lot of latency. And the whole idea was CEP was, you know, right now. So, that I'm a little curious about. But I'm sorry, what was your question? >> Well no, let's address that. So CEP David, I agree. But I don't want to turn this into a general discussion and CEP. It's got its own set of issues. >> It's clear there have got to be complex models created. And those are going to be created in a large environment, almost certainly in a tertiary type environment. And those are going to be created by the vendors of those particular problem solvers at the primary edge. To a large extent, they're going to provide solutions in that area. And they're going to have to update those. And so, they are going to have to have lots and lots of test data for themselves and maybe some companies will provide test data if it's convenient for those, for a fee or whatever it is, to those vendors. But the primary model itself is going to be in the tertiary level, and that's going to be pushed down to the primary level itself. >> I'm going to make an assertion here that the, the way I think about this Neil is that the data coming off at the primary level is going to be the sensor data, the sensor said it was good. Then that is recorded as an event, we let somebody in the building. And that's going to be a key feature of what happens at the secondary level. I think a lot of complex processing is likely to end up at that secondary level. >> Absolutely. >> Then the data gets pushed up to the tertiary level and it becomes part of an overall social understanding of the business, it's behavior data. So increasingly, what did we do as a consequence of letting this person in the building? Oh we tried to stop him. That's going to be more of the behavioral data that ends up at the tertiary level, will still do complex event processing there. It's going to be interesting to see whether or not we end up with CEP directly in the sensor tower. Might under certain circumstances, that's a cost question though. So let me now turn it in the last few minutes here Neil back to you. At the end of the day, we've seen for years the question of how much security is enough security? And businesses said, "Oh I want to be 100% secure." And sometimes see-so said "We got that. You gave me the money, we've now made you 100% secure." But we know it's not true. Same thing is going to exist here. How much fidelity is enough fidelity down at the edge? How do we ensure that business decisions can be translated into design decisions that lead to an appropriate and optimized overall approach to the way the system operates? From a business standpoint back, what types of conversations are going to take place in the boardroom that the rest of the organization's going to have to translate into design decisions? >> You know, boy, bad actors are going to be bad actors. I don't think you can do anything to eliminate it. The best you can do is use the best processes and the best techniques to keep it from happening and hope for the best. I'm sorry, that's all I can really say about it. >> There's quite a lot of work going on at the moment from Arm, in particular. They've got a security device image ability. So, there's a lot of work going on in that very space. It's obviously interesting from an IT perspective is how do you link the different security systems, both from an Arm point of view and then from a X86 as you go further up the chain. How are they going to be controlled and how's that going to be managed? That's going to be a big IT issue. >> Yeah, I think the transmission is the weak point. >> Male Voice: What do you mean by that Neil? >> Well the data has to flow across networks, that would be the easiest place for someone to intercept it and, you know, and do something nefarious. >> Right yeah, so that's purely in a security thing. I was trying to use that as an analogy. So, at the end of the day, the business is going to have to decide how much data do we have to capture off the edge to ensure that we have the kinds of models we want, so that we can realize the specificity of actions and behaviors that we want in our business? That's partly a technology question, partly a cost question. Different sensors are able to operate at different speeds for example. But ultimately, we have to be able to bring those, that list of decisions or business issues that Dave Vellante raised, down to some of the design questions. But it's not going to be throw a $300 micro processor everything. There's going to be very, very concrete decisions that have to take place. So, George do you agree with that? >> Yes, two issues though. One, there's the existing devices that can't get re-instrumented, that they already have their software, hardware stack. >> There's a legacy in place? >> Yes. But there's another thing which is, some of the most advanced research that's been going on that produced much of today's distributed computing and big data infrastructure, like the Berkeley Analytics lab, and say their contributions spark in related technologies. They're saying we have to throw everything out and start over for secure real-time systems. That you have to build from hardware all the way up. In other words, you're starting from the sand to re-think something that's secure and real-time that you can't layer it on. >> So very quickly David, that's a great point George. Building on what George has said very quickly, the primary responsibility for bonding the behavior or the attributes of these devices are going to be with the vendor. >> Of creating the solution? >> Correct. >> That's going to be the primary responsibility. But obviously from an IT point of view, you need to make sure that that device is doing the job that's important for your business, not too much, not too little, is doing that job, and that you are able to collect the necessary data from it that is going to be of value to you. So that's a question of qualification of the devices themselves. >> Alright so, David Vellante, Neil Raden, David Floyer, George Gilbert, action item round. I want one action item from you guys from this conversation. Keep it quick, keep it short, keep it to the point. David Floyer, what's your action item? >> So my action item is don't go into areas that you don't need to. You do not need to become experts, IT in general does not need to become experts at the edge itself. Rely on partners, rely on vendors to do that unless of course you're one of those vendors. In which case, you'll need very, very deep knowledge. >> Or you choose that that's where you're value stream your differentiations is going to be which means you just became one of those values. >> Yes, exactly. >> George Gilbert. >> I would build on that and I would say that if you look at the skills required to build these full stack solutions, there's data science, there's application development, there's the analytics. Very few of those solutions are going to have skills all in one company. So the go-to market model for building these is going to be something that, at least at this point in time, we're going to have to look to like combinations like IBM working with sort of supply chain masters. >> Good. Neil Raden, action item. >> The question is not necessarily one of technology because that's going to evolve. But I think as an organization, you need to look at it from this end which is, would employing this create a new business opportunity for us? Something we're not already doing. Or number two, change our operations in some significant way. Or number three, you know, the old red queen thing. We have to do it to keep up with the competition. >> Male Voice: David Vellante, action item. >> Okay well look, at the risk of sounding trite, you got to start the planning process from the customer on in, and so often people don't. You got to understand where you're going to add value for customers and constructing and external and internal ecosystem that can really juice that value creation. >> Alright, fantastic guys. So let me quickly summarize. This week on the Wikibon Friday research meeting in the cube, we discussed a new way of thinking about data characteristics that will inform system design and a business value that's created. We observe that data is not all the same when we think about these very complex, highly distributed, and decentralized systems that we're going to build. That there's a difference between primary data, secondary data, and tertiary data. Primary data is data that is generated from real world events or measurements and then turned into signals that can be acted upon very proximate to that real world set of conditions. A lot of sensors will be there, a lot of processing will be moved down there, and a lot of actuators and actions will take place without referencing other locations within the cloud. However, we will see circumstances where the events that are taken, or the decisions that are taken on those vents, will be captured in some sort of secondary tier that will then record something about the characteristics of the actions and events that were taken, and then summarized and then pushed up to a tertiary tier where that data can then be further integrated in other attributes and elements of the business. The technology to do this is broadly available but not universally successfully applied. We expect to see a lot of new combinations of edge-related device to work with primary data. That is going to be a combination of currently successful firms in the OT or operational technology world, most likely in partnership with a lot of other vendors that have demonstrated significant expertise and understanding the problems, especially the business problems, associated with the fidelity of what happens at the edge. The IT industry is going to approach very aggressively and very close to this at that secondary level, through gateways and other types of technologies. And even though we'll see IT technology continue to move down to the primary level, it's not clear exactly how vendors will be able to follow that. More likely, we'll see the adoption of IT approaches to doing things at the primary level by vendors that have the main expertise in how that level works. We will however see significantly interesting true private cloud and public cloud data end up from the tertiary level end up with a whole new sets of systems that are going to be very important from an administration and management standpoint because they have to work within the context of the fidelity of this overall system together. The final point we want to make is that these are not technology problems by themselves. While significant technology problems are on the horizon about how we think about handling this distribution of data, managing it appropriately, our ability, ultimately, to present the appropriate authority at different levels within that distributive fabric to ensure the proper working condition in a way that nonetheless we can recreate if we need to. But these are, at bottom, fundamentally business problems. They're business problems related to who owns the intellectual property that's being created, they're business problem related to what level in that stack do I want to show my differentiation to my customers and they're business problems from a liability and legal standpoint as well. The action item is, all firms will in one form or another be impacted by the emergence of the edge as a dominate design as consideration for their infrastructure but also for their business. Three ways, or a taxonomy that looks at three classes of data, primary, secondary, and tertiary, will help businesses sort out who's responsible, what partnerships I need to put in place, what technologies and I going to employ, and very importantly, what overall business exposure I'm going to accommodate as I think ultimately about the nature of the processing and business promises that I'm making to my marketplace. Once again, this has been the Wikibon Friday research meeting here on theCUBE. I want to thank all the analysts who were here today, but especially thank you for paying attention and working with us. And by all means, let's hear those comments back about how we're doing and what you think about this important question of different classes of data driven by different needs of the edge. (funky electronic music)
SUMMARY :
A great example of that is the popular notion And for the most part, they need to be designed present it to business. that are appropriate to the edge that will be taking and learned to drive, I was told of what it's going to require to stop that car. Is that an autonomous car, or is that But that's not going to be primary data. That car is going to tell you I'm primary inside your car. Have I got that right? the big picture about the traffic conditions and the weather and the routes What's the business model that's going to allow you to succeed? Now, the one thing I'd add to that the benefit of, goes back to Mercedes. of the liability issues cause that's going to be important. and you take that as a unit, and also, that you don't go and- So those are the two characteristics Those are the two things you're trying to design that- and then that'll also be a record of to HR maybe, air condition therefore is going to look like this, a specific result, and then there's AI to optimize high bandwidth, cheap, constrained, proximate to the event. Is it all going to be, you know, looking like clients and PCs? Those are the people that are going to make money in that area. Well the gateway's, sound like you're suggesting, at the same degree. How is this going to play out? Well, I'd like to point out that I worked in insurance Male Voice: I didn't want to date ya Neil. And the whole idea was CEP was, you know, right now. But I don't want to turn this into be in the tertiary level, and that's going to be And that's going to be a key feature of That's going to be more of the behavioral data and the best techniques to keep it from happening and how's that going to be managed? Well the data has to flow across networks, capture off the edge to ensure that we have can't get re-instrumented, that they already have their some of the most advanced research that's been going on are going to be with the vendor. the necessary data from it that is going to be of value to you. Keep it quick, keep it short, keep it to the point. IT in general does not need to Or you choose that that's where you're is going to be something that, at least at this point in time, Neil Raden, action item. We have to do it to keep up with the competition. You got to understand where you're going to add value sets of systems that are going to be very important
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave | PERSON | 0.99+ |
David Floyer | PERSON | 0.99+ |
Neil | PERSON | 0.99+ |
Neil Raden | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
David Vellante | PERSON | 0.99+ |
David | PERSON | 0.99+ |
George | PERSON | 0.99+ |
George Gilbert | PERSON | 0.99+ |
Peter Burris | PERSON | 0.99+ |
Mercedes | ORGANIZATION | 0.99+ |
BMW | ORGANIZATION | 0.99+ |
100% | QUANTITY | 0.99+ |
$300 | QUANTITY | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
10 lengths | QUANTITY | 0.99+ |
two characteristics | QUANTITY | 0.99+ |
Berkeley Analytics | ORGANIZATION | 0.99+ |
next week | DATE | 0.99+ |
4,000 people | QUANTITY | 0.99+ |
two issues | QUANTITY | 0.99+ |
Peter | PERSON | 0.99+ |
today | DATE | 0.99+ |
each level | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
one suggestion | QUANTITY | 0.98+ |
Three ways | QUANTITY | 0.98+ |
five car | QUANTITY | 0.98+ |
both | QUANTITY | 0.98+ |
This week | DATE | 0.97+ |
two things | QUANTITY | 0.97+ |
this week | DATE | 0.97+ |
30 years ago | DATE | 0.97+ |
one | QUANTITY | 0.97+ |
Wikibon | ORGANIZATION | 0.97+ |
Wikibons | ORGANIZATION | 0.97+ |
trillions of new devices | QUANTITY | 0.97+ |
single unit | QUANTITY | 0.97+ |
one sensor | QUANTITY | 0.96+ |
one form | QUANTITY | 0.96+ |
first | QUANTITY | 0.94+ |
one company | QUANTITY | 0.94+ |
Apple TV | COMMERCIAL_ITEM | 0.92+ |
one action item | QUANTITY | 0.92+ |
three classes | QUANTITY | 0.91+ |
intel | ORGANIZATION | 0.89+ |
Wikibon | EVENT | 0.86+ |
one more | QUANTITY | 0.79+ |
second | QUANTITY | 0.76+ |
past 10 years | DATE | 0.75+ |
CEP | ORGANIZATION | 0.75+ |
Deli MC | ORGANIZATION | 0.73+ |
CEP | TITLE | 0.68+ |
Arm | ORGANIZATION | 0.65+ |
Wikibon Friday | EVENT | 0.64+ |
Alexa | TITLE | 0.64+ |
years | QUANTITY | 0.62+ |
few bucks | QUANTITY | 0.6+ |