Casimir Wierzynski, Intel | RSAC USA 2020
>>Fly from San Francisco. It's the cube covering RSA conference, 2020 San Francisco brought to you by Silicon angle media. >>Hello and welcome back to the cube coverage here in San Francisco, the Moscone center for RSA Congress 2020 for all the coverage period for three days. I'm John, host of the cube. You know, as cybersecurity goes to the next level, as cloud computing goes, continues to go more enterprise, large scale AI and machine learning have become critical managing the data. We've got a great guest here from Intel, Kaz Borzynski, senior director of the AI price with Intel. Thanks for joining us. Oh thanks. So data is a huge, huge data problem when it comes down to cybersecurity, uh, and and generally across the enterprise. Now it's well known, well documented, but you're here giving a talk about machine learning privacy because everyone wants to know who the bad guys are. So do the bad guys deserve privacy? Okay, we'll get to that later. But first tell about your talk and give a talk here at RSA. >>We'll get into other stuff later. I gave a talk, so thanks for having me. I gave a talk on a whole suite of exciting new techniques known as privacy preserving machine learning. So this is a set of machine learning techniques that help people realize the promise of AI and machine learning. But we know that machine learning systems rely on underlying data to train. So how can you also respect the privacy and the security of the underlying data while still being able to train and use AI systems and just take it, where are you within the Intel sphere? Because Intel osseous surgery obviously chips and power to all the enterprises in large. Skip. How are you on the software side AI group? Explain where you are. And so I'm in the AI group at Intel, but I have the most fun job at Intel. I think so cause I work in the CTO office of the AI group, which means I get to think about more futuristic, you know, where is AI going? >>What are some of the major inflection points? One of these that we've been looking at for the last couple of years is this kind of collision course between the need for data to train machine learning systems to unlock all the power of AI, but still the need to keep data private. Yeah, and I think that's generally consistent with our editorial in our research, which is the confluence of cloud native, large scale cloud computing, multi-cloud and AI or machine learning, all kinds of coming together. Those are multigenerational technologies that are coming. So that's, this wave is big. That's right. And I think one thing that's kind of maybe underappreciated about machine learning, especially in production is it's almost always a multi-party interaction. So you'll have, let's say one party that owns data and other party may own a model. They're running a system on somebody else's hardware. So because of the nature of digital data, if you want to share things, you have to worry about what other parties may be doing with those data. >>Because you bring up a great point I want to get your reaction and thoughts on is that, is that it's multidisciplinary. Now as people aren't breaking into the field. I mean people are really excited about AI. I mean you talk to someone who's 12 years old, they see a Tesla, they see software, they see all these things, they see all this cool stuff. So machine learning, which powers AI is very enticing to anyone that's got kind of technical or nerdy background and social attracting a lot of young people. So it's not just getting a computer science degree. There's so much more to AI because you talk about why, what someone needs to be successful too. And to engage in the AI wave. You don't need to just be a code or you could be outside the scope because it's an integrated model or is it's very much, so my group at Intel is better, very heterogeneous. >>So what have got a, you know, kind of mathematicians, but I also have coders. I have, uh, an attorney who's a public policy expert. I have cryptographers. Uh, I think there's a number of ways to get involved in, in meaning my, my background is actually a neuroscience. So, um, it makes sense. Good. Stitch it all together. Yeah. Well, societal changes has to be the, the algorithm needs training they need to learn. So having the most diverse input seems to me to be a, a posture the industry is taking and what's, is that right? Is that the right way to think about it? How should we be thinking about how to make AI highly effective versus super scary? Right. Well, one of the efforts that we're making, part of my message here is that to make these systems better, generally more data helps, right? If you can expand the availability of data, that's always going to help machine learning systems. >>And so we're trying to unlock data silos that may exist across countries, across the organizations. So for example, you know, in healthcare you could have multiple hospitals that have patient data. If somehow they could pool all their data together, you would get much more effective models, much better patient outcomes, but for very good privacy reasons, they're not allowed to do that. So there's these interesting ideas like federated learning where you could somehow decentralize the machine learning process so that you can still respect privacy but get the statistical power. That's a double down on that for a second cause I want to explore that. I think this is the most important story that's not being talked about. It's nuance a little bit. Yeah. You know, healthcare, you had HIPAA, which was built for all the right reasons back then, but now when you start to get into much more of a cross pollination of data, you need to manage the benefit of why it existed with privacy. >>So encryption, homomorphic encryption for instance, data and use. Yes. Okay. When it's being used, not just in flight or being arrested becomes, now you have the three triads of data. Yes. This is now causing a new formula for encryption privacy. What are some of the state of the art mindset thinkings around how to make data open a usable but yet either secure, encrypted or protected. That's right. So it's kind of this paradox of how do I use the data but not actually get the data. You mentioned homomorphic encryption. So this is one of the most kind of leading edge techniques in this area where somehow you're able to, there are ways of doing math on the data while it stays encrypted and the answer that comes out, it's still encrypted and it's only the actual owner of the data who can reveal the answer. So it seems like magic, but with this capability you enable all kinds of new use cases that wouldn't be possible before where third parties can act on, you know, your sensitive data without ever being exposed to it in any way. >>So discovery and leverage of the days that what you're getting at in terms of the benefits, I mean use cases. So stay on that. They used cases of the, of this new idea. Yeah. Is discovery and usage. How would that work? Well, so when we talked about federated learning and pooling across hospitals, that's one set of techniques. Homomorphic encryption would be, for example, suppose that some AI system has already been trained, but I'd like to use it on sensitive data. How do I do that in such a way that the third party service isn't, you know, this what makes, I think machine learning different from different types of data. You know, security problems is that machine learning, you have to operate on the data. You're not just storing it, you're not just moving it around. So how do you, yeah, and this is a key thing. >>So I've got to ask you the question because one of the things that's a real interesting trade off these days is AI and machine learning is really can create great benefits, but also people just go the knee jerk reaction of, you know, Oh my God, it's scary. My privacy. So it's a frontline with Amazon, just facial recognition. Oh my God, it's evil. Yeah. So there's a lot of scared people that might not be informed. Yeah. How should companies invest in machine learning and AI from your opinion? On how should they think about the next 10 year trajectory starting today, thinking about how to invest, what's the right way to think about it, build a team. Yeah. What's your thoughts on that? Because, and this is the number one challenge right now. Yeah. Well I think the, uh, some of this scary issues that you mentioned, you know, there are legitimately scary. >>They're going to have to be resolved, not by companies, but probably, you know, by society and kind of our delegates. So lawmakers, regulators, part of what we're trying to do at the technical level is give society and regulators a, a more flexible set of tools around which you can slice and dice data privacy and so on, so that it's not just all or none. Right. I think that's kind of my main goal as a, as an organization. I think again, the, this idea of having a heterogeneous set of talents, you know, you're going to need policy experts and applied mathematicians and linguists and you know, neuroscientists. So diversity is a huge opportunity, very much so. Not just diversity of people, but diverse data, diverse data, diverse kind of mindsets, approaches to problems that are hard but very promising. If so. Okay. Let's flip to the other side of the spectrum, which is what should people not do? >>What does, what's a, what's a fail failure formula one dimensional thinking? What's a, what's an identification of something that's not, may not go in the right way? Well, you know, one, uh, distinguishing feature of the machine learning field, and it's kind of a cultural thing, but it's given it a lot of traction is it's fundamentally, it had been a very open culture. So there's a lot of, uh, sharing of methods. It's a very, uh, collaborative academic field. So I think within a company you want to kind of be re you want to be part of that culture too. So every company is going to have its secret sauce. It's things that it needs to keep proprietary, but it's very important for companies to engage this broader community of researchers. So you're saying, which I would want, maybe I'm what I would agree with, but I'll just say it. >>You can agree or disagree to be successful, you got to be open. If you're data-driven, you've gotta be open. That's right. There's more JD equals better data. That's why more data, more approaches to data, kind of more eyes on the problem. But you know, still you can definitely keep your proprietary, you know, it kind of forces organizations to think about what are our core strengths that we really want to keep proprietary. But then other things let's, you know, open. All right. So what's the coolest thing you've working on right now? What are some of the fun projects you guys are digging into and you've got a great job. Sounds like you're excited about it. I mean, AI I think is the most exciting thing. I mean I wish I could be 20 again in computer science or whatever field. Cause I think AI is more than a multigenerational things. >>Super exciting as a technical person. But what are you working on that you're excited about? So I'm very excited about taking some of these things like homomorphic encryption and making them much more available to developers, to data scientists because it's asking too much for a data scientist to also be a kind of a post quantum crypto expert. So we've written an open source package called H E transformer, H G for homomorphic encryption. It allows the data scientists to kind of do their normal data science and Python or whatever they're used to, but then they kind of flick a switch and suddenly their model is able to run on encrypted data. Can you just take a minute to explain why homomorphic encryption trend right now is really important? I mean, give a peek into the why because this is something that is now becoming much more real. >>Yeah. The data in use kind of philosophy. Why now? Why is it so important right now? Well, I think, uh, the, because of cloud in the, the power of cloud and the fact that you know, data are collected in one place and possibly processed in another place, you're going to have to, you know, your data are moving around and they're being operated on. So if you can know that, you know, as long as my data are moving around and people are operating on it but it's staying encrypted the whole time, you know, not just in transit, that gives a much higher level of comfort around and the applications are going to probably be onboarded. I mean you can almost imagine new applications will emerge from this application discovery cataloging and API integration points. I mean you can almost imagine the trust will go up and you can also kind of end up with these different business models where you have entities that compete in some spheres but they may decide to collaborate in other ways. >>So for example, banks could compete on, you know, lending and so on under normal activities. But in terms of fraud detection, they may decide, Hey, maybe we can make some Alliance where we cross check with each other as models on certain transactions, but I'm not actually giving you any transaction data. So that's maybe okay. Right. So that's a very powerful, it's really interesting. I mean I think the uh, the compute power has allowed, the overhead seems to be much more robust because people are working on this for in the eighties and nineties I remember. Yes. But it was just so expensive overhead while that's right. Yeah. So you bring up a great point here. So, and this is one of the areas where Intel is really pushing, my team is pushing these techniques have been around for 20 years. Initially they were maybe like 10 million times slower than real time. >>So people thought, okay, this is interesting, you know, mathematically, but not practical. There've been massive improvements just in the last two years where now things are running, you know, a hundred times slower than, than kind of un-encrypted math. But still that, that means that something that you know would take 50 milliseconds now takes five seconds. That's still not an unreasonable, you're my new friend. Now, my best friend on AI. Um, and I got a business to run and I'm going to ask you, what should I do? I really want to leverage machine learning and AI in my business. Okay, I'm investing in more tech. I got cloud and building my own software. How should I be investing? How do I build out a great machine learning AI scene and then ultimately capabilities? How should I do that? Okay, well I would start with a team that has a kind of a research mindset, not because you want them to come in and like write research papers, but the path from research into production is so incredibly short in AI. >>You know, you have things that are papers one year and they're going into production at Google search and within a year. So you kind of need that research mindset. I think another thing is that you want to, uh, you're gonna, you're going to require a very close collaboration between this data science team and your CIO and kind of, you know, systems. And a lot of the challenges around AI are not just coming up with the model, but how do you actually scale it up and you know, go to production with it and interesting about the research. I totally agree with you. I think, you know, you can almost call that product management kind of new fangled Prague product management because if it's applied research, you kind of have your eye on a market generally, but you're not making hardcore product decisions. You're researching it, you're writing it so that you got to, got to do the homework, you know, dream it before you can build it. >>Well, I'm just saying that the field is moving so fast that you're going to need on your team, uh, people who can kind of consume the latest papers. Oh, you're saying consume the research as well. Yeah, I mean if they can contribute, that's great too. I mean, I think this is this kind of open culture where, you know, people consume, they find some improvement. They can then publish it at the next year's conference. It's just been this incredibly healthy eco software. Acceleration's a big part of the cloud. Awesome. Well I really appreciate your insight. This is great topic. I could go for an hour. One of my favorite things. I love the homophobic uh, encryption. I think that's going to be a game changer. I think we're going to start to see some interesting discoveries there. Uh, give a quick plug for Intel. What are you working on now? >>What are you looking to do? What's your plans, highs hiring, doing more research, what's going on? Well, so we think that this intersection of privacy and AI is kind of at the core of, of Intel's data centric mission. So we're trying to figure out, you know, whatever it takes to enable the community, whether it's, you know, uh, optimize software libraries. It could be custom Silicon, it could be even services where, you know, we really want to listen to customers, figure out what they need. Funding. Moore's law is always going to be around the next wave is going to have more compute. It's never going away. More storage, more data. It's just gets better and better. Yeah. Thanks for coming on Catherine. Thanks for having can we have Intel inside the cube breaking down the future of AI. Really exciting stuff on the technology front security day. That's all going to happen at large scale. Of course, it's the cube bringing you all the data here at RSA. I'm John furrier. Thanks for watching.
SUMMARY :
RSA conference, 2020 San Francisco brought to you by Silicon So do the bad guys deserve privacy? So how can you also respect So because of the nature of digital data, I mean you talk to someone who's 12 years old, they see a Tesla, they see software, So what have got a, you know, kind of mathematicians, but I also have coders. So for example, you know, in healthcare you could have multiple So it seems like magic, but with this capability you enable all kinds of new use cases So discovery and leverage of the days that what you're getting at in terms of the benefits, So I've got to ask you the question because one of the things that's a real interesting trade off these days They're going to have to be resolved, not by companies, but probably, you know, by society and kind you know, one, uh, distinguishing feature of the machine learning field, You can agree or disagree to be successful, you got to be open. But what are you working on that you're excited about? I mean you can almost imagine the trust will go up and you can also kind of end up So for example, banks could compete on, you know, lending and so on under normal activities. So people thought, okay, this is interesting, you know, mathematically, but not practical. I think, you know, you can almost call that product management kind of new fangled Prague product Well, I'm just saying that the field is moving so fast that you're going to need on your team, So we're trying to figure out, you know, whatever it takes to enable the community,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Casimir Wierzynski | PERSON | 0.99+ |
Kaz Borzynski | PERSON | 0.99+ |
five seconds | QUANTITY | 0.99+ |
Catherine | PERSON | 0.99+ |
San Francisco | LOCATION | 0.99+ |
three days | QUANTITY | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
John | PERSON | 0.99+ |
50 milliseconds | QUANTITY | 0.99+ |
Intel | ORGANIZATION | 0.99+ |
RSA | ORGANIZATION | 0.99+ |
Tesla | ORGANIZATION | 0.99+ |
John furrier | PERSON | 0.99+ |
Python | TITLE | 0.99+ |
20 | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
one | QUANTITY | 0.98+ |
One | QUANTITY | 0.98+ |
one year | QUANTITY | 0.98+ |
HIPAA | TITLE | 0.97+ |
RSAC | ORGANIZATION | 0.97+ |
RSA | EVENT | 0.97+ |
today | DATE | 0.97+ |
RSA Congress 2020 | EVENT | 0.97+ |
Prague | LOCATION | 0.97+ |
12 years old | QUANTITY | 0.97+ |
next year | DATE | 0.96+ |
eighties | DATE | 0.96+ |
one set | QUANTITY | 0.95+ |
20 years | QUANTITY | 0.95+ |
USA | LOCATION | 0.94+ |
three triads | QUANTITY | 0.94+ |
second | QUANTITY | 0.93+ |
10 million times | QUANTITY | 0.92+ |
an hour | QUANTITY | 0.92+ |
Moscone | LOCATION | 0.92+ |
one party | QUANTITY | 0.9+ |
Moore | PERSON | 0.89+ |
a year | QUANTITY | 0.88+ |
ORGANIZATION | 0.87+ | |
one thing | QUANTITY | 0.85+ |
next 10 year | DATE | 0.84+ |
last two years | DATE | 0.81+ |
one place | QUANTITY | 0.8+ |
Silicon angle media | ORGANIZATION | 0.78+ |
hundred times | QUANTITY | 0.74+ |
nineties | DATE | 0.74+ |
AI | ORGANIZATION | 0.68+ |
last couple | DATE | 0.65+ |
years | DATE | 0.65+ |
E | OTHER | 0.65+ |
CTO | ORGANIZATION | 0.64+ |
2020 | EVENT | 0.61+ |
AI wave | EVENT | 0.58+ |
G | TITLE | 0.54+ |
number one | QUANTITY | 0.52+ |
JD | PERSON | 0.51+ |
2020 | ORGANIZATION | 0.45+ |
next | DATE | 0.44+ |
H | TITLE | 0.35+ |
wave | EVENT | 0.3+ |