Cecilia Aragon, University of Washington | WiDS Worldwide Conference 2022
>>Hey, everyone. Welcome to the cubes coverage of women in data science, 2022. I'm Lisa Martin. And I'm here with one of the key featured keynotes for this year is with events. So the Aragon, the professor and department of human centered design and engineering at the university of Washington Cecilia, it's a pleasure to have you on the cube. >>Thank you so much, Lisa Lisa, it's a pleasure to be here as well. >>You got an amazing background that I want to share with the audience. You are a professor, you are a data scientist, an aerobatic pilot, and an author with expertise in human centered, data science, visual analytics, aviation safety, and analysis of extremely large and complex data sets. That's quite the background. >>Well, thank you so much. It's it's all very interesting and fun. So, >>And as a professor, you study how people make sense of vast data sets, including a combination of computer science and art, which I love. And as an author, you write about interesting things. You write about how to overcome fear, which is something that everybody can benefit from and how to expand your life until it becomes amazing. I need to take a page out of your book. You were also honored by president Obama a few years back. My goodness. >>Thank you so much. Yes. I I've had quite a journey to come here, but I feel really fortunate to be here today. >>Talk about that journey. I'd love to understand if you were always interested in stem, if it was something that you got into later, I know that you are the co-founder of Latinas in computing, a passionate advocate for girls and women in stem. Were you always interested in stem or was it something that you got into in a kind of a non-linear path? >>I was always interested in it when I was a young girl. I grew up in a small Midwestern town and my parents are both immigrants and I was one of the few Latinas in a mostly white community. And I was, um, I loved math, but I also wanted to be an astronaut. And I remember I, when we were asked, I think it was in second grade. What would you like to be when you grow up? I said, oh, I want to be an astronaut. And my teacher said, oh, you can't do that. You're a girl pick something else. And um, so I picked math and she was like, okay. >>Um, so I always wanted to, well, maybe it would be better to say I never really quite lost my love of being up in the air and potentially space. But, um, but I ended up working in math and science and, um, I, I loved it because one of the great advantages of math is that it's kind of like a magic trick for young people, especially if you're a girl or if you are from an underrepresented group, because if you get the answers right on a math test, no one can mark you wrong. It doesn't matter what the color of your skin is or what your gender is. Math is powerful that way. And I will say there's nothing like standing in a room in front of a room of people who think little of you and you silence them with your love with numbers. >>I love that. I never thought about math as power before, but it clearly is. But also, you know, and, and I wish we had more time because I would love to get into how you overcame that fear. And you write books about that, but being told you can't be an astronaut. You're a girl and maybe laughing at you because you liked Matt. How did you overcome that? And so nevermind I'm doing it anyway. >>Well, that's a, it's a, okay. The short answer is I had incredible imposter syndrome. I didn't believe that I was smart enough to get a PhD in math and computer science. But what enabled me to do that was becoming a pilot and I B I learned how to fly small airplanes. I learned how to fly them upside down and pointing straight at the ground. And I know this might sound kind of extreme. So this is not what I recommend to everybody. But if you are brought up in a way where everybody thinks little of you, one of the best things you can possibly do is take on a challenge. That's scary. I was afraid of everything, but by learning to fly and especially learning to fly loops and rolls, it gave me confidence to do everything else because I thought I appointed the airplane at the ground at 250 miles an hour and waited, why am I afraid to get a PhD in computer science? >>Wow. How empowering is that? >>Yeah, it really was. So that's really how I overcame the fear. And I will say that, you know, I encountered situations getting my PhD in computer science where I didn't believe that I was good enough to finish the degree. I didn't believe that I was smart enough. And what I've learned later on is that was just my own emotional, you know, residue from my childhood and from people telling me that they, you know, that they, that I couldn't achieve >>As I look what, look what you've achieved so far. It's amazing. And we're going to be talking about some of the books that you've written, but I want to get into data science and AI and get your thoughts on this. Why is it necessary to think about human issues and data science >>And what are your thoughts there? So there's been a lot of work in data science recently looking at societal impacts. And if you just address data science as a purely technical field, and you don't think about unintended consequences, you can end up with tremendous injustices and societal harms and harms to individuals. And I think any of us who has dealt with an inflexible algorithm, even if you just call up, you know, customer service and you get told, press five for this press four for that. And you say, well, I don't fit into any of those categories, you know, or have the system hang up on you after an hour. I think you'll understand that any type of algorithmic approach, especially on very large data sets has the risk of impacting people, particularly from low income or marginalized groups, but really any of us can be impacted in a negative way. >>And so, as a developer of algorithms that work over very large data sets, I've always found it really important to consider the humans on the other end of the algorithm. And that's why I believe that all data science is truly human centered or should be human centered, should be human centered and also involves both technical issues as well as social issues. Absolutely correct. So one example is that, um, many of us who started working in data science, including I have to admit me when I started out assume that data is unbiased. It's scrubbed of human influence. It is pure in some ways, however, that's really not true as I've started working with datasets. And this is generally known in the field that data sets are touched by humans everywhere. As a matter of fact, in our, in the recent book that we're, that we're coming out with human centered data science, we talk about five important points where humans touch data, no matter how scrubbed of human influence it's support it's supposed to be. >>Um, so the first one is discovery. So when a human encounters, a data set and starts to use it, it's a human decision. And then there's capture, which is the process of searching for a data set. So any data that has to be selected and chosen by an individual, um, then once that data set is brought in there's curation, a human will have to select various data sets. They'll have to decide what is, what is the proper set to use. And they'll be making judgements on this the time. And perhaps one of the most important ways the data is changed and touched by humans is what we call the design of data. And what that means is whenever you bring in a data set, you have to categorize it. No, for example, let's suppose you are, um, a geologist and you are classifying soil data. >>Well, you don't just take whatever the description of the soil data is. You actually may put it into a previously established taxonomy and you're making human judgments on that. So even though you think, oh, geology data, that's just rocks. You know, that's soil. It has nothing to do with people, but it really does. Um, and finally, uh, people will label the data that they have. And this is especially critical when humans are making subjective judgments, such as what race is the person in this dataset. And they may judge it based on looking at the individual skin color. They may try to apply an algorithm to it, but you know what? We all have very different skin colors, categorizing us into race boxes, really diminishes us and makes us less than we truly are. So it's very important to realize that humans touch the data. We interpret the data. It is not scrubbed of bias. And when we make algorithmic decisions, even the very fact of having an algorithm that makes a judgment say on whether a prisoner's likely to offend again, the judge just by having an algorithm, even if the algorithm makes a recommended statement, they are impacted by that algorithms recommendation. And that has obviously an impact on that human's life. So we consider all of this. >>So you just get given five solid reasons why data science and AI are inevitably human centric should be, but in the past, what's led to the separation between data science and humans. >>Well, I think a lot of it simply has to do with incorrect mental models. So many of us grew up thinking that, oh, humans have biases, but computers don't. And so if we just take decision-making out of people's hands and put it into the hands of an algorithm, we will be having less biased results. However, recent work in the field of data science and artificial intelligence has shown that that's simply not true that algorithmic algorithms reinforce human biases. They amplify them. So algorithmic biases can be much worse than human biases and can greater impact. >>So how do we pull ethics into all of this data science and AI and that ethical component, which seems to be that it needs to be foundational. >>It absolutely has to be foundational. And this is why we believe. And what we teach at the university of Washington in our data science courses is that ethical and human centered approaches and ideas have to be brought in at the very beginning of the algorithm. It's not something you slap on at the end or say, well, I'll wait for the ethicists to weigh in on this. Now we are all human. We can all make human decisions. We can all think about the unintended consequences of our algorithms as we develop them. And we should do that at the very beginning. And all algorithm designers really need to spend some time thinking about the impact that their algorithm may have. >>Right. Do you, do you find that people are still in need of convincing of that or is it generally moving in that direction of understanding? We need to bring ethics in from the beginning, >>It's moving in that direction, but there are still people who haven't modified their mental models yet. So we're working on it. And we hope that with the publication of our book, that it will be used as a supplemental textbook in many data science courses that are focused exclusively on the algorithms and that they can open up the idea that considering the human centered approaches at the beginning of learning about algorithms and data science and the mathematical and statistical techniques, that the next generation of data scientists and artificial intelligence developers will be able to mitigate some of the potentially harmful effects. And we're very excited about this. This is why I'm a professor, because I want to teach the next generation of data scientists and artificial intelligence experts, how to make sure that their work really achieves what they intended it to, which is to make the world a better place, not a worse place, but to enable humans to do better and to mitigate biases and really to lead us into this century in a positive way. >>So the book, human centered data science, you can see it there over Sicily, his right shoulder. When does this come out and how can folks get a copy of it? >>So it came out March 1st and it's available in bookstores everywhere. It was published by MIT press, and you can go online or you can go to your local independent bookstore, or you can order it from your university bookstore as well. >>Excellent. Got to, got to get a copy of, get my hands on that. Got cut and get a copy and dig into that. Cause it sounds so interesting, but also so thoughtful and, um, clear in the way that you described that. And also all the opportunities that, that AI data science and humans are gonna unlock for the world and humans and jobs and, and great things like that. So I'm sure there's lots of great information there. Last question I mentioned, you are keynoting at this year's conference. Talk to me about like the top three takeaways that the audience is going to get from your keynote. >>So I'm very excited to have been invited to wins this year, which of course is a wonderful conference to support women in data science. And I've been a big fan of the conference since it was first developed here, uh, here at Stanford. Um, the three, the three top takeaways I would say is to really consider the data. Science can be rigorous and mathematical and human centered and ethical. It's not a trade-off, it's both at the same time. And that's really the, the number one that, that I'm hoping to keynote will bring to, to the entire audience. And secondly, I hope that it will encourage women or people who've been told that maybe you're not a science person or this isn't for you, or you're not good at math. I hope it will encourage them to disbelieve those views. And to realize that if you, as a member of any type of unread, underrepresented group have ever felt, oh, I'm not good enough for this. >>I'm not smart enough. It's not for me that you will reconsider because I firmly believe that everyone can be good at math. And it's a matter of having the information presented to you in a way that honors your, the background you had. So when I started out my, my high school didn't have AP classes and I needed to learn in a somewhat different way than other people around me. And it's really, it's really something. That's what I tell young people today is if you are struggling in a class, don't think it's because you're not good enough. It might just be that the teacher is not presenting it in a way that is best for someone with your particular background. So it doesn't mean they're a bad teacher. It doesn't mean you're unintelligent. It just means the, maybe you need to find someone else that can explain it to you in a simple and clear way, or maybe you need to get some scaffolding that is Tate, learn extra, take extra classes that will help you. Not necessarily remedial classes. I believe very strongly as a teacher in giving students very challenging classes, but then giving them the scaffolding so that they can learn that difficult material. And I have longer stories on that, but I think I've already talked a bit too long. >>I love that. The scaffolding, I th I think the, the one, one of the high level takeaways that we're all going to get from your keynote is inspiration. Thank you so much for sharing your path to stem, how you got here, why humans, data science and AI are, have to be foundationally human centered, looking forward to the keynote. And again, Cecilia, Aragon. Thank you so much for spending time with me today. >>Thank you so much, Lisa. It's been a pleasure, >>Likewise versus silly Aragon. I'm Lisa Martin. You're watching the cubes coverage of women in data science, 2022.
SUMMARY :
of Washington Cecilia, it's a pleasure to have you on the cube. You are a professor, you are a data scientist, Well, thank you so much. And as a professor, you study how people make sense of vast data sets, including a combination of computer Thank you so much. if it was something that you got into later, I know that you are the co-founder of Latinas in computing, And my teacher said, oh, you can't do that. And I will say there's nothing like standing in And you write books about that, but being told you can't be an astronaut. And I know this might sound kind of extreme. And I will say that, you know, I encountered situations And we're going to be talking about some of the books that you've written, but I want to get into data science and AI And you say, well, I don't fit into any of those categories, you know, And so, as a developer of algorithms that work over very large data sets, And what that means is whenever you bring in a And that has obviously an impact on that human's life. So you just get given five solid reasons why data science and AI Well, I think a lot of it simply has to do with incorrect So how do we pull ethics into all of this data science and AI and that ethical And all algorithm designers really need to spend some time thinking about the is it generally moving in that direction of understanding? that considering the human centered approaches at the beginning So the book, human centered data science, you can see it there over Sicily, his right shoulder. or you can go to your local independent bookstore, or you can order it from your university takeaways that the audience is going to get from your keynote. And I've been a big fan of the conference since it was first developed here, the information presented to you in a way that honors your, to stem, how you got here, why humans, data science and AI women in data science, 2022.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Cecilia | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Aragon | PERSON | 0.99+ |
March 1st | DATE | 0.99+ |
Lisa | PERSON | 0.99+ |
2022 | DATE | 0.99+ |
three | QUANTITY | 0.99+ |
Lisa Lisa | PERSON | 0.99+ |
president | PERSON | 0.99+ |
Cecilia Aragon | PERSON | 0.99+ |
Sicily | LOCATION | 0.99+ |
Matt | PERSON | 0.99+ |
both | QUANTITY | 0.99+ |
five important points | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
today | DATE | 0.98+ |
250 miles an hour | QUANTITY | 0.98+ |
one | QUANTITY | 0.97+ |
MIT press | ORGANIZATION | 0.97+ |
second grade | QUANTITY | 0.97+ |
five solid reasons | QUANTITY | 0.97+ |
one example | QUANTITY | 0.97+ |
an hour | QUANTITY | 0.97+ |
three top takeaways | QUANTITY | 0.97+ |
four | QUANTITY | 0.96+ |
five | QUANTITY | 0.96+ |
first one | QUANTITY | 0.95+ |
this year | DATE | 0.94+ |
University of Washington | ORGANIZATION | 0.94+ |
this year | DATE | 0.94+ |
Midwestern | LOCATION | 0.93+ |
three takeaways | QUANTITY | 0.88+ |
WiDS Worldwide Conference 2022 | EVENT | 0.87+ |
few years back | DATE | 0.8+ |
university of Washington Cecilia | ORGANIZATION | 0.77+ |
Stanford | LOCATION | 0.76+ |
university of Washington | ORGANIZATION | 0.75+ |
silly | PERSON | 0.74+ |
Obama | PERSON | 0.74+ |
Tate | PERSON | 0.71+ |
Aragon | ORGANIZATION | 0.69+ |
top | QUANTITY | 0.6+ |
Latinas | PERSON | 0.57+ |
Latinas | OTHER | 0.57+ |