Data Science Coast to Coast: Urban Informatics
April 8 @ 3:00 pm - 4:00 pm
The data science institutes at seven universities invite you to join the 2021 Data Science Coast-to-Coast seminar series. These data science institutes have been planning ways to foster a broad-reaching data science community that can collaborate extensively to advance our missions for research, education, and data science for social good. Using Zoom to eliminate geographical constraints, this seminar series is one of our first steps toward the goal.
In the first half of 2021, we will host five seminars, each featuring one faculty member and one postdoctoral fellow from two universities. Each speaker will give a 20-minute talk about ongoing projects and motivating issues, followed by 20 minutes of discussion with the audience. These seminars will be the launching point for follow-on research discussion meetings which will hopefully lead to fruitful collaborative research.
We strongly encourage you to sign up ahead of time, and indicate your area of research and your interest in follow-up discussions.
Quantifying and Mitigating Sources of Bias in a Decision-Support System: Applications of AI decision-support systems are increasingly shaping the fabric of our society. These systems can exhibit and exacerbate undesired biases that might hurt the under-represented population. Therefore, it is critical to evaluate these systems not only from a lens of predictive power and the rate of error but also from a lens of trustworthiness and fairness. In this talk, I will focus on two specific sources of bias in a decision-support system and propose mitigation strategies. In the first part, I will discuss biases originated from historical decisions and are reflected in data. I propose a metric of quantifying disparity in data and illustrate how we can alleviate these historical biases by applying simple modification to a decision-making system. In the second part, I will shed light on biases that are originated from predictive models. Predictive models are a central part of any decision-making system. The end-user act based on the information provided by these models. Biased or untrustworthy information mislead the end-user or incentivize the public to mistrust the system. I will present our mitigation method KiTE. KiTE is a hypothesis-testing framework with provable guarantees that enables practitioners to (i) test whether a model provides trustworthy information with respect to each sub-group of a population and (ii) estimate and correct for prediction bias at the individual and group levels.
Revealing the “Big Lie”: Collaborative Data Science for Rapid Response to Online Disinformation: In this talk, I’ll present preliminary research results from ongoing efforts to understand the spread of disinformation about the 2020 Election. First, I’ll describe the mission, structure, and everyday work practices of the Election Integrity Partnership — a multi-stakeholder collaboration that addressed mis- and disinformation about the 2020 U.S. election in (near) real-time through rapid response data science. Next, I’ll take you through some of our analyses to show how the “Big Lie” — the sustained effort to sow doubt in the results of the 2020 election — took shape on social media platforms throughout the latter half of 2020. I’ll highlight the participatory nature of this disinformation campaign and reveal some of the “super spreader” accounts that helped produce and sustain it. Finally, I’ll note how some of the social media platforms have evolved their strategies to address this kind of disinformation and wrap up by talking about what might come next, both in terms of platform policies and future collaborations for rapid response to disinformation.