A People’s Guide to Finding Algorithmic Bias
Pioneers at the forefront of algorithmic bias & justice
Algorithmic Audits & Exposés
-
Julia Angwin
Award-winning investigative journalist Julia Angwin is behind some of the most influential exposés on algorithmic bias. The editor in chief of The Markup, she has uncovered biases in algorithms used in pretrial risk assessments, insurance premium determinations, ride-sharing apps, social media advertising, and more. Celebrated for her ability to demystify complex algorithmic injustices and hold powerful corporations accountable, Angwin has shaped discourse around the disparate impact of algorithms by providing numerous striking and concrete examples.
-
Joy Buolamwini
Joy Buolamwini is the founder of the Algorithmic Justice League, an award-winning researcher, and poet of code. She advises world leaders, policymakers, and executives on redressing algorithmic harms. Her work is featured in global exhibitions and the documentary Coded Bias, available on Netflix. Her seminal research paper with Timnit Gebru, Gender Shades, exposed how poorly big-brand facial recognition systems perform on darker-skinned women's faces. The study's findings prompted swift action by Microsoft, IBM, and Amazon to address bias in their designs. Their continued advocacy has drawn massive public awareness to algorithmic biases in everyday technologies.
-
Virginia Eubanks
Virginia Eubanks' writing on technology and social justice highlights the effects of technological discrimination against working-class and poor communities. Her book, Automating Inequality: How High-Tech Tools Profile, Police, and Punish the Poor, is an eye-opening survey of how corporations and governments weaponize big data tools against the most marginalized people.
-
Safiya Noble
Professor Safiya Noble is the author of Algorithms of Oppression, which documents how Google search reinforces racial and gender stereotypes. She persuasively argues that the field needs a Black feminist theory for technology studies. Without widespread adoption of such a lens, daily-use technologies will continue reinforcing Black women and girls' marginalization.
-
Latanya Sweeney
Latanya Sweeney was the first scientist to demonstrate racial biases in commercial algorithms. She found that Google AdSense delivered ads that suggested an individual's arrest more often when searching for Black sounding names than for white-sounding ones. Sweeney found a Black-identifying name to be 25% more likely to get an ad suggestive of an arrest record. This early work established the field of algorithmic fairness.
-
Timnit Gebru
Timnit Gebru is the founder and executive director of the Distributed Artificial Intelligence Research Institute (DAIR). Prior to that she was fired by Google in December 2020 for raising issues of discrimination in the workplace, where she was serving as co-lead of the Ethical AI research team. She received her PhD from Stanford University, and did a postdoc at Microsoft Research, New York City in the FATE (Fairness Accountability Transparency and Ethics in AI) group, where she studied algorithmic bias and the ethical implications underlying projects aiming to gain insights from data. Timnit also co-founded Black in AI, a nonprofit that works to increase the presence, inclusion, visibility and health of Black people in the field of AI, and is on the board of AddisCoder, a nonprofit dedicated to teaching algorithms and computer programming to Ethiopian highschool students, free of charge.
Defining & Measuring Algorithmic Bias
-
Solon Barocas
Solon Barocas co-founded the annual workshop on Fairness, Accountability, and Transparency in Machine Learning (FAT/ML) and later established the ACM conference on Fairness, Accountability, and Transparency (FAccT). His most cited paper, Big Data's Disparate Impact, was among the first to demonstrate how high stakes algorithmic decisions can disproportionately harm historically marginalized groups.
-
Cynthia Dwork
Cynthia Dwork is a computer scientist known for, among many things, her work in mathematically defining and remediating unfairness in machine learning classification. Her most recent work has demonstrated that classifiers that are fair in isolation do not necessarily compose into fair systems. Similarly, her research shows that seemingly unfair components may be carefully combined to construct fair systems.
-
Batya Friedman & Helen Nissenbaum
Batya Friedman and Helen Nissenbaum's groundbreaking work, Bias in Computer Systems, informs much of how we define and detect algorithmic bias today. In this widely cited theoretical paper, they outline how biases in computational tools can be due to (1) pre existing societal biases, (2) technical constraints, or (3) the context of use.
-
Jon Kleinberg
Computer scientist Jon Kleinberg has worked to define fairness in algorithmic decision making concerning equity and efficiency and the role algorithms can play in detecting and preventing real-world discrimination. He also studies how machine predictions influence human decision-makers and the risks inherent in those processes when biases exist.
-
Arvind Narayanan
Arvind Narayanan, along with Baracoas (above), wrote Fairness and Machine Learning: Limitations and Opportunities—the most comprehensive attempt to develop a framework for machine learning development rooted in a theory of fairness. The book is the first of its kind -- a textbook written for and by computer scientists that takes a critical look at bias in computational systems and proposes technical solutions for achieving fairness.
Debiasing Solutions
-
Alexander Amini
Alexander Amini and his co-authors developed a technical solution to algorithmic bias to detect which training inputs come from marginalized groups (e.g., faces with dark skin). This method then samples from those inputs more frequently. This approach results in more balanced representations of outputs that approximate parity despite inequities in training sets.
-
IBM team at AIF360
The IBM team at AIF360 has put together one of the most comprehensive toolkits for detecting and mitigating algorithmic bias in various data science applications. Among their solutions, they offer data-scientist-oriented introductions to the concepts and techniques that undergird their toolkit, which is made free to the public.
-
Sarah Hajian
Sarah Hajian works at the forefront of bias mitigation through data mining techniques. Her work has paved the way for reducing bias from unrepresentative training data sets by detecting and minimizing discrimination in the data collection and curation phases—otherwise known as data mining processes that are discrimination-conscious by design.
-
Weiwei Pan & Finale Doshi-Velez
Weiwei Pan and Finale Doshi-Velez, two computer scientists at the intersection of machine learning and healthcare, have developed uncertainty-aware prediction models. Scientists can use this class of models to make predictions that are just as accurate as traditional models while also providing a metric for how confident a model is on any given prediction. In high-stakes tasks, such as medical diagnosis or judicial sentencing, uncertainty-aware methods can cue human decision-makers to the shortcomings of machine-learning models and mitigate real-world repercussions.
-
Brian Hu Zhang, Blake Lemoine & Margaret Mitchell
Brian Hu Zhang, Blake Lemoine, and Margaret Mitchell advanced adversarial debiasing, a popular debiasing tool used in development. In short, this method aims to learn a representation of the training set that is independent of some protected attribute (e.g., gender) by jointly training an adversary model to prevent the prediction of that same attribute.
Theoretical Frameworks for Critical Algorithm Studies
-
Charlton D. McIlwain
Charlton D. McIlwain demonstrates in his most recent book, Black Software, that racial organizing and social justice movements' online roots span much farther back than most contemporary discourse suggests. He shows that, though we may characterize the Black community's utilization of digital platforms to advance social justice causes as an excellent use case of technology, we cannot discretely separate movements for civil rights with technological advancements. Under oppressive forces, but armed with agency and ingenuity, Black Americans, McIlwain argues, have played an integral role in creating and developing modern-day computing.
-
Meredith Broussard
Meredith Broussard’s book, Artificial Unintelligence: How Computers Misunderstand the World, makes a compelling and accessible case for why society cannot rely on computational models to solve social problems. She provides a framework for holding technologists accountable for technologically driven social inequities. Beyond dismantling myths about artificial intelligence for popular audiences, she complicates the idea that technology can and will be useful only by existing. In Broussard's framework, we can no longer ask, "can we build it?" but instead, should we build it.
-
Ruha Benjamin
Sociologist Ruha Benjamin fuses an abolitionist politic, rooted in Black liberation, with science and technology studies. With roots in the health sciences, her most recent book, Race After Technology: Abolitionist Tools for The New Jim Code, lays out an empowering framework for "reimagining science and technology for liberatory ends.”
-
Kate Crawford & Meredith Whittaker
Kate Crawford & Meredith Whittaker, co-founders of NYU's AI Now Institute, research and teach broadly about AI's various social implications. Employing an intersectional framework, Crawford and Whittaker center a critique of asymmetrical power structures in investigations of big data systems while casting doubt on proposed solutions like increased transparency and explainability. Crawford's newest book, Atlas of AI explores both the hidden human and environmental costs of artificial intelligence. Whittaker, also an author and distinguished professor, is a recognized tech worker political organizer.
-
Catherine D'Ignazio
Catherine D'Ignazio, a co-author of the book, Data Feminism, is a data literacy advocate and researcher who maintains that data science is a form of power. Specifically, she roots her critique of data science inequities in challenging the male/female binary as a means to dismantle other hierarchical classification systems at the heart of the field.
-
Avriel Epps-Darling
Avriel Epps-Darling’s research explores how biases in autonomous technologies differently affect youth of color compared to adults. Specifically, she positions age as a dimension of intersecting identity by which AI discriminates. Her work fuses together developmental psychology and data science to push the narratives around young people and technology beyond the “screen time” and “cyber bullying” debates.