Joy Buolamwini

Last updated

Joy Buolamwini
Joy Buolamwini - Wikimania 2018 01.jpg
Buolamwini at Wikimania 2018
Born
Joy Adowaa Buolamwini

1989or1990(age 33–34)
Education Cordova High School
Alma mater Georgia Institute of Technology (BS)
Jesus College, Oxford (MS)
Massachusetts Institute of Technology (MS, PhD)
Known for Algorithmic Justice League
Scientific career
FieldsMedia Arts & Sciences
Computer science
Algorithmic bias
Institutions MIT Media Lab
Theses
Doctoral advisor Ethan Zuckerman [1]
Website www.poetofcode.com OOjs UI icon edit-ltr-progressive.svg

Joy Adowaa Buolamwini is a Ghanaian-American-Canadian computer scientist and digital activist based at the MIT Media Lab. [2] She founded the Algorithmic Justice League (AJL), an organization that works to challenge bias in decision-making software, using art, advocacy, and research to highlight the social implications and harms of artificial intelligence (AI). [3]

Contents

Early life and education

Buolamwini was born in Edmonton, Alberta, grew up in Mississippi, and attended Cordova High School in Cordova, Tennessee. [4] At age nine, she was inspired by Kismet, the MIT robot, and taught herself XHTML, JavaScript and PHP. [5] [6] She was a competitive pole vaulter. [7]

As an undergraduate, Buolamwini studied computer science at the Georgia Institute of Technology, where she researched health informatics. [8] Buolamwini graduated as a Stamps President's Scholar [9] from Georgia Tech in 2012, [10] and was the youngest finalist of the Georgia Tech InVenture Prize in 2009. [11]

Buolamwini is a Rhodes Scholar, a Fulbright fellow, a Stamps Scholar, an Astronaut Scholar, and an Anita Borg Institute scholar. [12] As a Rhodes Scholar, she studied learning and technology at the University of Oxford, where she was a student based at Jesus College, Oxford. [13] [14] During her scholarship she took part in the first formal Service Year, working on community focused projects. [14] [15] She was awarded a Master's Degree in Media Arts & Sciences from MIT in 2017 for research supervised by Ethan Zuckerman. [1] She was awarded a PhD degree in Media Arts & Sciences from the MIT Media Lab in 2022 with a thesis on Facing the Coded Gaze with Evocative Audits and Algorithmic Audits. [16]

Career and research

In 2011, Buolamwini worked with the trachoma program at the Carter Center to develop an Android-based assessment system for use in Ethiopia. [17] [5]

Joy Buolamwini at Wikimania 2018 in Cape Town Joy Buolamwini - Wikimania 2018 02.jpg
Joy Buolamwini at Wikimania 2018 in Cape Town

As a Fulbright fellow, in 2013 she worked with local computer scientists in Zambia to help Zambian youth become technology creators. [18] On September 14, 2016, Buolamwini appeared at the White House summit on Computer Science for All.[ citation needed ]

She is a researcher at the MIT Media Lab, where she works to identify bias in algorithms and to develop practices for accountability during their design; [19] at the lab, Buolamwini is a member of Ethan Zuckerman's Center for Civic Media group. [20] [21] During her research, Buolamwini showed 1,000 faces to facial recognition systems and asked the systems to identify whether faces were female or male, and found that the software found it hard to identify dark-skinned women. [22] Her project, Gender Shades, became part of her MIT thesis. [1] [23] Her 2018 paper Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification [24] prompted responses from IBM and Microsoft, which swiftly improved their software. [25] [26] She also created the Aspire Mirror, a device that lets users see a reflection of themselves based on what inspires them. [27] Her program, Algorithmic Justice League, aims to highlight the bias in code that can lead to discrimination against underrepresented groups. [28] She has created two films, Code4Rights and Algorithmic Justice League: Unmasking Bias. [29] [30] She served as Chief Technology Officer (CTO) for Techturized Inc., a hair-care technology company. [8]

Buolamwini's research was cited in 2020 as an influence for Google and Microsoft in addressing gender and race bias in their products and processes. [31]

She also served as an advisor to President Biden ahead of his administration's Executive Order 14110, released October 30, 2023. The order is also known as the Executive Order on Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence (sometimes referred to as "Executive Order on Artificial Intelligence"). [32] [33]

In 2023, she published her first book, Unmasking AI: My Mission to Protect What Is Human in a World of Machines, which chronicles her research. [34]

Activism

Buolamwini founded the Algorithmic Justice League (AJL) in 2016 to promote equitable and accountable artificial intelligence (AI). [35] The AJL organization combines art and research to point out potential societal implications and harms of AI. The company works to raise public awareness of the impacts of AI, and promote further research in the area.

The AJL website provides information and a live blog. [36] There are several sections on the site where users can share stories, and donate or write to US Congressional representatives. In 2019, Buolamwini testified before the United States House Committee on Oversight and Reform about the risks of facial recognition technology. [37]

Voicing Erasure

The Voicing Erasure section on the AJL website hosts spoken pieces by Buolamwini, Allison Koenecke, Safiya Noble, Ruha Benjamin, Kimberlé Crenshaw, Megan Smith, and Sasha Costanza-Chock about bias in voice systems. [38] [39] Buolamwini and Koenecke are the lead researchers on the website working to uncovering biases of voice systems. They've written that speech recognition systems have the most trouble with African-American Vernacular English speakers, and that these systems are secretly listening to users' conversations.[ citation needed ] They have also written about what they regard as harmful gender stereotypes perpetuated by the voice recognition systems in Siri, Amazon Alexa, and Microsoft Cortana.[ citation needed ]

The Coded Gaze

The Coded Gaze is a mini-documentary that debuted at the Museum of Fine Arts, Boston in 2016, and is currently available via YouTube. Buolamwini uses the mini documentary to talk about the bias that she believes lies in artificial intelligence's function. The inspiration for the mini documentary and her research came when she was at MIT, creating her art "Aspire Mirror," which uses facial recognition to reflect another person who inspires a user, onto that user's face. [40] Buolamwini anticipated having Serena Williams, another dark-skinned woman, reflected onto her face. However, the technology did not recognize her face. Buolamwini's research investigated why this happened, and consequently led Buolamwini to conclude that the exclusion of people who look like her was a result of a practice she called the "Coded Gaze." [41] She further discusses this concept in the mini documentary, "The Coded Gaze." The documentary explores how AI can be subject to racial and gender biases that reflect the views and cultural backgrounds of those who develop it. [42]

Coded Bias

Coded Bias is a documentary film directed by Shalini Kantayya that features Buolamwini’s research about AI inaccuracies in facial recognition technology and automated assessment software. [43] [36] It focuses on what the film's creators regard as a lack of regulation of facial recognition tools sold by IBM, Microsoft, and Amazon, and which they say perpetuates racial and gender bias. The film describes a dispute between Brooklyn tenants and a building management company that tried to use facial recognition to control entry to a building. The film featured Weapons of Math Destruction author Cathy O'Neill and members of Big Brother Watch in London, including Silkie Carlo. On April 5, 2021, the documentary was made available to stream on Netflix. [44]

Exhibitions

Projects conducted by Algorithmic Justice League have been exhibited at art institutions including the Barbican Centre in London, UK, and Ars Electronica in Linz, Austria. [45]

Awards and honors

In 2017, Buolamwini was awarded the grand prize in the professional category in the Search for Hidden Figures contest, tied to the release of the film Hidden Figures in December 2016. [51] The contest, sponsored by PepsiCo and 21st Century Fox, was intended to "help uncover the next generation of female leaders in science, technology, engineering and math," [52] and attracted 7,300 submissions from young women across the United States. [10]

Buolamwini delivered a TEDx talk at Beacon Street entitled How I'm fighting bias in algorithms. [53] [54] [55] In 2018, she appeared on the TED Radio Hour. [56] She was featured on Amy Poehler's Smart Girls in 2018. [4] Fast Company magazine listed her as one of four "design heroes who are defending democracy online." [57] She was listed as one of BBC's 100 Women in 2018. [58]

In 2019, Buolamwini was listed in Fortune magazine's 2019 list of the "World's 50 Greatest Leaders." The magazine also described her as "the conscience of the A.I. revolution." [59] She also made the inaugural Time 100 Next list in 2019. [60] In 2020, Buolamwini featured in a women's empowerment campaign by the clothing company Levi's for International Women's Day. [61] She was also featured in the documentary Coded Bias. [62]

In 2022, Buolamwini was named the ASQ Hutchens Medalist. [63] In 2023, she was listed in the Time 100 AI. [64]

Personal life

Buolamwini has lived in Ghana; Barcelona, Spain; Oxford, United Kingdom; and, in the U.S., Memphis, Tennessee, and Atlanta, Georgia. [11]

Related Research Articles

<span class="mw-page-title-main">Facial recognition system</span> Technology capable of matching a face from an image against a database of faces

A facial recognition system is a technology potentially capable of matching a human face from a digital image or a video frame against a database of faces. Such a system is typically employed to authenticate users through ID verification services, and works by pinpointing and measuring facial features from a given image.

<span class="mw-page-title-main">Kate Crawford</span> Australian writer, composer, and academic

Kate Crawford is a researcher, writer, composer, producer and academic, who studies the social and political implications of artificial intelligence. She is based in New York and works as a principal researcher at Microsoft Research, the co-founder and former director of research at the AI Now Institute at NYU, a visiting professor at the MIT Center for Civic Media, a senior fellow at the Information Law Institute at NYU, and an associate professor in the Journalism and Media Research Centre at the University of New South Wales. She is also a member of the WEF's Global Agenda Council on Data-Driven Development.

DeepFace is a deep learning facial recognition system created by a research group at Facebook. It identifies human faces in digital images. The program employs a nine-layer neural network with over 120 million connection weights and was trained on four million images uploaded by Facebook users. The Facebook Research team has stated that the DeepFace method reaches an accuracy of 97.35% ± 0.25% on Labeled Faces in the Wild (LFW) data set where human beings have 97.53%. This means that DeepFace is sometimes more successful than human beings. As a result of growing societal concerns Meta announced that it plans to shut down Facebook facial recognition system, deleting the face scan data of more than one billion users. This change will represent one of the largest shifts in facial recognition usage in the technology's history. Facebook planned to delete by December 2021 more than one billion facial recognition templates, which are digital scans of facial features. However, it did not plan to eliminate DeepFace which is the software that powers the facial recognition system. The company has also not ruled out incorporating facial recognition technology into future products, according to Meta spokesperson.

<span class="mw-page-title-main">Ruha Benjamin</span> American sociologist

Ruha Benjamin is a sociologist and a professor in the Department of African American Studies at Princeton University. The primary focus of her work is the relationship between innovation and equity, particularly focusing on the intersection of race, justice and technology. Benjamin is the author of numerous publications, including the books People's Science: Bodies and Rights on the Stem Cell Frontier (2013), Race After Technology: Abolitionist Tools for the New Jim Code (2019) and Viral Justice: How We Grow the World We Want (2022).

<span class="mw-page-title-main">Shalini Kantayya</span> American filmmaker and activist

Shalini Kantayya is an American filmmaker and environmental activist based in Brooklyn, New York, whose films explore human rights at the intersection of water, food, and renewable energy. Kantayya is best known for her debut feature documentary, Catching the Sun.

Labeled data is a group of samples that have been tagged with one or more labels. Labeling typically takes a set of unlabeled data and augments each piece of it with informative tags. For example, a data label might indicate whether a photo contains a horse or a cow, which words were uttered in an audio recording, what type of action is being performed in a video, what the topic of a news article is, what the overall sentiment of a tweet is, or whether a dot in an X-ray is a tumor.

<span class="mw-page-title-main">Algorithmic bias</span> Technological phenomenon with social implications

Algorithmic bias describes systematic and repeatable errors in a computer system that create "unfair" outcomes, such as "privileging" one category over another in ways different from the intended function of the algorithm.

<span class="mw-page-title-main">Meredith Broussard</span> Data journalism professor

Meredith Broussard is a data journalism professor at the Arthur L. Carter Journalism Institute at New York University. Her research focuses on the role of artificial intelligence in journalism.

<span class="mw-page-title-main">Meredith Whittaker</span> American artificial intelligence research scientist

Meredith Whittaker is the president of the Signal Foundation and serves on its board of directors. She was formerly the Minderoo Research Professor at New York University (NYU), and the co-founder and faculty director of the AI Now Institute. She also served as a senior advisor on AI to Chair Lina Khan at the Federal Trade Commission. Whittaker was employed at Google for 13 years, where she founded Google's Open Research group and co-founded the M-Lab. In 2018, she was a core organizer of the Google Walkouts and resigned from the company in July 2019.

<span class="mw-page-title-main">Timnit Gebru</span> Computer scientist (born 1983)

Timnit Gebru is an Eritrean Ethiopian-born computer scientist who works in the fields of artificial intelligence (AI), algorithmic bias and data mining. She is an advocate for diversity in technology and co-founder of Black in AI, a community of Black researchers working in AI. She is the founder of the Distributed Artificial Intelligence Research Institute (DAIR).

Amazon Rekognition is a cloud-based software as a service (SaaS) computer vision platform that was launched in 2016. It has been sold to, and used by, a number of United States government agencies, including U.S. Immigration and Customs Enforcement (ICE) and Orlando, Florida police, as well as private entities.

Olga Russakovsky is an associate professor of computer science at Princeton University. Her research investigates computer vision and machine learning. She was one of the leaders of the ImageNet Large Scale Visual Recognition challenge and has been recognised by MIT Technology Review as one of the world's top young innovators.

<i>Coded Bias</i> 2020 American documentary film

Coded Bias is an American documentary film directed by Shalini Kantayya that premiered at the 2020 Sundance Film Festival. The film includes contributions from researchers Joy Buolamwini, Deborah Raji, Meredith Broussard, Cathy O’Neil, Zeynep Tufekci, Safiya Noble, Timnit Gebru, Virginia Eubanks, and Silkie Carlo, and others.

<span class="mw-page-title-main">Algorithmic Justice League</span> Digital advocacy non-profit organization

The Algorithmic Justice League (AJL) is a digital advocacy non-profit organization based in Cambridge, Massachusetts. Founded in 2016 by computer scientist Joy Buolamwini, the AJL uses research, artwork, and policy advocacy to increase societal awareness regarding the use of artificial intelligence (AI) in society and the harms and biases that AI can pose to society. The AJL has engaged in a variety of open online seminars, media appearances, and tech advocacy initiatives to communicate information about bias in AI systems and promote industry and government action to mitigate against the creation and deployment of biased AI systems. In 2021, Fast Company named AJL as one of the 10 most innovative AI companies in the world.

<span class="mw-page-title-main">Rashida Richardson</span> American attorney and scholar

Rashida Richardson is a visiting scholar at Rutgers Law School and the Rutgers Institute for Information Policy and the Law and an attorney advisor to the Federal Trade Commission. She is also an assistant professor of law and political science at the Northeastern University School of Law and the Northeastern University Department of Political Science in the College of Social Sciences and Humanities.

<span class="mw-page-title-main">Deborah Raji</span> Nigerian-Canadian computer scientist and activist

Inioluwa Deborah Raji is a Nigerian-Canadian computer scientist and activist who works on algorithmic bias, AI accountability, and algorithmic auditing. Raji has previously worked with Joy Buolamwini, Timnit Gebru, and the Algorithmic Justice League on researching gender and racial bias in facial recognition technology. She has also worked with Google’s Ethical AI team and been a research fellow at the Partnership on AI and AI Now Institute at New York University working on how to operationalize ethical considerations in machine learning engineering practice. A current Mozilla fellow, she has been recognized by MIT Technology Review and Forbes as one of the world's top young innovators.

Karen Hao is an American journalist and data scientist. Currently a contributing writer for The Atlantic and previously a foreign correspondent based in Hong Kong for The Wall Street Journal and senior artificial intelligence editor at the MIT Technology Review, she is best known for her coverage on AI research, technology ethics and the social impact of AI. Hao also co-produces the podcast In Machines We Trust and writes the newsletter The Algorithm.

<span class="mw-page-title-main">Virginia Eubanks</span> American political scientist, author

Virginia Eubanks is an American political scientist, professor, and author studying technology and social justice. She is an associate professor in the Department of Political Science at the University at Albany, SUNY. Previously Eubanks was a Fellow at New America researching digital privacy, economic inequality, and data-based discrimination.

<span class="mw-page-title-main">Tawana Petty</span>

Tawana Petty is an American author, poet, social justice organizer, mother and youth advocate who works to counter systemic racism. Petty formerly served as Director of Policy and Advocacy for the Algorithmic Justice League representing AJL in national and international processes shaping AI governance.

<span class="mw-page-title-main">Allison Koenecke</span> American computer scientist and academic

Allison Koenecke is an American computer scientist and an assistant professor in the Department of Information Science at Cornell University. Her research considers computational social science and algorithmic fairness. In 2022, Koenecke was named one of the Forbes 30 Under 30 in Science.

References

  1. 1 2 3 Buolamwini, Joy Adowaa (2017). Gender shades : intersectional phenotypic and demographic evaluation of face datasets and gender classifiers (MS thesis). MIT. hdl:1721.1/114068. OCLC   1026503582. Lock-green.svg
  2. "Joy Buolamwini". forbes.com. Retrieved March 19, 2022.
  3. "Algorithmic Justice League - Unmasking AI harms and biases". Algorithmic Justice League - Unmasking AI harms and biases. Retrieved May 15, 2021.
  4. 1 2 "The Future of Computer Science and Tech: 12 Young Women to Watch — Part 2". amysmartgirls.com. Amy Poehler’s Smart Girls. February 19, 2018. Retrieved March 24, 2018.
  5. 1 2 "Joy Buolamwini | TryComputing.org". trycomputing.org. Archived from the original on March 25, 2018. Retrieved March 24, 2018.
  6. "Meet The Digital Activist That's Taking Human Prejudice Out of Our Machines". Bloomberg.com. June 26, 2017. Retrieved March 24, 2018.
  7. "CHS Pole Vaulting - Joy Buolamwini". vault.awardspace.com. Archived from the original on March 25, 2018. Retrieved March 24, 2018.
  8. 1 2 "Tech Startup of The Week: Techturized Wins With Hair Care Company". Black Enterprise. March 15, 2013. Retrieved March 24, 2018.
  9. "Stamps President's Scholars Program". stampsps.gatech.edu.
  10. 1 2 "Joy Buolamwini wins national contest for her work fighting bias in machine learning". MIT News. Retrieved March 24, 2018.
  11. 1 2 "Admissions Conquered | InVenture Prize". inventureprize.gatech.edu. Retrieved September 25, 2021.
  12. "Scholar Spotlight: Joy Buolamwini | Astronaut Scholarship Foundation" . Retrieved September 25, 2021.
  13. Buolamwini, Joy Adowaa (2014). Increasing participation in graduate level computer science education : a case study of the Georgia Institute of Technology's master of computer science. ox.ac.uk (MSc thesis). University of Oxford. OCLC   908967245.
  14. 1 2 "Joy Buolamwini Profile". rhodesproject.com. The Rhodes Project. Retrieved March 24, 2018.
  15. "Oxford Launchpad: Confessions of an Entrepreneur: Joy Buolamwini | Enterprising Oxford". eship.ox.ac.uk. Archived from the original on March 25, 2018. Retrieved March 24, 2018.
  16. Buolamwini, Joy (2022). Facing the Coded Gaze with Evocative Audits and Algorithmic Audits. mit.edu (PhD thesis). hdl:1721.1/143396.
  17. "Scholar Spotlight: Joy Buolamwini | Astronaut Scholarship Foundation". astronautscholarship.org. Retrieved March 24, 2018.
  18. ZamrizeMedia (April 28, 2013), Joy Buolamwini | Fulbright Fellow 2013 | Zambia , retrieved March 24, 2018
  19. "Project Overview ‹ Algorithmic Justice League – MIT Media Lab". MIT Media Lab. Retrieved March 24, 2018.
  20. "interview: joy buolamwini | MIT Admissions". mitadmissions.org. March 5, 2017. Retrieved March 24, 2018.
  21. "Group People ‹ Civic Media – MIT Media Lab". MIT Media Lab. Retrieved March 24, 2018.
  22. "Photo Algorithms ID White Men Fine—Black Women, Not So Much". wired.com. Wired magazine. Retrieved March 24, 2018.
  23. Kleinman, Zoe (April 14, 2017). "Is artificial intelligence racist?". bbc.co.uk. BBC News . Retrieved March 24, 2018.
  24. Buolamwini, Joy (2018). "Gender shades: Intersectional accuracy disparities in commercial gender classification". Conference on Fairness, Accountability and Transparency. 81: 77–91 via mir.press.
  25. "Mitigating Bias in Artificial Intelligence (AI) Models -- IBM Research". ibm.com. May 16, 2016. Retrieved March 24, 2018.
  26. "Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification" (PDF). Proceedings of Machine Learning Research. 2018. Retrieved March 24, 2018.
  27. "Aspire Mirror". Aspire Mirror. Retrieved March 24, 2018.
  28. International, Youth Radio-- Youth Media (February 28, 2017). "A Search For 'Hidden Figures' Finds Joy". huffingtonpost.com. Huffington Post . Retrieved March 24, 2018.
  29. "Filmmakers Collaborative | Code4Rights". filmmakerscollab.org. Retrieved March 24, 2018.
  30. "Filmmakers Collaborative | Algorithmic Justice League: Unmasking Bias". filmmakerscollab.org. Retrieved March 24, 2018.
  31. Mar 2; Burt, 2020 | Chris (March 2, 2020). "Tech giants pressured to follow Google in removing gender labels from computer vision services". biometricupdate.com. Biometric Update. Retrieved March 9, 2020.{{cite web}}: CS1 maint: numeric names: authors list (link)
  32. Boak, Josh; Press, MATT O’BRIEN Associated (October 30, 2023). "Biden wants to move fast on AI safeguards and signs an executive order to address his concerns". Los Angeles Times. Retrieved January 22, 2024.
  33. "Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence". Federal Register. October 30, 2023.
  34. Twitter; Mastodon; Email (November 3, 2023). "Column: She set out to build robots. She ended up exposing big tech". Los Angeles Times. Retrieved January 22, 2024.{{cite web}}: |last= has generic name (help)
  35. "Mission, Team and Story - The Algorithmic Justice League". ajl.org. Retrieved May 9, 2021.
  36. 1 2 "Spotlight - Coded Bias Documentary". ajl.org. Retrieved May 9, 2021.
  37. Quach, Katyanna (2019). "We listened to more than 3 hours of US Congress testimony on facial recognition so you didn't have to go through it". theregister.com. Retrieved January 13, 2022.
  38. Voicing Erasure - A Spoken Word Piece Exploring Bias in Voice Recognition Technology , retrieved May 9, 2021
  39. "Voicing Erasure". ajl.org. Retrieved May 9, 2021.
  40. "The Coded Gaze: Unpacking Biases in Algorithms That Perpetuate Inequity". rockefellerfoundation.org. Retrieved May 15, 2021.
  41. "The Coded Gaze: Unpacking Biases in Algorithms That Perpetuate Inequity". The Rockefeller Foundation. Retrieved June 20, 2021.
  42. "Here's AOC calling out the vicious circle of white men building biased face AI". fastcompany.com. May 22, 2019. Retrieved May 15, 2021.
  43. "Coded Bias | Films | PBS". Independent Lens. Retrieved May 9, 2021.
  44. Trenholm, Richard (March 31, 2021). "Eye-opening documentary Coded Bias, streaming on Netflix April 5, faces racist technology". CNET .
  45. "Art and Film - The Algorithmic Justice League". ajl.org. Retrieved March 25, 2022.
  46. "apexart Exhibition: The Criminal Type". apexart.org.
  47. "Understanding AI". Ars Electronica Center.
  48. "AI: More than Human | Barbican". www.barbican.org.uk.
  49. "Nine Moments for Now". coopergallery.fas.harvard.edu.
  50. "Big Bang Data". MIT Museum.
  51. "Hidden No More: STEM Spotlight Shines On 'Hidden Figures' Like MIT's Joy Buolamwini". youthradio.org. Youth Radio. February 27, 2017. Retrieved March 24, 2018.
  52. ""Hidden Figures" Inspires A Scholarship Contest For Minority STEM Aspirants". fastcompany.com. Fast Company. January 19, 2017. Retrieved March 24, 2018.
  53. "Speaker Joy Buolamwini: How I'm Fighting Bias in Algorithms". scholar.harvard.edu. Retrieved March 24, 2018.
  54. Buolamwini, Joy. "How I'm fighting bias in algorithms – MIT Media Lab". MIT Media Lab. Retrieved March 24, 2018.
  55. TED (March 29, 2017), How I'm fighting bias in algorithms | Joy Buolamwini , retrieved March 24, 2018
  56. Joy Buolamwini: How Does Facial Recognition Software See Skin Color? , retrieved March 24, 2018
  57. Schwab, Katharine (July 3, 2018), "Meet 4 design heroes who are defending democracy online", fastcompany.com, Fast Company Magazine, retrieved July 21, 2018
  58. "BBC 100 Women 2018: Who is on the list?". BBC News. November 19, 2018. Retrieved November 21, 2018.
  59. "Joy Buolamwini". Fortune. Retrieved November 26, 2019.
  60. "TIME 100 Next 2019: Joy Buolamwini". time.com. Time magazine . Retrieved December 16, 2019.
  61. "She's Rewriting the Code". Off The Cuff. Retrieved March 9, 2020.
  62. Lee, Jennifer 8. (February 8, 2020). "New Documentary 'Coded Bias' Explores How Tech Can Be Racist And Sexist : Code Switch". npr.org. NPR . Retrieved December 12, 2020.{{cite web}}: CS1 maint: numeric names: authors list (link)
  63. "Hutchens MEdalists". January 4, 2024.
  64. "The 100 Most Influential People in AI 2023". Time. Retrieved February 18, 2024.