Title: An Interdisciplinary Approach to Understanding and Mitigating Harms Against Marginalized Groups in Intelligent Systems

Date: Tuesday, December 3, 2024

Time:  11:30am-1:30pm(Eastern Time)

Location (in-person): Coda 9th Floor, Room C0908 Home Park

Location (virtual): https://gatech.zoom.us/j/98489741017?pwd=AcLSKTibcaMckqZFWc5AmUfMzB7nn2.1

 

 

Camille Harris

School of Interactive Computing

College of Computing

Georgia Tech

 

Committee

Dr. Diyi Yang (Advisor) - School of Computer Science, Stanford University

Dr. Neha Kumar (Advisor) - School of Interactive Computing, Georgia Tech

Dr. Munmun De Choudhury - School of Interactive Computing, Georgia Tech

Dr. Kartik Goyal - School of Interactive Computing, Georgia Tech

Dr. Andre Brock - School of Literature Media and Communication , Georgia Tech

Dr. Christina Harrington - Human Computer Interaction Institute, Carnegie Mellon University 

 

Abstract

 

As AI systems become increasingly adopted in new domains and a greater part of everyday life, examining, understanding, and ultimately preventing these systems from replicate existing societal harms of sexism, racism, and other oppressive systems is an increasing concern. Within a United States context, these issues pose a large potential to harm to groups historically oppressed marginalized groups such as Black/African people and Indigenous people, and other People of Color (BIPOC). Within Natural Language Processing (NLP), dialects of English spoken by minority populations such as African American English and Chicano English that have distinct grammar, vocabulary and other linguistic differences from White Mainstream English (WME) also called Standard American English (SAE). While most works that explore such biases concerns in NLP or other machine learning (ML) bias and fairness evaluations focus on the model level, most marginalized users who experience the impacts of such systems experience it at the level of the downstream application, where multiple other components including the context of application, the party creating and implementing the model, the user interface, etc. all impact the end user in addition to model performance issues alone. For this reason, my work focuses on downstream applications of intelligent systems, with an emphasis on NLP systems, and the harms of these applications towards marginalized groups in the United States. Further, my work engages traditional machine learning fairness and NLP methods to better quantify harmful outcomes against marginalized groups with these technologies, along with qualitative methodologies which provide an analysis that centers the lived experiences of marginalized groups.