Professional Documents
Culture Documents
me@hal3.name
By surveying 146 papers analyzing “bias” in NLP Sandeep Attree. 2019. Gendered ambiguous pronouns
shared task: Boosting model confidence by evidence
systems, we found that (a) their motivations are pooling. In Proceedings of the Workshop on Gen-
often vague, inconsistent, and lacking in norma- der Bias in Natural Language Processing, Florence,
tive reasoning; and (b) their proposed quantitative Italy.
techniques for measuring or mitigating “bias” are Pinkesh Badjatiya, Manish Gupta, and Vasudeva
poorly matched to their motivations and do not en- Varma. 2019. Stereotypical bias removal for hate
gage with the relevant literature outside of NLP. speech detection task using knowledge-based gen-
To help researchers and practitioners avoid these eralizations. In Proceedings of the International
World Wide Web Conference, pages 49–59, San Fran-
pitfalls, we proposed three recommendations that cisco, CA.
should guide work analyzing “bias” in NLP sys-
tems, and, for each, provided several concrete re- Eugene Bagdasaryan, Omid Poursaeed, and Vitaly
Shmatikov. 2019. Differential Privacy Has Dis-
search questions. These recommendations rest on parate Impact on Model Accuracy. In Proceedings
a greater recognition of the relationships between of the Conference on Neural Information Processing
language and social hierarchies—a step that we Systems, Vancouver, Canada.
see as paramount to establishing a path forward. April Baker-Bell. 2019. Dismantling anti-black lin-
guistic racism in English language arts classrooms:
Acknowledgments Toward an anti-racist black language pedagogy. The-
ory Into Practice.
This paper is based upon work supported by the
National Science Foundation Graduate Research David Bamman, Sejal Popat, and Sheng Shen. 2019.
Fellowship under Grant No. 1451512. Any opin- An annotated dataset of literary entities. In Proceed-
ings of the North American Association for Com-
ion, findings, and conclusions or recommenda- putational Linguistics (NAACL), pages 2138–2144,
tions expressed in this material are those of the Minneapolis, MN.
authors and do not necessarily reflect the views of
Xingce Bao and Qianqian Qiao. 2019. Transfer Learn-
the National Science Foundation. We thank the ing from Pre-trained BERT for Pronoun Resolution.
reviewers for their useful feedback, especially the In Proceedings of the Workshop on Gender Bias
suggestion to include additional details about our in Natural Language Processing, pages 82–88, Flo-
method. rence, Italy.
Shaowen Bardzell and Jeffrey Bardzell. 2011. Towards
a Feminist HCI Methodology: Social Science, Femi-
References nism, and HCI. In Proceedings of the Conference on
Human Factors in Computing Systems (CHI), pages
Artem Abzaliev. 2019. On GAP coreference resolu- 675–684, Vancouver, Canada.
tion shared task: insights from the 3rd place solution.
In Proceedings of the Workshop on Gender Bias in Solon Barocas, Asia J. Biega, Benjamin Fish, J˛edrzej
Natural Language Processing, pages 107–112, Flo- Niklas, and Luke Stark. 2020. When Not to De-
rence, Italy. sign, Build, or Deploy. In Proceedings of the Confer-
ence on Fairness, Accountability, and Transparency,
ADA. 2018. Guidelines for Writing About Peo- Barcelona, Spain.
ple With Disabilities. ADA National Network.
https://bit.ly/2KREbkB. Solon Barocas, Kate Crawford, Aaron Shapiro, and
Hanna Wallach. 2017. The Problem With Bias: Al-
Oshin Agarwal, Funda Durupinar, Norman I. Badler, locative Versus Representational Harms in Machine
and Ani Nenkova. 2019. Word embeddings (also) Learning. In Proceedings of SIGCIS, Philadelphia,
encode human personality stereotypes. In Proceed- PA.
ings of the Joint Conference on Lexical and Com-
putational Semantics, pages 205–211, Minneapolis, Christine Basta, Marta R. Costa-jussà, and Noe Casas.
MN. 2019. Evaluating the underlying gender bias in con-
textualized word embeddings. In Proceedings of
H. Samy Alim. 2004. You Know My Steez: An Ethno- the Workshop on Gender Bias for Natural Language
graphic and Sociolinguistic Study of Styleshifting in Processing, pages 33–39, Florence, Italy.
John Baugh. 2018. Linguistics in Pursuit of Justice. Homemaker? Debiasing Word Embeddings. In Pro-
Cambridge University Press. ceedings of the Conference on Neural Information
Processing Systems, pages 4349–4357, Barcelona,
Emily M. Bender. 2019. A typology of ethi- Spain.
cal risks in language technology with an eye to-
wards where transparent documentation can help. Tolga Bolukbasi, Kai-Wei Chang, James Zou,
Presented at The Future of Artificial Intelli- Venkatesh Saligrama, and Adam Kalai. 2016b.
gence: Language, Ethics, Technology Workshop. Quantifying and reducing stereotypes in word
https://bit.ly/2P9t9M6. embeddings. In Proceedings of the ICML Workshop
on #Data4Good: Machine Learning in Social Good
Ruha Benjamin. 2019. Race After Technology: Aboli- Applications, pages 41–45, New York, NY.
tionist Tools for the New Jim Code. John Wiley &
Sons. Shikha Bordia and Samuel R. Bowman. 2019. Identify-
ing and reducing gender bias in word-level language
Ruha Benjamin. 2020. 2020 Vision: Reimagining the
models. In Proceedings of the NAACL Student Re-
Default Settings of Technology & Society. Keynote
search Workshop, pages 7–15, Minneapolis, MN.
at ICLR.
Cynthia L. Bennett and Os Keyes. 2019. What is the Marc-Etienne Brunet, Colleen Alkalay-Houlihan, Ash-
Point of Fairness? Disability, AI, and The Com- ton Anderson, and Richard Zemel. 2019. Under-
plexity of Justice. In Proceedings of the ASSETS standing the Origins of Bias in Word Embeddings.
Workshop on AI Fairness for People with Disabili- In Proceedings of the International Conference on
ties, Pittsburgh, PA. Machine Learning, pages 803–811, Long Beach,
CA.
Camiel J. Beukeboom and Christian Burgers. 2019.
How Stereotypes Are Shared Through Language: A Mary Bucholtz, Dolores Inés Casillas, and Jin Sook
Review and Introduction of the Social Categories Lee. 2016. Beyond Empowerment: Accompani-
and Stereotypes Communication (SCSC) Frame- ment and Sociolinguistic Justice in a Youth Research
work. Review of Communication Research, 7:1–37. Program. In Robert Lawson and Dave Sayers, edi-
tors, Sociolinguistic Research: Application and Im-
Shruti Bhargava and David Forsyth. 2019. Expos- pact, pages 25–44. Routledge.
ing and Correcting the Gender Bias in Image
Captioning Datasets and Models. arXiv preprint Mary Bucholtz, Dolores Inés Casillas, and Jin Sook
arXiv:1912.00578. Lee. 2019. California Latinx Youth as Agents of
Sociolinguistic Justice. In Netta Avineri, Laura R.
Jayadev Bhaskaran and Isha Bhallamudi. 2019. Good Graham, Eric J. Johnson, Robin Conley Riner, and
Secretaries, Bad Truck Drivers? Occupational Gen- Jonathan Rosa, editors, Language and Social Justice
der Stereotypes in Sentiment Analysis. In Proceed- in Practice, pages 166–175. Routledge.
ings of the Workshop on Gender Bias in Natural Lan-
guage Processing, pages 62–68, Florence, Italy. Mary Bucholtz, Audrey Lopez, Allina Mojarro, Elena
Skapoulli, Chris VanderStouwe, and Shawn Warner-
Su Lin Blodgett, Lisa Green, and Brendan O’Connor. Garcia. 2014. Sociolinguistic Justice in the Schools:
2016. Demographic Dialectal Variation in Social Student Researchers as Linguistic Experts. Lan-
Media: A Case Study of African-American English. guage and Linguistics Compass, 8:144–157.
In Proceedings of Empirical Methods in Natural
Language Processing (EMNLP), pages 1119–1130, Kaylee Burns, Lisa Anne Hendricks, Kate Saenko,
Austin, TX. Trevor Darrell, and Anna Rohrbach. 2018. Women
also Snowboard: Overcoming Bias in Captioning
Su Lin Blodgett and Brendan O’Connor. 2017. Racial
Models. In Procedings of the European Conference
Disparity in Natural Language Processing: A Case
on Computer Vision (ECCV), pages 793–811, Mu-
Study of Social Media African-American English.
nich, Germany.
In Proceedings of the Workshop on Fairness, Ac-
countability, and Transparency in Machine Learning
(FAT/ML), Halifax, Canada. Aylin Caliskan, Joanna J. Bryson, and Arvind
Narayanan. 2017. Semantics derived automatically
Su Lin Blodgett, Johnny Wei, and Brendan O’Connor. from language corpora contain human-like biases.
2018. Twitter Universal Dependency Parsing for Science, 356(6334).
African-American and Mainstream American En-
glish. In Proceedings of the Association for Compu- Kathryn Campbell-Kibler. 2009. The nature of so-
tational Linguistics (ACL), pages 1415–1425, Mel- ciolinguistic perception. Language Variation and
bourne, Australia. Change, 21(1):135–156.
Tolga Bolukbasi, Kai-Wei Chang, James Zou, Yang Trista Cao and Hal Daumé, III. 2019. To-
Venkatesh Saligrama, and Adam Kalai. 2016a. ward gender-inclusive coreference resolution. arXiv
Man is to Computer Programmer as Woman is to preprint arXiv:1910.13913.
Rakesh Chada. 2019. Gendered pronoun resolution us- Thomas Davidson, Debasmita Bhattacharya, and Ing-
ing bert and an extractive question answering formu- mar Weber. 2019. Racial bias in hate speech and
lation. In Proceedings of the Workshop on Gender abusive language detection datasets. In Proceedings
Bias in Natural Language Processing, pages 126– of the Workshop on Abusive Language Online, pages
133, Florence, Italy. 25–35, Florence, Italy.
Kaytlin Chaloner and Alfredo Maldonado. 2019. Mea- Maria De-Arteaga, Alexey Romanov, Hanna Wal-
suring Gender Bias in Word Embedding across Do- lach, Jennifer Chayes, Christian Borgs, Alexandra
mains and Discovering New Gender Bias Word Cat- Chouldechova, Sahin Geyik, Krishnaram Kentha-
egories. In Proceedings of the Workshop on Gender padi, and Adam Tauman Kalai. 2019. Bias in bios:
Bias in Natural Language Processing, pages 25–32, A case study of semantic representation bias in a
Florence, Italy. high-stakes setting. In Proceedings of the Confer-
ence on Fairness, Accountability, and Transparency,
Anne H. Charity Hudley. 2017. Language and Racial-
pages 120–128, Atlanta, GA.
ization. In Ofelia García, Nelson Flores, and Mas-
similiano Spotti, editors, The Oxford Handbook of Sunipa Dev, Tao Li, Jeff Phillips, and Vivek Sriku-
Language and Society. Oxford University Press. mar. 2019. On Measuring and Mitigating Biased
Won Ik Cho, Ji Won Kim, Seok Min Kim, and Inferences of Word Embeddings. arXiv preprint
Nam Soo Kim. 2019. On measuring gender bias in arXiv:1908.09369.
translation of gender-neutral pronouns. In Proceed-
Sunipa Dev and Jeff Phillips. 2019. Attenuating Bias in
ings of the Workshop on Gender Bias in Natural Lan-
Word Vectors. In Proceedings of the International
guage Processing, pages 173–181, Florence, Italy.
Conference on Artificial Intelligence and Statistics,
Shivang Chopra, Ramit Sawhney, Puneet Mathur, and pages 879–887, Naha, Japan.
Rajiv Ratn Shah. 2020. Hindi-English Hate Speech
Detection: Author Profiling, Debiasing, and Practi- Mark Díaz, Isaac Johnson, Amanda Lazar, Anne Marie
cal Perspectives. In Proceedings of the AAAI Con- Piper, and Darren Gergle. 2018. Addressing age-
ference on Artificial Intelligence (AAAI), New York, related bias in sentiment analysis. In Proceedings
NY. of the Conference on Human Factors in Computing
Systems (CHI), Montréal, Canada.
Marika Cifor, Patricia Garcia, T.L. Cowan, Jas-
mine Rault, Tonia Sutherland, Anita Say Emily Dinan, Angela Fan, Adina Williams, Jack Ur-
Chan, Jennifer Rode, Anna Lauren Hoffmann, banek, Douwe Kiela, and Jason Weston. 2019.
Niloufar Salehi, and Lisa Nakamura. 2019. Queens are Powerful too: Mitigating Gender
Feminist Data Manifest-No. Retrieved from Bias in Dialogue Generation. arXiv preprint
https://www.manifestno.com/. arXiv:1911.03842.
Patricia Hill Collins. 2000. Black Feminist Thought: Carl DiSalvo, Andrew Clement, and Volkmar Pipek.
Knowledge, Consciousness, and the Politics of Em- 2013. Communities: Participatory Design for, with
powerment. Routledge. and by communities. In Jesper Simonsen and Toni
Robertson, editors, Routledge International Hand-
Justin T. Craft, Kelly E. Wright, Rachel Elizabeth book of Participatory Design, pages 182–209. Rout-
Weissler, and Robin M. Queen. 2020. Language ledge.
and Discrimination: Generating Meaning, Perceiv-
ing Identities, and Discriminating Outcomes. An- Lucas Dixon, John Li, Jeffrey Sorensen, Nithum Thain,
nual Review of Linguistics, 6(1). and Lucy Vasserman. 2018. Measuring and mitigat-
ing unintended bias in text classification. In Pro-
Kate Crawford. 2017. The Trouble with Bias. Keynote
ceedings of the Conference on Artificial Intelligence,
at NeurIPS.
Ethics, and Society (AIES), New Orleans, LA.
Kimberle Crenshaw. 1989. Demarginalizing the Inter-
section of Race and Sex: A Black Feminist Critique Jacob Eisenstein. 2013. What to do about bad lan-
of Antidiscrmination Doctrine, Feminist Theory and guage on the Internet. In Proceedings of the North
Antiracist Politics. University of Chicago Legal Fo- American Association for Computational Linguistics
rum. (NAACL), pages 359–369.
Amanda Cercas Curry and Verena Rieser. 2018. Kawin Ethayarajh. 2020. Is Your Classifier Actually
#MeToo: How Conversational Systems Respond to Biased? Measuring Fairness under Uncertainty with
Sexual Harassment. In Proceedings of the Workshop Bernstein Bounds. In Proceedings of the Associa-
on Ethics in Natural Language Processing, pages 7– tion for Computational Linguistics (ACL).
14, New Orleans, LA.
Kawin Ethayarajh, David Duvenaud, and Graeme Hirst.
Karan Dabas, Nishtha Madaan, Gautam Singh, Vi- 2019. Understanding Undesirable Word Embedding
jay Arya, Sameep Mehta, and Tanmoy Chakraborty. Assocations. In Proceedings of the Association for
2020. Fair Transfer of Multiple Style Attributes in Computational Linguistics (ACL), pages 1696–1705,
Text. arXiv preprint arXiv:2001.06693. Florence, Italy.
Joseph Fisher. 2019. Measuring social bias in Oguzhan Gencoglu. 2020. Cyberbullying Detec-
knowledge graph embeddings. arXiv preprint tion with Fairness Constraints. arXiv preprint
arXiv:1912.02761. arXiv:2005.06625.
Nelson Flores and Sofia Chaparro. 2018. What counts Alexandra Reeve Givens and Meredith Ringel Morris.
as language education policy? Developing a materi- 2020. Centering Disability Perspecives in Algorith-
alist Anti-racist approach to language activism. Lan- mic Fairness, Accountability, and Transparency. In
guage Policy, 17(3):365–384. Proceedings of the Conference on Fairness, Account-
ability, and Transparency, Barcelona, Spain.
Omar U. Florez. 2019. On the Unintended Social Bias
of Training Language Generation Models with Data Hila Gonen and Yoav Goldberg. 2019. Lipstick on a
from Local Media. In Proceedings of the NeurIPS Pig: Debiasing Methods Cover up Systematic Gen-
Workshop on Human-Centric Machine Learning, der Biases in Word Embeddings But do not Remove
Vancouver, Canada. Them. In Proceedings of the North American As-
Joel Escudé Font and Marta R. Costa-jussà. 2019. sociation for Computational Linguistics (NAACL),
Equalizing gender biases in neural machine trans- pages 609–614, Minneapolis, MN.
lation with word embeddings techniques. In Pro- Hila Gonen and Kellie Webster. 2020. Auto-
ceedings of the Workshop on Gender Bias for Natu- matically Identifying Gender Issues in Machine
ral Language Processing, pages 147–154, Florence, Translation using Perturbations. arXiv preprint
Italy. arXiv:2004.14065.
Batya Friedman and David G. Hendry. 2019. Value
Ben Green. 2019. “Good” isn’t good enough. In Pro-
Sensitive Design: Shaping Technology with Moral
ceedings of the AI for Social Good Workshop, Van-
Imagination. MIT Press.
couver, Canada.
Batya Friedman, Peter H. Kahn Jr., and Alan Borning.
2006. Value Sensitive Design and Information Sys- Lisa J. Green. 2002. African American English: A Lin-
tems. In Dennis Galletta and Ping Zhang, editors, guistic Introduction. Cambridge University Press.
Human-Computer Interaction in Management Infor-
Anthony G. Greenwald, Debbie E. McGhee, and Jor-
mation Systems: Foundations, pages 348–372. M.E.
dan L.K. Schwartz. 1998. Measuring individual dif-
Sharpe.
ferences in implicit cognition: The implicit associa-
Nikhil Garg, Londa Schiebinger, Dan Jurafsky, and tion test. Journal of Personality and Social Psychol-
James Zou. 2018. Word Embeddings Quantify 100 ogy, 74(6):1464–1480.
Years of Gender and Ethnic Stereotypes. Proceed-
ings of the National Academy of Sciences, 115(16). Enoch Opanin Gyamfi, Yunbo Rao, Miao Gou, and
Yanhua Shao. 2020. deb2viz: Debiasing gender in
Sahaj Garg, Vincent Perot, Nicole Limtiaco, Ankur word embedding data using subspace visualization.
Taly, Ed H. Chi, and Alex Beutel. 2019. Counter- In Proceedings of the International Conference on
factual fairness in text classification through robust- Graphics and Image Processing.
ness. In Proceedings of the Conference on Artificial
Intelligence, Ethics, and Society (AIES), Honolulu, Foad Hamidi, Morgan Klaus Scheuerman, and Stacy M.
HI. Branham. 2018. Gender Recognition or Gender Re-
ductionism? The Social Implications of Automatic
Aparna Garimella, Carmen Banea, Dirk Hovy, and Gender Recognition Systems. In Proceedings of the
Rada Mihalcea. 2019. Women’s syntactic resilience Conference on Human Factors in Computing Sys-
and men’s grammatical luck: Gender bias in part-of- tems (CHI), Montréal, Canada.
speech tagging and dependency parsing data. In Pro-
ceedings of the Association for Computational Lin- Alex Hanna, Emily Denton, Andrew Smart, and Jamila
guistics (ACL), pages 3493–3498, Florence, Italy. Smith-Loud. 2020. Towards a Critical Race Method-
ology in Algorithmic Fairness. In Proceedings of the
Andrew Gaut, Tony Sun, Shirlyn Tang, Yuxin Huang, Conference on Fairness, Accountability, and Trans-
Jing Qian, Mai ElSherief, Jieyu Zhao, Diba parency, pages 501–512, Barcelona, Spain.
Mirza, Elizabeth Belding, Kai-Wei Chang, and
William Yang Wang. 2020. Towards Understand- Madeline E. Heilman, Aaaron S. Wallen, Daniella
ing Gender Bias in Relation Extraction. In Proceed- Fuchs, and Melinda M. Tamkins. 2004. Penalties
ings of the Association for Computational Linguis- for Success: Reactions to Women Who Succeed at
tics (ACL). Male Gender-Typed Tasks. Journal of Applied Psy-
chology, 89(3):416–427.
R. Stuart Geiger, Kevin Yu, Yanlai Yang, Mindy Dai,
Jie Qiu, Rebekah Tang, and Jenny Huang. 2020. Jane H. Hill. 2008. The Everyday Language of White
Garbage In, Garbage Out? Do Machine Learn- Racism. Wiley-Blackwell.
ing Application Papers in Social Computing Report
Where Human-Labeled Training Data Comes From? Dirk Hovy, Federico Bianchi, and Tommaso Fornaciari.
In Proceedings of the Conference on Fairness, Ac- 2020. Can You Translate that into Man? Commer-
countability, and Transparency, pages 325–336. cial Machine Translation Systems Include Stylistic
Biases. In Proceedings of the Association for Com- Anna Jørgensen, Dirk Hovy, and Anders Søgaard. 2015.
putational Linguistics (ACL). Challenges of studying and processing dialects in
social media. In Proceedings of the Workshop on
Dirk Hovy and Anders Søgaard. 2015. Tagging Per- Noisy User-Generated Text, pages 9–18, Beijing,
formance Correlates with Author Age. In Proceed- China.
ings of the Association for Computational Linguis-
tics and the International Joint Conference on Nat- Anna Jørgensen, Dirk Hovy, and Anders Søgaard. 2016.
ural Language Processing, pages 483–488, Beijing, Learning a POS tagger for AAVE-like language. In
China. Proceedings of the North American Association for
Computational Linguistics (NAACL), pages 1115–
Dirk Hovy and Shannon L. Spruit. 2016. The social 1120, San Diego, CA.
impact of natural language processing. In Proceed-
ings of the Association for Computational Linguis- Pratik Joshi, Sebastian Santy, Amar Budhiraja, Kalika
tics (ACL), pages 591–598, Berlin, Germany. Bali, and Monojit Choudhury. 2020. The State and
Fate of Linguistic Diversity and Inclusion in the
Po-Sen Huang, Huan Zhang, Ray Jiang, Robert NLP World. In Proceedings of the Association for
Stanforth, Johannes Welbl, Jack W. Rae, Vishal Computational Linguistics (ACL).
Maini, Dani Yogatama, and Pushmeet Kohli. 2019.
Reducing Sentiment Bias in Language Models Jaap Jumelet, Willem Zuidema, and Dieuwke Hupkes.
via Counterfactual Evaluation. arXiv preprint 2019. Analysing Neural Language Models: Contex-
arXiv:1911.03064. tual Decomposition Reveals Default Reasoning in
Number and Gender Assignment. In Proceedings
Xiaolei Huang, Linzi Xing, Franck Dernoncourt, and of the Conference on Natural Language Learning,
Michael J. Paul. 2020. Multilingual Twitter Cor- Hong Kong, China.
pus and Baselines for Evaluating Demographic Bias
in Hate Speech Recognition. In Proceedings of Marie-Odile Junker. 2018. Participatory action re-
the Language Resources and Evaluation Conference search for Indigenous linguistics in the digital age.
(LREC), Marseille, France. In Shannon T. Bischoff and Carmen Jany, editors,
Insights from Practices in Community-Based Re-
Christoph Hube, Maximilian Idahl, and Besnik Fetahu. search, pages 164–175. De Gruyter Mouton.
2020. Debiasing Word Embeddings from Sentiment
David Jurgens, Yulia Tsvetkov, and Dan Jurafsky. 2017.
Associations in Names. In Proceedings of the Inter-
Incorporating Dialectal Variability for Socially Equi-
national Conference on Web Search and Data Min-
table Language Identification. In Proceedings of the
ing, pages 259–267, Houston, TX.
Association for Computational Linguistics (ACL),
pages 51–57, Vancouver, Canada.
Ben Hutchinson, Vinodkumar Prabhakaran, Emily
Denton, Kellie Webster, Yu Zhong, and Stephen De- Masahiro Kaneko and Danushka Bollegala. 2019.
nuyl. 2020. Social Biases in NLP Models as Barriers Gender-preserving debiasing for pre-trained word
for Persons with Disabilities. In Proceedings of the embeddings. In Proceedings of the Association for
Association for Computational Linguistics (ACL). Computational Linguistics (ACL), pages 1641–1650,
Florence, Italy.
Matei Ionita, Yury Kashnitsky, Ken Krige, Vladimir
Larin, Dennis Logvinenko, and Atanas Atanasov. Saket Karve, Lyle Ungar, and João Sedoc. 2019. Con-
2019. Resolving gendered ambiguous pronouns ceptor debiasing of word representations evaluated
with BERT. In Proceedings of the Workshop on on WEAT. In Proceedings of the Workshop on Gen-
Gender Bias in Natural Language Processing, pages der Bias in Natural Language Processing, pages 40–
113–119, Florence, Italy. 48, Florence, Italy.
Hailey James-Sorenson and David Alvarez-Melis. Michael Katell, Meg Young, Dharma Dailey, Bernease
2019. Probabilistic Bias Mitigation in Word Embed- Herman, Vivian Guetler, Aaron Tam, Corinne Bintz,
dings. In Proceedings of the Workshop on Human- Danielle Raz, and P.M. Krafft. 2020. Toward sit-
Centric Machine Learning, Vancouver, Canada. uated interventions for algorithmic equity: lessons
from the field. In Proceedings of the Conference on
Shengyu Jia, Tao Meng, Jieyu Zhao, and Kai-Wei Fairness, Accountability, and Transparency, pages
Chang. 2020. Mitigating Gender Bias Amplification 45–55, Barcelona, Spain.
in Distribution by Posterior Regularization. In Pro-
ceedings of the Association for Computational Lin- Stephen Kemmis. 2006. Participatory action research
guistics (ACL). and the public sphere. Educational Action Research,
14(4):459–476.
Taylor Jones, Jessica Rose Kalbfeld, Ryan Hancock,
and Robin Clark. 2019. Testifying while black: An Os Keyes. 2018. The Misgendering Machines:
experimental study of court reporter accuracy in tran- Trans/HCI Implications of Automatic Gender
scription of African American English. Language, Recognition. Proceedings of the ACM on Human-
95(2). Computer Interaction, 2(CSCW).
Os Keyes, Josephine Hoy, and Margaret Drouhard. Paul Pu Liang, Irene Li, Emily Zheng, Yao Chong Lim,
2019. Human-Computer Insurrection: Notes on an Ruslan Salakhutdinov, and Louis-Philippe Morency.
Anarchist HCI. In Proceedings of the Conference on 2019. Towards Debiasing Sentence Representations.
Human Factors in Computing Systems (CHI), Glas- In Proceedings of the NeurIPS Workshop on Human-
gow, Scotland, UK. Centric Machine Learning, Vancouver, Canada.
Jae Yeon Kim, Carlos Ortiz, Sarah Nam, Sarah Santi- Rosina Lippi-Green. 2012. English with an Ac-
ago, and Vivek Datta. 2020. Intersectional Bias in cent: Language, Ideology, and Discrimination in the
Hate Speech and Abusive Language Datasets. In United States. Routledge.
Proceedings of the Association for Computational
Linguistics (ACL).
Bo Liu. 2019. Anonymized BERT: An Augmentation
Svetlana Kiritchenko and Saif M. Mohammad. 2018. Approach to the Gendered Pronoun Resolution Chal-
Examining Gender and Race Bias in Two Hundred lenge. In Proceedings of the Workshop on Gender
Sentiment Analysis Systems. In Proceedings of the Bias in Natural Language Processing, pages 120–
Joint Conference on Lexical and Computational Se- 125, Florence, Italy.
mantics, pages 43–53, New Orleans, LA.
Haochen Liu, Jamell Dacon, Wenqi Fan, Hui Liu, Zi-
Moshe Koppel, Shlomo Argamon, and Anat Rachel tao Liu, and Jiliang Tang. 2019. Does Gender Mat-
Shimoni. 2002. Automatically Categorizing Writ- ter? Towards Fairness in Dialogue Systems. arXiv
ten Texts by Author Gender. Literary and Linguistic preprint arXiv:1910.10486.
Computing, 17(4):401–412.
Felipe Alfaro Lois, José A.R. Fonollosa, and Costa-jà.
Keita Kurita, Nidhi Vyas, Ayush Pareek, Alan W.
2019. BERT Masked Language Modeling for Co-
Black, and Yulia Tsvetkov. 2019. Measuring bias
reference Resolution. In Proceedings of the Work-
in contextualized word representations. In Proceed-
shop on Gender Bias in Natural Language Process-
ings of the Workshop on Gender Bias for Natu-
ing, pages 76–81, Florence, Italy.
ral Language Processing, pages 166–172, Florence,
Italy.
Brandon C. Loudermilk. 2015. Implicit attitudes and
Sonja L. Lanehart and Ayesha M. Malik. 2018. Black the perception of sociolinguistic variation. In Alexei
Is, Black Isn’t: Perceptions of Language and Black- Prikhodkine and Dennis R. Preston, editors, Re-
ness. In Jeffrey Reaser, Eric Wilbanks, Karissa Woj- sponses to Language Varieties: Variability, pro-
cik, and Walt Wolfram, editors, Language Variety in cesses and outcomes, pages 137–156.
the New South. University of North Carolina Press.
Anastassia Loukina, Nitin Madnani, and Klaus Zech-
Brian N. Larson. 2017. Gender as a variable in natural- ner. 2019. The many dimensions of algorithmic fair-
language processing: Ethical considerations. In Pro- ness in educational applications. In Proceedings of
ceedings of the Workshop on Ethics in Natural Lan- the Workshop on Innovative Use of NLP for Build-
guage Processing, pages 30–40, Valencia, Spain. ing Educational Applications, pages 1–10, Florence,
Italy.
Anne Lauscher and Goran Glavaš. 2019. Are We Con-
sistently Biased? Multidimensional Analysis of Bi-
ases in Distributional Word Vectors. In Proceedings Kaiji Lu, Peter Mardziel, Fangjing Wu, Preetam Aman-
of the Joint Conference on Lexical and Computa- charla, and Anupam Datta. 2018. Gender bias in
tional Semantics, pages 85–91, Minneapolis, MN. neural natural language processing. arXiv preprint
arXiv:1807.11714.
Anne Lauscher, Goran Glavaš, Simone Paolo Ponzetto,
and Ivan Vulić. 2019. A General Framework for Im- Anne Maass. 1999. Linguistic intergroup bias: Stereo-
plicit and Explicit Debiasing of Distributional Word type perpetuation through language. Advances in
Vector Spaces. arXiv preprint arXiv:1909.06092. Experimental Social Psychology, 31:79–121.
Christopher A. Le Dantec, Erika Shehan Poole, and Su- Nitin Madnani, Anastassia Loukina, Alina von Davier,
san P. Wyche. 2009. Values as Lived Experience: Jill Burstein, and Aoife Cahill. 2017. Building Bet-
Evolving Value Sensitive Design in Support of Value ter Open-Source Tools to Support Fairness in Auto-
Discovery. In Proceedings of the Conference on Hu- mated Scoring. In Proceedings of the Workshop on
man Factors in Computing Systems (CHI), Boston, Ethics in Natural Language Processing, pages 41–
MA. 52, Valencia, Spain.
Nayeon Lee, Andrea Madotto, and Pascale Fung. 2019.
Exploring Social Bias in Chatbots using Stereotype Thomas Manzini, Yao Chong Lim, Yulia Tsvetkov, and
Knowledge. In Proceedings of the Workshop on Alan W. Black. 2019. Black is to Criminal as Cau-
Widening NLP, pages 177–180, Florence, Italy. casian is to Police: Detecting and Removing Multi-
class Bias in Word Embeddings. In Proceedings of
Wesley Y. Leonard. 2012. Reframing language recla- the North American Association for Computational
mation programmes for everybody’s empowerment. Linguistics (NAACL), pages 801–809, Minneapolis,
Gender and Language, 6(2):339–367. MN.
Ramón Antonio Martínez and Alexander Feliciano Dong Nguyen, Rilana Gravel, Dolf Trieschnigg, and
Mejía. 2019. Looking closely and listening care- Theo Meder. 2013. “How Old Do You Think I
fully: A sociocultural approach to understanding Am?”: A Study of Language and Age in Twitter. In
the complexity of Latina/o/x students’ everyday lan- Proceedings of the Conference on Web and Social
guage. Theory Into Practice. Media (ICWSM), pages 439–448, Boston, MA.
Rowan Hall Maudslay, Hila Gonen, Ryan Cotterell, Malvina Nissim, Rik van Noord, and Rob van der Goot.
and Simone Teufel. 2019. It’s All in the Name: Mit- 2020. Fair is better than sensational: Man is to doc-
igating Gender Bias with Name-Based Counterfac- tor as woman is to doctor. Computational Linguis-
tual Data Substitution. In Proceedings of Empirical tics.
Methods in Natural Language Processing (EMNLP), Debora Nozza, Claudia Volpetti, and Elisabetta Fersini.
pages 5270–5278, Hong Kong, China. 2019. Unintended Bias in Misogyny Detection. In
Proceedings of the Conference on Web Intelligence,
Chandler May, Alex Wang, Shikha Bordia, Samuel R. pages 149–155.
Bowman, and Rachel Rudinger. 2019. On Measur-
ing Social Biases in Sentence Encoders. In Proceed- Alexandra Olteanu, Carlos Castillo, Fernando Diaz,
ings of the North American Association for Compu- and Emre Kıcıman. 2019. Social Data: Biases,
tational Linguistics (NAACL), pages 629–634, Min- Methodological Pitfalls, and Ethical Boundaries.
neapolis, MN. Frontiers in Big Data, 2.
Elijah Mayfield, Michael Madaio, Shrimai Prab- Alexandra Olteanu, Kartik Talamadupula, and Kush R.
humoye, David Gerritsen, Brittany McLaughlin, Varshney. 2017. The Limits of Abstract Evaluation
Ezekiel Dixon-Roman, and Alan W. Black. 2019. Metrics: The Case of Hate Speech Detection. In
Equity Beyond Bias in Language Technologies for Proceedings of the ACM Web Science Conference,
Education. In Proceedings of the Workshop on Inno- Troy, NY.
vative Use of NLP for Building Educational Appli- Orestis Papakyriakopoulos, Simon Hegelich, Juan Car-
cations, Florence, Italy. los Medina Serrano, and Fabienne Marco. 2020.
Bias in word embeddings. In Proceedings of the
Katherine McCurdy and Oğuz Serbetçi. 2017. Gram- Conference on Fairness, Accountability, and Trans-
matical gender associations outweigh topical gender parency, pages 446–457, Barcelona, Spain.
bias in crosslinguistic word embeddings. In Pro-
ceedings of the Workshop for Women & Underrepre- Ji Ho Park, Jamin Shin, and Pascale Fung. 2018. Re-
sented Minorities in Natural Language Processing, ducing Gender Bias in Abusive Language Detection.
Vancouver, Canada. In Proceedings of Empirical Methods in Natural
Language Processing (EMNLP), pages 2799–2804,
Ninareh Mehrabi, Thamme Gowda, Fred Morstatter, Brussels, Belgium.
Nanyun Peng, and Aram Galstyan. 2019. Man is to
Person as Woman is to Location: Measuring Gender Ellie Pavlick and Tom Kwiatkowski. 2019. Inherent
Bias in Named Entity Recognition. arXiv preprint Disagreements in Human Textual Inferences. Trans-
arXiv:1910.10872. actions of the Association for Computational Lin-
guistics, 7:677–694.
Michela Menegatti and Monica Rubini. 2017. Gender Xiangyu Peng, Siyan Li, Spencer Frazier, and Mark
bias and sexism in language. In Oxford Research Riedl. 2020. Fine-Tuning a Transformer-Based Lan-
Encyclopedia of Communication. Oxford University guage Model to Avoid Generating Non-Normative
Press. Text. arXiv preprint arXiv:2001.08764.
Inom Mirzaev, Anthony Schulte, Michael Conover, and Radomir Popović, Florian Lemmerich, and Markus
Sam Shah. 2019. Considerations for the interpreta- Strohmaier. 2020. Joint Multiclass Debiasing of
tion of bias measures of word embeddings. arXiv Word Embeddings. In Proceedings of the Interna-
preprint arXiv:1906.08379. tional Symposium on Intelligent Systems, Graz, Aus-
tria.
Salikoko S. Mufwene, Guy Bailey, and John R. Rick-
Vinodkumar Prabhakaran, Ben Hutchinson, and Mar-
ford, editors. 1998. African-American English:
garet Mitchell. 2019. Perturbation Sensitivity Anal-
Structure, History, and Use. Routledge.
ysis to Detect Unintended Model Biases. In Pro-
ceedings of Empirical Methods in Natural Language
Michael J. Muller. 2007. Participatory Design: The Processing (EMNLP), pages 5744–5749, Hong
Third Space in HCI. In The Human-Computer Inter- Kong, China.
action Handbook, pages 1087–1108. CRC Press.
Shrimai Prabhumoye, Elijah Mayfield, and Alan W.
Moin Nadeem, Anna Bethke, and Siva Reddy. Black. 2019. Principled Frameworks for Evaluating
2020. StereoSet: Measuring stereotypical bias Ethics in NLP Systems. In Proceedings of the Work-
in pretrained language models. arXiv preprint shop on Innovative Use of NLP for Building Educa-
arXiv:2004.09456. tional Applications, Florence, Italy.
Marcelo Prates, Pedro Avelar, and Luis C. Lamb. 2019. without Access to Protected Attributes. In Proceed-
Assessing gender bias in machine translation: A ings of the North American Association for Com-
case study with google translate. Neural Computing putational Linguistics (NAACL), pages 4187–4195,
and Applications. Minneapolis, MN.
Rasmus Précenth. 2019. Word embeddings and gender Jonathan Rosa. 2019. Contesting Representations of
stereotypes in Swedish and English. Master’s thesis, Migrant “Illegality” through the Drop the I-Word
Uppsala University. Campaign: Rethinking Language Change and So-
cial Change. In Netta Avineri, Laura R. Graham,
Dennis R. Preston. 2009. Are you really smart (or Eric J. Johnson, Robin Conley Riner, and Jonathan
stupid, or cute, or ugly, or cool)? Or do you just talk Rosa, editors, Language and Social Justice in Prac-
that way? Language attitudes, standardization and tice. Routledge.
language change. Oslo: Novus forlag, pages 105–
129. Jonathan Rosa and Christa Burdick. 2017. Language
Ideologies. In Ofelia García, Nelson Flores, and
Flavien Prost, Nithum Thain, and Tolga Bolukbasi. Massimiliano Spotti, editors, The Oxford Handbook
2019. Debiasing Embeddings for Reduced Gender of Language and Society. Oxford University Press.
Bias in Text Classification. In Proceedings of the
Workshop on Gender Bias in Natural Language Pro- Jonathan Rosa and Nelson Flores. 2017. Unsettling
cessing, pages 69–75, Florence, Italy. race and language: Toward a raciolinguistic perspec-
tive. Language in Society, 46:621–647.
Reid Pryzant, Richard Diehl Martinez, Nathan Dass,
Sadao Kurohashi, Dan Jurafsky, and Diyi Yang. Sara Rosenthal and Kathleen McKeown. 2011. Age
2020. Automatically Neutralizing Subjective Bias Prediction in Blogs: A Study of Style, Content, and
in Text. In Proceedings of the AAAI Conference on Online Behavior in Pre- and Post-Social Media Gen-
Artificial Intelligence (AAAI), New York, NY. erations. In Proceedings of the North American As-
Arun K. Pujari, Ansh Mittal, Anshuman Padhi, An- sociation for Computational Linguistics (NAACL),
shul Jain, Mukesh Jadon, and Vikas Kumar. 2019. pages 763–772, Portland, OR.
Debiasing Gender biased Hindi Words with Word-
Candace Ross, Boris Katz, and Andrei Barbu.
embedding. In Proceedings of the International
2020. Measuring Social Biases in Grounded Vi-
Conference on Algorithms, Computing and Artificial
sion and Language Embeddings. arXiv preprint
Intelligence, pages 450–456.
arXiv:2002.08911.
Yusu Qian, Urwa Muaz, Ben Zhang, and Jae Won
Hyun. 2019. Reducing gender bias in word-level Richard Rothstein. 2017. The Color of Law: A For-
language models with a gender-equalizing loss func- gotten History of How Our Government Segregated
tion. In Proceedings of the ACL Student Research America. Liveright Publishing.
Workshop, pages 223–228, Florence, Italy.
David Rozado. 2020. Wide range screening of algo-
Shauli Ravfogel, Yanai Elazar, Hila Gonen, Michael rithmic bias in word embedding models using large
Twiton, and Yoav Goldberg. 2020. Null It Out: sentiment lexicons reveals underreported bias types.
Guarding Protected Attributes by Iterative Nullspace PLOS One.
Projection. In Proceedings of the Association for
Computational Linguistics (ACL). Elayne Ruane, Abeba Birhane, and Anthony Ven-
tresque. 2019. Conversational AI: Social and Ethi-
John R. Rickford and Sharese King. 2016. Language cal Considerations. In Proceedings of the Irish Con-
and linguistics on trial: Hearing Rachel Jeantel (and ference on Artificial Intelligence and Cognitive Sci-
other vernacular speakers) in the courtroom and be- ence, Galway, Ireland.
yond. Language, 92(4):948–988.
Rachel Rudinger, Chandler May, and Benjamin
Anthony Rios. 2020. FuzzE: Fuzzy Fairness Evalua- Van Durme. 2017. Social bias in elicited natural lan-
tion of Offensive Language Classifiers on African- guage inferences. In Proceedings of the Workshop
American English. In Proceedings of the AAAI Con- on Ethics in Natural Language Processing, pages
ference on Artificial Intelligence (AAAI), New York, 74–79, Valencia, Spain.
NY.
Rachel Rudinger, Jason Naradowsky, Brian Leonard,
Gerald Roche. 2019. Articulating language oppres- and Benjamin Van Durme. 2018. Gender Bias
sion: colonialism, coloniality and the erasure of Ti- in Coreference Resolution. In Proceedings of the
betâĂŹs minority languages. Patterns of Prejudice. North American Association for Computational Lin-
guistics (NAACL), pages 8–14, New Orleans, LA.
Alexey Romanov, Maria De-Arteaga, Hanna Wal-
lach, Jennifer Chayes, Christian Borgs, Alexandra Elizabeth B.N. Sanders. 2002. From user-centered to
Chouldechova, Sahin Geyik, Krishnaram Kentha- participatory design approaches. In Jorge Frascara,
padi, Anna Rumshisky, and Adam Tauman Kalai. editor, Design and the Social Sciences: Making Con-
2019. What’s in a Name? Reducing Bias in Bios nections, pages 18–25. CRC Press.
Brenda Salenave Santana, Vinicius Woloszyn, and Le- Emily Sheng, Kai-Wei Chang, Premkumar Natarajan,
andro Krug Wives. 2018. Is there gender bias and and Nanyun Peng. 2019. The Woman Worked as
stereotype in Portuguese word embeddings? In a Babysitter: On Biases in Language Generation.
Proceedings of the International Conference on the In Proceedings of Empirical Methods in Natural
Computational Processing of Portuguese Student Re- Language Processing (EMNLP), pages 3398–3403,
search Workshop, Canela, Brazil. Hong Kong, China.
Maarten Sap, Dallas Card, Saadia Gabriel, Yejin Choi, Emily Sheng, Kai-Wei Chang, Premkumar Natarajan,
and Noah A. Smith. 2019. The risk of racial bias in and Nanyun Peng. 2020. Towards Controllable
hate speech detection. In Proceedings of the Asso- Biases in Language Generation. arXiv preprint
ciation for Computational Linguistics (ACL), pages arXiv:2005.00268.
1668–1678, Florence, Italy.
Seungjae Shin, Kyungwoo Song, JoonHo Jang, Hyemi
Maarten Sap, Saadia Gabriel, Lianhui Qin, Dan Juraf- Kim, Weonyoung Joo, and Il-Chul Moon. 2020.
sky, Noah A. Smith, and Yejin Choi. 2020. Social Neutralizing Gender Bias in Word Embedding with
Bias Frames: Reasoning about Social and Power Im- Latent Disentanglement and Counterfactual Genera-
plications of Language. In Proceedings of the Asso- tion. arXiv preprint arXiv:2004.03133.
ciation for Computational Linguistics (ACL).
Jesper Simonsen and Toni Robertson, editors. 2013.
Hanna Sassaman, Jennifer Lee, Jenessa Irvine, and Routledge International Handbook of Participatory
Shankar Narayan. 2020. Creating Community- Design. Routledge.
Based Tech Policy: Case Studies, Lessons Learned,
and What Technologists and Communities Can Do Gabriel Stanovsky, Noah A. Smith, and Luke Zettle-
Together. In Proceedings of the Conference on Fair- moyer. 2019. Evaluating gender bias in machine
ness, Accountability, and Transparency, Barcelona, translation. In Proceedings of the Association for
Spain. Computational Linguistics (ACL), pages 1679–1684,
Florence, Italy.
Danielle Saunders and Bill Byrne. 2020. Reducing
Gender Bias in Neural Machine Translation as a Do- Yolande Strengers, Lizhe Qu, Qiongkai Xu, and Jarrod
main Adaptation Problem. In Proceedings of the As- Knibbe. 2020. Adhering, Steering, and Queering:
sociation for Computational Linguistics (ACL). Treatment of Gender in Natural Language Genera-
tion. In Proceedings of the Conference on Human
Tyler Schnoebelen. 2017. Goal-Oriented Design for Factors in Computing Systems (CHI), Honolulu, HI.
Ethical Machine Learning and NLP. In Proceedings
of the Workshop on Ethics in Natural Language Pro- Tony Sun, Andrew Gaut, Shirlyn Tang, Yuxin Huang,
cessing, pages 88–93, Valencia, Spain. Mai ElSherief, Jieyu Zhao, Diba Mirza, Elizabeth
Belding, Kai-Wei Chang, and William Yang Wang.
Sabine Sczesny, Magda Formanowicz, and Franziska 2019. Mitigating Gender Bias in Natural Lan-
Moser. 2016. Can gender-fair language reduce gen- guage Processing: Literature Review. In Proceed-
der stereotyping and discrimination? Frontiers in ings of the Association for Computational Linguis-
Psychology, 7. tics (ACL), pages 1630–1640, Florence, Italy.
João Sedoc and Lyle Ungar. 2019. The Role of Pro- Adam Sutton, Thomas Lansdall-Welfare, and Nello
tected Class Word Lists in Bias Identification of Con- Cristianini. 2018. Biased embeddings from wild
textualized Word Representations. In Proceedings data: Measuring, understanding and removing.
of the Workshop on Gender Bias in Natural Lan- In Proceedings of the International Symposium
guage Processing, pages 55–61, Florence, Italy. on Intelligent Data Analysis, pages 328–339, ’s-
Hertogenbosch, Netherlands.
Procheta Sen and Debasis Ganguly. 2020. Towards So-
cially Responsible AI: Cognitive Bias-Aware Multi- Chris Sweeney and Maryam Najafian. 2019. A Trans-
Objective Learning. In Proceedings of the AAAI parent Framework for Evaluating Unintended De-
Conference on Artificial Intelligence (AAAI), New mographic Bias in Word Embeddings. In Proceed-
York, NY. ings of the Association for Computational Linguis-
tics (ACL), pages 1662–1667, Florence, Italy.
Deven Shah, H. Andrew Schwartz, and Dirk Hovy.
2020. Predictive Biases in Natural Language Pro- Chris Sweeney and Maryam Najafian. 2020. Reduc-
cessing Models: A Conceptual Framework and ing sentiment polarity for demographic attributes
Overview. In Proceedings of the Association for in word embeddings using adversarial learning. In
Computational Linguistics (ACL). Proceedings of the Conference on Fairness, Ac-
countability, and Transparency, pages 359–368,
Judy Hanwen Shen, Lauren Fratamico, Iyad Rahwan, Barcelona, Spain.
and Alexander M. Rush. 2018. Darling or Baby-
girl? Investigating Stylistic Bias in Sentiment Anal- Nathaniel Swinger, Maria De-Arteaga, Neil Thomas
ysis. In Proceedings of the Workshop on Fairness, Heffernan, Mark D.M. Leiserson, and Adam Tau-
Accountability, and Transparency (FAT/ML), Stock- man Kalai. 2019. What are the biases in my word
holm, Sweden. embedding? In Proceedings of the Conference on
Artificial Intelligence, Ethics, and Society (AIES), Kellie Webster, Marta R. Costa-jussà, Christian Hard-
Honolulu, HI. meier, and Will Radford. 2019. Gendered Ambigu-
ous Pronoun (GAP) Shared Task at the Gender Bias
Samson Tan, Shafiq Joty, Min-Yen Kan, and Richard in NLP Workshop 2019. In Proceedings of the Work-
Socher. 2020. It’s Morphin’ Time! Combating shop on Gender Bias in Natural Language Process-
Linguistic Discrimination with Inflectional Perturba- ing, pages 1–7, Florence, Italy.
tions. In Proceedings of the Association for Compu-
tational Linguistics (ACL). Kellie Webster, Marta Recasens, Vera Axelrod, and Ja-
son Baldridge. 2018. Mind the GAP: A balanced
Yi Chern Tan and L. Elisa Celis. 2019. Assessing corpus of gendered ambiguous pronouns. Transac-
Social and Intersectional Biases in Contextualized tions of the Association for Computational Linguis-
Word Representations. In Proceedings of the Con- tics, 6:605–618.
ference on Neural Information Processing Systems,
Vancouver, Canada. Walt Wolfram and Natalie Schilling. 2015. American
English: Dialects and Variation, 3 edition. Wiley
J. Michael Terry, Randall Hendrick, Evangelos Evan- Blackwell.
gelou, and Richard L. Smith. 2010. Variable
dialect switching among African American chil- Austin P. Wright, Omar Shaikh, Haekyu Park, Will Ep-
dren: Inferences about working memory. Lingua, person, Muhammed Ahmed, Stephane Pinel, Diyi
120(10):2463–2475. Yang, and Duen Horng (Polo) Chau. 2020. RE-
CAST: Interactive Auditing of Automatic Toxicity
Joel Tetreault, Daniel Blanchard, and Aoife Cahill. Detection Models. In Proceedings of the Con-
2013. A Report on the First Native Language Iden- ference on Human Factors in Computing Systems
tification Shared Task. In Proceedings of the Work- (CHI), Honolulu, HI.
shop on Innovative Use of NLP for Building Educa-
tional Applications, pages 48–57, Atlanta, GA. Yinchuan Xu and Junlin Yang. 2019. Look again at
the syntax: Relational graph convolutional network
Mike Thelwall. 2018. Gender Bias in Sentiment Anal- for gendered ambiguous pronoun resolution. In Pro-
ysis. Online Information Review, 42(1):45–57. ceedings of the Workshop on Gender Bias in Natu-
ral Language Processing, pages 96–101, Florence,
Kristen Vaccaro, Karrie Karahalios, Deirdre K. Mul- Italy.
ligan, Daniel Kluttz, and Tad Hirsch. 2019. Con-
testability in Algorithmic Systems. In Conference Kai-Chou Yang, Timothy Niven, Tzu-Hsuan Chou, and
Companion Publication of the 2019 on Computer Hung-Yu Kao. 2019. Fill the GAP: Exploiting
Supported Cooperative Work and Social Computing, BERT for Pronoun Resolution. In Proceedings of
pages 523–527, Austin, TX. the Workshop on Gender Bias in Natural Language
Processing, pages 102–106, Florence, Italy.
Ameya Vaidya, Feng Mai, and Yue Ning. 2019. Em-
pirical Analysis of Multi-Task Learning for Reduc- Zekun Yang and Juan Feng. 2020. A Causal Inference
ing Model Bias in Toxic Comment Detection. arXiv Method for Reducing Gender Bias in Word Embed-
preprint arXiv:1909.09758v2. ding Relations. In Proceedings of the AAAI Con-
ference on Artificial Intelligence (AAAI), New York,
Eva Vanmassenhove, Christian Hardmeier, and Andy NY.
Way. 2018. Getting Gender Right in Neural Ma-
chine Translation. In Proceedings of Empirical Daisy Yoo, Anya Ernest, Sofia Serholt, Eva Eriksson,
Methods in Natural Language Processing (EMNLP), and Peter Dalsgaard. 2019. Service Design in HCI
pages 3003–3008, Brussels, Belgium. Research: The Extended Value Co-creation Model.
In Proceedings of the Halfway to the Future Sympo-
Jesse Vig, Sebastian Gehrmann, Yonatan Belinkov, sium, Nottingham, United Kingdom.
Sharon Qian, Daniel Nevo, Yaron Singer, and Stu-
art Shieber. 2020. Causal Mediation Analysis for Brian Hu Zhang, Blake Lemoine, and Margaret
Interpreting Neural NLP: The Case of Gender Bias. Mitchell. 2018. Mitigating unwanted biases with ad-
arXiv preprint arXiv:2004.12265. versarial learning. In Proceedings of the Conference
on Artificial Intelligence, Ethics, and Society (AIES),
Tianlu Wang, Xi Victoria Lin, Nazneen Fatema Ra- New Orleans, LA.
jani, Bryan McCann, Vicente Ordonez, and Caim-
ing Xiong. 2020. Double-Hard Debias: Tailoring Guanhua Zhang, Bing Bai, Junqi Zhang, Kun Bai, Con-
Word Embeddings for Gender Bias Mitigation. In ghui Zhu, and Tiejun Zhao. 2020a. Demographics
Proceedings of the Association for Computational Should Not Be the Reason of Toxicity: Mitigating
Linguistics (ACL). Discrimination in Text Classifications with Instance
Weighting. In Proceedings of the Association for
Zili Wang. 2019. MSnet: A BERT-based Network for Computational Linguistics (ACL).
Gendered Pronoun Resolution. In Proceedings of
the Workshop on Gender Bias in Natural Language Haoran Zhang, Amy X. Lu, Mohamed Abdalla,
Processing, pages 89–95, Florence, Italy. Matthew McDermott, and Marzyeh Ghassemi.
2020b. Hurtful Words: Quantifying Biases in Clin- A.1 Categorization details
ical Contextual Word Embeddings. In Proceedings
of the ACM Conference on Health, Inference, and In this section, we provide some additional de-
Learning. tails about our method—specifically, our catego-
rization.
Jieyu Zhao, Subhabrata Mukherjee, Saghar Hosseini,
Kai-Wei Chang, and Ahmed Hassan Awadallah. What counts as being covered by an NLP task?
2020. Gender Bias in Multilingual Embeddings and We considered a paper to cover a given NLP task
Cross-Lingual Transfer. In Proceedings of the Asso-
ciation for Computational Linguistics (ACL). if it analyzed “bias” with respect to that task, but
not if it only evaluated overall performance on
Jieyu Zhao, Tianlu Wang, Mark Yatskar, Ryan Cot- that task. For example, a paper examining the im-
terell, Vicente Ordonez, and Kai-Wei Chang. 2019. pact of mitigating “bias” in word embeddings on
Gender Bias in Contextualized Word Embeddings.
In Proceedings of the North American Association “bias” in sentiment analysis would be counted as
for Computational Linguistics (NAACL), pages 629– covering both NLP tasks. In contrast, a paper as-
634, Minneapolis, MN. sessing whether performance on sentiment analy-
sis degraded after mitigating “bias” in word em-
Jieyu Zhao, Tianlu Wang, Mark Yatskar, Vicente Or-
beddings would be counted only as focusing on
donez, and Kai-Wei Chang. 2017. Men also like
shopping: Reducing gender bias amplification us- embeddings.
ing corpus-level constraints. In Proceedings of
Empirical Methods in Natural Language Process- What counts as a motivation? We considered a
ing (EMNLP), pages 2979–2989, Copenhagen, Den- motivation to include any description of the prob-
mark. lem that motivated the paper or proposed quantita-
tive technique, including any normative reasoning.
Jieyu Zhao, Tianlu Wang, Mark Yatskar, Vicente Or-
donez, and Kai-Wei Chang. 2018a. Gender Bias We excluded from the “Vague/unstated” cate-
in Coreference Resolution: Evaluation and Debias- gory of motivations the papers that participated in
ing Methods. In Proceedings of the North American the Gendered Ambiguous Pronoun (GAP) Shared
Association for Computational Linguistics (NAACL), Task at the First ACL Workshop on Gender Bias
pages 15–20, New Orleans, LA.
in NLP. In an ideal world, shared task papers
Jieyu Zhao, Yichao Zhou, Zeyu Li, Wei Wang, and Kai- would engage with “bias” more critically, but
Wei Chang. 2018b. Learning Gender-Neutral Word given the nature of shared tasks it is understand-
Embeddings. In Proceedings of Empirical Methods able that they do not. As a result, we ex-
in Natural Language Processing (EMNLP), pages
4847–4853, Brussels, Belgium. cluded them from our counts for techniques as
well. We cite the papers here; most propose tech-
Alina Zhiltsova, Simon Caton, and Catherine Mulwa. niques we would have categorized as “Question-
2019. Mitigation of Unintended Biases against Non- able correlations,” with a few as “Other repre-
Native English Texts in Sentiment Analysis. In Pro-
ceedings of the Irish Conference on Artificial Intelli-
sentational harms” (Abzaliev, 2019; Attree, 2019;
gence and Cognitive Science, Galway, Ireland. Bao and Qiao, 2019; Chada, 2019; Ionita et al.,
2019; Liu, 2019; Lois et al., 2019; Wang, 2019;
Pei Zhou, Weijia Shi, Jieyu Zhao, Kuan-Hao Huang, Xu and Yang, 2019; Yang et al., 2019).
Muhao Chen, and Kai-Wei Chang. 2019. Examin-
We excluded Dabas et al. (2020) from our sur-
ing gender bias in languages with grammatical gen-
ders. In Proceedings of Empirical Methods in Nat- vey because we could not determine what this pa-
ural Language Processing (EMNLP), pages 5279– per’s user study on fairness was actually measur-
5287, Hong Kong, China. ing.
Finally, we actually categorized the motiva-
Ran Zmigrod, S. J. Mielke, Hanna Wallach, and Ryan
Cotterell. 2019. Counterfactual data augmentation tion for Liu et al. (2019) (i.e., the last row in Ta-
for mitigating gender stereotypes in languages with ble 3) as “Questionable correlations” due to a
rich morphology. In Proceedings of the Association sentence elsewhere in the paper; had the para-
for Computational Linguistics (ACL), pages 1651– graph we quoted been presented without more de-
1661, Florence, Italy.
tail, we would have categorized the motivation as
“Vague/unstated.”
A Appendix
In Table 3, we provide examples of the papers’ mo- A.2 Full categorization: Motivations
tivations and techniques across several NLP tasks. Allocational harms Hovy and Spruit (2016);
Categories
NLP task Stated motivation Motivations Techniques
Language “Existing biases in data can be amplified by models and the re- Allocational Questionable
modeling sulting output consumed by the public can influence them, en- harms, correlations
courage and reinforce harmful stereotypes, or distort the truth.
(Bordia and Bowman, stereotyping
2019) Automated systems that depend on these models can take prob-
lematic actions based on biased profiling of individuals.”
Sentiment “Other biases can be inappropriate and result in negative experi- Allocational Questionable
analysis ences for some groups of people. Examples include, loan eligibil- harms, other correlations
ity and crime recidivism prediction systems...and resumé sorting
(Kiritchenko and Mohammad, representational (differences in
2018) systems that believe that men are more qualified to be program- harms (system sentiment
mers than women (Bolukbasi et al., 2016). Similarly, sentiment performance intensity scores
and emotion analysis systems can also perpetuate and accentu- differences w.r.t. w.r.t. text about
ate inappropriate human biases, e.g., systems that consider utter- text written by different social
ances from one race or gender to be less positive simply because different social groups)
of their race or gender, or customer support systems that priori- groups)
tize a call from an angry male over a call from the equally angry
female.”
Machine “[MT training] may incur an association of gender-specified pro- Questionable Questionable
translation nouns (in the target) and gender-neutral ones (in the source) for correlations, correlations
(Cho et al., lexicon pairs that frequently collocate in the corpora. We claim other
2019) that this kind of phenomenon seriously threatens the fairness of a representational
translation system, in the sense that it lacks generality and inserts harms
social bias to the inference. Moreover, the input is not fully cor-
rect (considering gender-neutrality) and might offend the users
who expect fairer representations.”
Machine “Learned models exhibit social bias when their training data en- Stereotyping, Stereotyping,
translation code stereotypes not relevant for the task, but the correlations questionable other
(Stanovsky et al., are picked up anyway.” correlations representational
2019) harms (system
performance
differences),
questionable
correlations
Type-level “However, embeddings trained on human-generated corpora Allocational Stereotyping
embeddings have been demonstrated to inherit strong gender stereotypes that harms,
(Zhao et al., reflect social constructs....Such a bias substantially affects down- stereotyping,
2018b) stream applications....This concerns the practitioners who use other
the embedding model to build gender-sensitive applications such representational
as a resume filtering system or a job recommendation system harms
as the automated system may discriminate candidates based on
their gender, as reflected by their name. Besides, biased embed-
dings may implicitly affect downstream applications used in our
daily lives. For example, when searching for ‘computer scientist’
using a search engine...a search algorithm using an embedding
model in the backbone tends to rank male scientists higher than
females’ [sic], hindering women from being recognized and fur-
ther exacerbating the gender inequality in the community.”
Type-level “[P]rominent word embeddings such as word2vec (Mikolov et Vague Stereotyping
and contextu- al., 2013) and GloVe (Pennington et al., 2014) encode systematic
alized biases against women and black people (Bolukbasi et al., 2016;
embeddings Garg et al., 2018), implicating many NLP systems in scaling up
(May et al., social injustice.”
2019)
Dialogue “Since the goal of dialogue systems is to talk with users...if the Vague/unstated Stereotyping,
generation systems show discriminatory behaviors in the interactions, the other
(Liu et al., user experience will be adversely affected. Moreover, public com- representational
2019) mercial chatbots can get resisted for their improper speech.” harms,
questionable
correlations
Table 3: Examples of the categories into which the papers’ motivations and proposed quantitative techniques for
measuring or mitigating “bias” fall. Bold text in the quotes denotes the content that yields our categorizations.
Caliskan et al. (2017); Madnani et al. (2017); Dixon et al. (2018); Kiritchenko and Mohammad
Dixon et al. (2018); Kiritchenko and Mohammad (2018); Park et al. (2018); Shen et al. (2018);
(2018); Shen et al. (2018); Zhao et al. Thelwall (2018); Zhao et al. (2018b);
(2018b); Bhaskaran and Bhallamudi (2019); Badjatiya et al. (2019); Bagdasaryan et al. (2019);
Bordia and Bowman (2019); Brunet et al. Bamman et al. (2019); Cao and Daumé (2019);
(2019); Chaloner and Maldonado (2019); Chaloner and Maldonado (2019); Cho et al.
De-Arteaga et al. (2019); Dev and Phillips (2019); Davidson et al. (2019); De-Arteaga et al.
(2019); Font and Costa-jussà (2019); (2019); Fisher (2019); Font and Costa-jussà
James-Sorenson and Alvarez-Melis (2019); (2019); Garimella et al. (2019); Loukina et al.
Kurita et al. (2019); Mayfield et al. (2019); (2019); Mayfield et al. (2019); Mehrabi et al.
Pujari et al. (2019); Romanov et al. (2019); (2019); Nozza et al. (2019); Prabhakaran et al.
Ruane et al. (2019); Sedoc and Ungar (2019); (2019); Romanov et al. (2019); Ruane et al.
Sun et al. (2019); Zmigrod et al. (2019); (2019); Sap et al. (2019); Sheng et al. (2019);
Hutchinson et al. (2020); Papakyriakopoulos et al. Sun et al. (2019); Sweeney and Najafian (2019);
(2020); Ravfogel et al. (2020); Strengers et al. Vaidya et al. (2019); Gaut et al. (2020); Gencoglu
(2020); Sweeney and Najafian (2020); Tan et al. (2020); Hovy et al. (2020); Hutchinson et al.
(2020); Zhang et al. (2020b). (2020); Kim et al. (2020); Peng et al. (2020);
Rios (2020); Sap et al. (2020); Shah et al. (2020);
Stereotyping Bolukbasi et al. (2016a,b); Sheng et al. (2020); Tan et al. (2020); Zhang et al.
Caliskan et al. (2017); McCurdy and Serbetçi (2020a,b).
(2017); Rudinger et al. (2017); Zhao et al. (2017);
Curry and Rieser (2018); Díaz et al. (2018); Questionable correlations Jørgensen et al.
Santana et al. (2018); Sutton et al. (2018); (2015); Hovy and Spruit (2016); Madnani et al.
Zhao et al. (2018a,b); Agarwal et al. (2019); (2017); Rudinger et al. (2017); Zhao et al.
Basta et al. (2019); Bhaskaran and Bhallamudi (2017); Burns et al. (2018); Dixon et al. (2018);
(2019); Bordia and Bowman (2019); Kiritchenko and Mohammad (2018); Lu et al.
Brunet et al. (2019); Cao and Daumé (2018); Park et al. (2018); Shen et al. (2018);
(2019); Chaloner and Maldonado (2019); Zhang et al. (2018); Badjatiya et al. (2019);
Cho et al. (2019); Dev and Phillips Bhargava and Forsyth (2019); Cao and Daumé
(2019); Font and Costa-jussà (2019); Cho et al. (2019); Davidson et al.
(2019); Gonen and Goldberg (2019); (2019); Dev et al. (2019); Garimella et al.
James-Sorenson and Alvarez-Melis (2019); (2019); Garg et al. (2019); Huang et al.
Kaneko and Bollegala (2019); Karve et al. (2019); James-Sorenson and Alvarez-Melis
(2019); Kurita et al. (2019); Lauscher and Glavaš (2019); Kaneko and Bollegala (2019); Liu et al.
(2019); Lee et al. (2019); Manzini et al. (2019); (2019); Karve et al. (2019); Nozza et al. (2019);
Mayfield et al. (2019); Précenth (2019); Prabhakaran et al. (2019); Romanov et al. (2019);
Pujari et al. (2019); Ruane et al. (2019); Sap et al. (2019); Sedoc and Ungar (2019);
Stanovsky et al. (2019); Sun et al. (2019); Stanovsky et al. (2019); Sweeney and Najafian
Tan and Celis (2019); Webster et al. (2019); (2019); Vaidya et al. (2019); Zhiltsova et al.
Zmigrod et al. (2019); Gyamfi et al. (2020); (2019); Chopra et al. (2020); Gonen and Webster
Hube et al. (2020); Hutchinson et al. (2020); (2020); Gyamfi et al. (2020); Hube et al.
Kim et al. (2020); Nadeem et al. (2020); (2020); Ravfogel et al. (2020); Rios (2020);
Papakyriakopoulos et al. (2020); Ravfogel et al. Ross et al. (2020); Saunders and Byrne (2020);
(2020); Rozado (2020); Sen and Ganguly (2020); Sen and Ganguly (2020); Shah et al. (2020);
Shin et al. (2020); Strengers et al. (2020). Sweeney and Najafian (2020); Yang and Feng
(2020); Zhang et al. (2020a).
Other representational harms
Hovy and Søgaard (2015); Blodgett et al. (2016); Vague/unstated Rudinger et al. (2018);
Bolukbasi et al. (2016b); Hovy and Spruit (2016); Webster et al. (2018); Dinan et al. (2019); Florez
Blodgett and O’Connor (2017); Larson (2017); (2019); Jumelet et al. (2019); Lauscher et al.
Schnoebelen (2017); Blodgett et al. (2018); (2019); Liang et al. (2019); Maudslay et al.
Curry and Rieser (2018); Díaz et al. (2018); (2019); May et al. (2019); Prates et al.
(2019); Prost et al. (2019); Qian et al. (2019); (2020); Yang and Feng (2020); Zhao et al. (2020).
Swinger et al. (2019); Zhao et al. (2019);
Zhou et al. (2019); Ethayarajh (2020); Huang et al. Other representational harms Jørgensen et al.
(2020); Jia et al. (2020); Popović et al. (2020); (2015); Hovy and Søgaard (2015); Blodgett et al.
Pryzant et al. (2020); Vig et al. (2020); Wang et al. (2016); Blodgett and O’Connor (2017);
(2020); Zhao et al. (2020). Blodgett et al. (2018); Curry and Rieser (2018);
Dixon et al. (2018); Park et al. (2018); Thelwall
Surveys, frameworks, and meta-analyses (2018); Webster et al. (2018); Badjatiya et al.
Hovy and Spruit (2016); Larson (2017); (2019); Bagdasaryan et al. (2019); Bamman et al.
McCurdy and Serbetçi (2017); Schnoebelen (2019); Bhargava and Forsyth (2019);
(2017); Basta et al. (2019); Ethayarajh et al. Cao and Daumé (2019); Font and Costa-jussà
(2019); Gonen and Goldberg (2019); (2019); Garg et al. (2019); Garimella et al.
Lauscher and Glavaš (2019); Loukina et al. (2019); Liu et al. (2019); Loukina et al.
(2019); Mayfield et al. (2019); Mirzaev et al. (2019); Mehrabi et al. (2019); Nozza et al.
(2019); Prabhumoye et al. (2019); Ruane et al. (2019); Sap et al. (2019); Sheng et al. (2019);
(2019); Sedoc and Ungar (2019); Sun et al. Stanovsky et al. (2019); Vaidya et al. (2019);
(2019); Nissim et al. (2020); Rozado (2020); Webster et al. (2019); Ethayarajh (2020);
Shah et al. (2020); Strengers et al. (2020); Gaut et al. (2020); Gencoglu (2020); Hovy et al.
Wright et al. (2020). (2020); Huang et al. (2020); Kim et al.
(2020); Peng et al. (2020); Ravfogel et al.
B Full categorization: Techniques (2020); Rios (2020); Sap et al. (2020);
Saunders and Byrne (2020); Sheng et al. (2020);
Allocational harms De-Arteaga et al. (2019);
Sweeney and Najafian (2020); Tan et al. (2020);
Prost et al. (2019); Romanov et al. (2019).
Zhang et al. (2020a,b); Zhao et al. (2020).
Stereotyping Bolukbasi et al. (2016a,b);
Questionable correlations Jurgens et al.
Caliskan et al. (2017); McCurdy and Serbetçi
(2017); Madnani et al. (2017); Rudinger et al.
(2017); Díaz et al. (2018); Santana et al. (2018);
(2017); Zhao et al. (2017); Burns et al. (2018);
Sutton et al. (2018); Zhang et al. (2018);
Díaz et al. (2018); Kiritchenko and Mohammad
Zhao et al. (2018a,b); Agarwal et al. (2019);
(2018); Lu et al. (2018); Rudinger et al. (2018);
Basta et al. (2019); Bhaskaran and Bhallamudi
Shen et al. (2018); Bordia and Bowman
(2019); Brunet et al. (2019); Cao and Daumé
(2019); Cao and Daumé (2019); Cho et al.
(2019); Chaloner and Maldonado (2019);
(2019); Davidson et al. (2019); Dev et al.
Dev and Phillips (2019); Ethayarajh et al.
(2019); Dinan et al. (2019); Fisher (2019);
(2019); Gonen and Goldberg (2019);
Florez (2019); Font and Costa-jussà (2019);
James-Sorenson and Alvarez-Melis (2019);
Garg et al. (2019); Huang et al. (2019); Liu et al.
Jumelet et al. (2019); Kaneko and Bollegala
(2019); Nozza et al. (2019); Prabhakaran et al.
(2019); Karve et al. (2019); Kurita et al. (2019);
(2019); Qian et al. (2019); Sap et al. (2019);
Lauscher and Glavaš (2019); Lauscher et al.
Stanovsky et al. (2019); Sweeney and Najafian
(2019); Lee et al. (2019); Liang et al. (2019);
(2019); Swinger et al. (2019); Zhiltsova et al.
Liu et al. (2019); Manzini et al. (2019);
(2019); Zmigrod et al. (2019); Hube et al. (2020);
Maudslay et al. (2019); May et al. (2019);
Hutchinson et al. (2020); Jia et al. (2020);
Mirzaev et al. (2019); Prates et al. (2019);
Papakyriakopoulos et al. (2020); Popović et al.
Précenth (2019); Prost et al. (2019); Pujari et al.
(2020); Pryzant et al. (2020); Saunders and Byrne
(2019); Qian et al. (2019); Sedoc and Ungar
(2020); Sen and Ganguly (2020); Shah et al.
(2019); Stanovsky et al. (2019); Tan and Celis
(2020); Sweeney and Najafian (2020); Zhang et al.
(2019); Zhao et al. (2019); Zhou et al. (2019);
(2020b).
Chopra et al. (2020); Gyamfi et al. (2020);
Nadeem et al. (2020); Nissim et al. (2020); Vague/unstated None.
Papakyriakopoulos et al. (2020); Popović et al.
(2020); Ravfogel et al. (2020); Ross et al. (2020); Surveys, frameworks, and meta-analyses
Rozado (2020); Saunders and Byrne (2020); Hovy and Spruit (2016); Larson (2017);
Shin et al. (2020); Vig et al. (2020); Wang et al. McCurdy and Serbetçi (2017); Schnoebelen
(2017); Basta et al. (2019); Ethayarajh et al.
(2019); Gonen and Goldberg (2019);
Lauscher and Glavaš (2019); Loukina et al.
(2019); Mayfield et al. (2019); Mirzaev et al.
(2019); Prabhumoye et al. (2019); Ruane et al.
(2019); Sedoc and Ungar (2019); Sun et al.
(2019); Nissim et al. (2020); Rozado (2020);
Shah et al. (2020); Strengers et al. (2020);
Wright et al. (2020).