Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

Assessing gender bias in machine translation: a case study with Google Translate

Published: 01 May 2020 Publication History

Abstract

Recently there has been a growing concern in academia, industrial research laboratories and the mainstream commercial media about the phenomenon dubbed as machine bias, where trained statistical models—unbeknownst to their creators—grow to reflect controversial societal asymmetries, such as gender or racial bias. A significant number of Artificial Intelligence tools have recently been suggested to be harmfully biased toward some minority, with reports of racist criminal behavior predictors, Apple’s Iphone X failing to differentiate between two distinct Asian people and the now infamous case of Google photos’ mistakenly classifying black people as gorillas. Although a systematic study of such biases can be difficult, we believe that automated translation tools can be exploited through gender neutral languages to yield a window into the phenomenon of gender bias in AI. In this paper, we start with a comprehensive list of job positions from the U.S. Bureau of Labor Statistics (BLS) and used it in order to build sentences in constructions like “He/She is an Engineer” (where “Engineer” is replaced by the job position of interest) in 12 different gender neutral languages such as Hungarian, Chinese, Yoruba, and several others. We translate these sentences into English using the Google Translate API, and collect statistics about the frequency of female, male and gender neutral pronouns in the translated output. We then show that Google Translate exhibits a strong tendency toward male defaults, in particular for fields typically associated to unbalanced gender distribution or stereotypes such as STEM (Science, Technology, Engineering and Mathematics) jobs. We ran these statistics against BLS’ data for the frequency of female participation in each job position, in which we show that Google Translate fails to reproduce a real-world distribution of female workers. In summary, we provide experimental evidence that even if one does not expect in principle a 50:50 pronominal gender distribution, Google Translate yields male defaults much more frequently than what would be expected from demographic data alone. We believe that our study can shed further light on the phenomenon of machine bias and are hopeful that it will ignite a debate about the need to augment current statistical translation tools with debiasing techniques—which can already be found in the scientific literature.

References

[1]
Angwin J, Larson J, Mattu S, Kirchner L (2016) Machine bias: there’s software used across the country to predict future criminals and it’s biased against blacks. https://www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing. Last visited 2017-12-17
[2]
Bahdanau D, Cho K, Bengio Y (2014) Neural machine translation by jointly learning to align and translate. arxiv:1409.0473. Accessed 9 Mar 2019
[4]
Boitet C, Blanchon H, Seligman M, Bellynck V (2010) MT on and for the web. In: 2010 International conference on natural language processing and knowledge engineering (NLP-KE), IEEE, pp 1–10
[5]
Bolukbasi T, Chang KW, Zou JY, Saligrama V, Kalai AT (2016) Man is to computer programmer as woman is to homemaker? Debiasing word embeddings. In: Advances in neural information processing systems 29: annual conference on neural information processing systems 2016, December 5–10. Barcelona, Spain, pp 4349–4357
[6]
Boroditsky L, Schmidt LA, and Phillips W Getner D and Goldin-Meadow S Sex, syntax, and semantics Language in mind: advances in the study of language and thought 2003 Cambridge MIT Press 61-79
[7]
Bureau of Labor Statistics (2017) Table 11: employed persons by detailed occupation, sex, race, and Hispanic or Latino ethnicity, 2017. Labor force statistics from the current population survey, United States Department of Labor, Washington D.C
[8]
Carl M and Way A Recent advances in example-based machine translation 2003 Berlin Springer
[9]
Chomsky N (2011) The golden age: a look at the original roots of artificial intelligence, cognitive science, and neuroscience (partial transcript of an interview with N. Chomsky at MIT150 Symposia: Brains, minds and machines symposium). https://chomsky.info/20110616/. Last visited 26 Dec 2017
[10]
Clauburn T (2018) Boffins bash Google Translate for sexism. https://www.theregister.co.uk/2018/09/10/boffins_bash_google_translate_for_sexist_language/. Posted 10 Sep 2018
[11]
Dascal M Universal language schemes in England and France, 1600–1800 comments on James Knowlson Studia leibnitiana 1982 14 1 98-109
[12]
Diño G (2019) He said, she said: addressing gender in neural machine translation. https://slator.com/technology/he-said-she-said-addressing-gender-in-neural-machine-translation/. Posted 22 Jan 2019
[13]
Dryer MS and Haspelmath M WALS online 2013 Leipzig Max Planck Institute for Evolutionary Anthropology
[14]
Firat O, Cho K, Sankaran B, Yarman-Vural FT, and Bengio Y Multi-way, multilingual neural machine translation Comput Speech Lang 2017 45 236-252
[15]
Garcia M Racist in the machine: the disturbing implications of algorithmic bias World Policy J 2016 33 4 111-117
[16]
Google: language support for the neural machine translation model (2017). https://cloud.google.com/translate/docs/languages#languages-nmt. Last visited 19 Mar 2018
[17]
Gordin MD Scientific Babel: how science was done before and after global English 2015 Chicago University of Chicago Press
[18]
Hajian S, Bonchi F, Castillo C (2016) Algorithmic bias: from discrimination discovery to fairness-aware data mining. In: Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, ACM, pp 2125–2126
[19]
Hutchins WJ Machine translation: past, present, future 1986 Chichester Ellis Horwood
[20]
Johnson M, Schuster M, Le QV, Krikun M, Wu Y, Chen Z, Thorat N, Viégas FB, Wattenberg M, Corrado G, Hughes M, and Dean J Google’s multilingual neural machine translation system: enabling zero-shot translation TACL 2017 5 339-351
[21]
Kay P and Kempton W What is the Sapir–Whorf hypothesis? Am Anthropol 1984 86 1 65-79
[22]
Kelman S (2014) Translate community: help us improve google translate!. https://search.googleblog.com/2014/07/translate-community-help-us-improve.html. Last visited 12 Mar 2018
[23]
Kirkpatrick K Battling algorithmic bias: how do we ensure algorithms treat us fairly? Commun ACM 2016 59 10 16-17
[25]
Koehn P Statistical machine translation 2009 Cambridge Cambridge University Press
[26]
Koehn P, Hoang H, Birch A, Callison-Burch C, Federico M, Bertoldi N, Cowan B, Shen W, Moran C, Zens R, Dyer C, Bojar O, Constantin A, Herbst E (2007) Moses: open source toolkit for statistical machine translation. In: ACL 2007, Proceedings of the 45th annual meeting of the association for computational linguistics, June 23–30, 2007, Prague, Czech Republic. http://aclweb.org/anthology/P07-2045. Accessed 9 Mar 2019
[27]
Locke WN and Booth AD Machine translation of languages: fourteen essays 1955 New York Wiley
[28]
Mills KA (2017) ’Racist’ soap dispenser refuses to help dark-skinned man wash his hands—but Twitter blames ’technology’. http://www.mirror.co.uk/news/world-news/racist-soap-dispenser-refuses-help-11004385. Last visited 17 Dec 2017
[29]
Moss-Racusin CA, Molenda AK, and Cramer CR Can evidence impact attitudes? Public reactions to evidence of gender bias in stem fields Psychol Women Q 2015 39 2 194-209
[30]
Norvig P (2017) On Chomsky and the two cultures of statistical learning. http://norvig.com/chomsky.html. Last visited 17 Dec 2017
[31]
Olson P (2018) The algorithm that helped Google Translate become sexist. https://www.forbes.com/sites/parmyolson/2018/02/15/the-algorithm-that-helped-google-translate-become-sexist/#1c1122c27daa. Last visited 12 Mar 2018
[32]
Papenfuss M (2017) Woman in China says colleague’s face was able to unlock her iPhone X. http://www.huffpostbrasil.com/entry/iphone-face-recognition-double_us_5a332cbce4b0ff955ad17d50. Last visited 17 Dec 2017
[33]
Rixecker K (2018) Google Translate verstärkt sexistische vorurteile. https://t3n.de/news/google-translate-verstaerkt-sexistische-vorurteile-1109449/. Posted 11 Sep 2018
[34]
Santacreu-Vasut E, Shoham A, and Gay V Do female/male distinctions in language matter? Evidence from gender political quotas Appl Econ Lett 2013 20 5 495-498
[35]
Schiebinger L Scientific research must take gender into account Nature 2014 507 7490 9
[36]
Shankland S (2017) Google Translate now serves 200 million people daily. https://www.cnet.com/news/google-translate-now-serves-200-million-people-daily/. Last visited 12 Mar 2018
[37]
Thompson AJ Linguistic relativity: can gendered languages predict sexist attitudes? 2014 Montclair Linguistics Department, Montclair State University
[38]
Wang Y and Kosinski M Deep neural networks are more accurate than humans at detecting sexual orientation from facial images J Personal Soc Psychol 2018 114 2 246-257
[39]
Weaver W (1955) Translation. In: Locke WN, Booth AD (eds) Machine translation of languages, vol 14. Technology Press, MIT, Cambridge, pp 15–23. http://www.mt-archive.info/Weaver-1949.pdf. Last visited 17 Dec 2017
[40]
Women’s Bureau – United States Department of Labor (2017) Traditional and nontraditional occupations. https://www.dol.gov/wb/stats/nontra_traditional_occupations.htm. Last visited 30 May 2018

Cited By

View all
  • (2025)Gender Bias in Natural Language Processing and Computer Vision: A Comparative SurveyACM Computing Surveys10.1145/370043857:6(1-36)Online publication date: 10-Feb-2025
  • (2024)PositionProceedings of the 41st International Conference on Machine Learning10.5555/3692070.3692594(13072-13085)Online publication date: 21-Jul-2024
  • (2024)Fairness Testing of Machine Translation SystemsACM Transactions on Software Engineering and Methodology10.1145/366460833:6(1-27)Online publication date: 27-Jun-2024
  • Show More Cited By

Index Terms

  1. Assessing gender bias in machine translation: a case study with Google Translate
      Index terms have been assigned to the content through auto-classification.

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image Neural Computing and Applications
      Neural Computing and Applications  Volume 32, Issue 10
      May 2020
      1068 pages
      ISSN:0941-0643
      EISSN:1433-3058
      Issue’s Table of Contents

      Publisher

      Springer-Verlag

      Berlin, Heidelberg

      Publication History

      Published: 01 May 2020
      Accepted: 09 March 2019
      Received: 19 October 2018

      Author Tags

      1. Machine bias
      2. Gender bias
      3. Machine learning
      4. Machine translation

      Qualifiers

      • Research-article

      Funding Sources

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)0
      • Downloads (Last 6 weeks)0
      Reflects downloads up to 16 Feb 2025

      Other Metrics

      Citations

      Cited By

      View all
      • (2025)Gender Bias in Natural Language Processing and Computer Vision: A Comparative SurveyACM Computing Surveys10.1145/370043857:6(1-36)Online publication date: 10-Feb-2025
      • (2024)PositionProceedings of the 41st International Conference on Machine Learning10.5555/3692070.3692594(13072-13085)Online publication date: 21-Jul-2024
      • (2024)Fairness Testing of Machine Translation SystemsACM Transactions on Software Engineering and Methodology10.1145/366460833:6(1-27)Online publication date: 27-Jun-2024
      • (2024)Ethics in the Age of AI: An Analysis of AI Practitioners’ Awareness and ChallengesACM Transactions on Software Engineering and Methodology10.1145/363571533:3(1-35)Online publication date: 15-Mar-2024
      • (2024)Learning About Social Context From Smartphone Data: Generalization Across Countries and Daily Life MomentsProceedings of the 2024 CHI Conference on Human Factors in Computing Systems10.1145/3613904.3642444(1-18)Online publication date: 11-May-2024
      • (2024)Travel Demand Forecasting: A Fair AI ApproachIEEE Transactions on Intelligent Transportation Systems10.1109/TITS.2024.339506125:10(14611-14627)Online publication date: 1-Oct-2024
      • (2024)Neural machine translation systems for English to KhasiExpert Systems with Applications: An International Journal10.1016/j.eswa.2023.121813238:PAOnline publication date: 15-Mar-2024
      • (2024)Gender bias perpetuation and mitigation in AI technologies: challenges and opportunitiesAI & Society10.1007/s00146-023-01675-439:4(2045-2057)Online publication date: 1-Aug-2024
      • (2024)Gender-Neutral English to Portuguese Machine Translator: Promoting Inclusive LanguageIntelligent Systems10.1007/978-3-031-79038-6_13(180-195)Online publication date: 17-Nov-2024
      • (2024)Bias Detection and Mitigation in Textual Data: A Study on Fake News and Hate Speech DetectionAdvances in Information Retrieval10.1007/978-3-031-56063-7_29(374-383)Online publication date: 24-Mar-2024
      • Show More Cited By

      View Options

      View options

      Figures

      Tables

      Media

      Share

      Share

      Share this Publication link

      Share on social media