Women Are Beautiful, Men Are Leaders: Gender Stereotypes in Machine Translation and Language Modeling

ACL ARR 2024 June Submission393 Authors

10 Jun 2024 (modified: 02 Aug 2024)ACL ARR 2024 June SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Abstract: We present GEST -- a new dataset for measuring gender-stereotypical reasoning in language models and machine translation systems. GEST contains samples for 16 gender stereotypes about men and women (e.g., Women are beautiful, Men are leaders) that are compatible with the English language and 9 Slavic languages. The definition of said stereotypes was informed by gender experts. We used GEST to evaluate English and Slavic masked LMs, English generative LMs, and machine translation systems. We discovered significant and consistent amounts of gender-stereotypical reasoning in almost all the evaluated models and languages. Our experiments confirm the previously postulated hypothesis that the larger the model, the more biased it usually is.
Paper Type: Long
Research Area: Ethics, Bias, and Fairness
Research Area Keywords: model bias/fairness evaluation
Contribution Types: Model analysis & interpretability, Approaches to low-resource settings, Data resources
Languages Studied: English,Belarusian,Russian,Ukrainian,Croatian,Serbian,Slovene,Czech,Polish,Slovak
Submission Number: 393
Loading