Bias Transmission in Large Language Models: Evidence from Gender-Occupation Bias in GPT-4

Published: 28 Jun 2024, Last Modified: 25 Jul 2024NextGenAISafety 2024 PosterEveryoneRevisionsBibTeXCC BY 4.0
Keywords: LLM Bias, Implicit Bias, Gender Bias, Generative AI, GPT
TL;DR: Association-based biases in LLMs (e.g., surgeon=male, nurse=female) do not automatically translate into biases in model behavior (e.g.,male surgical applicants receiving stronger materials or ratings than female applicants).
Abstract: Recent advances in generative AI are poised to reduce the burden of important and arduous tasks, including drafting job application materials. In this paper, we examine whether GPT-4 produces job cover letters that systematically advantage some users and disadvantage others. To test this, we introduce a novel method designed to probe LLMs for gender-occupation biases. Using our method, we show that GPT-4, like humans, possesses strong gender-occupation associations (e.g., surgeon = male, nurse = female). However, surprisingly, we find that biased associations do not necessarily translate into biased results. That is, we find that GPT-4 can (a) produce reasonable evaluations of cover letters, (b) evaluate information written by men and women equally, unlike humans, and (c) generate equally strong cover letters for male and female applicants. Our work calls for more systematic studies of the connection between association bias and outcome bias in generative AI models.
Submission Number: 66
Loading