Track: Technical
Keywords: GenAI, Large language models, Systematic bias, Bias probing, ChatGPT, Responsible AI, Healthcare
TL;DR: This paper advocates for a systematic approach to bias probing in LLMs by analyzing output distributions across various prompts using healthcare as a case study to better predict generalization of biased behaviors.
Abstract: There is a growing body of literature exposing social biases of LLMs. However, these works often focus on a specific protected group, a specific prompt type and a specific decision task. Given the large and complex input-output space of LLMs, case-by-case analyses alone may not paint a picture of the systematic biases of these models.
In this paper, we argue for broad and systematic bias probing. We propose to do so by comparing the distribution of outputs
over a wide range of prompts, multiple protected attributes and across different realistic decision making settings in the
same application domain. We demonstrate this approach for three personalized healthcare advice-seeking settings.
We argue that studying the complex patterns of bias across tasks helps us better anticipate
the way behaviors (specifically biased behaviors) of LLMs might generalize to new tasks.
Submission Number: 70
Loading