Can Bias in Large Language Models for Tabular Data Classification Be Mitigated?

ACL ARR 2024 December Submission737 Authors

15 Dec 2024 (modified: 05 Feb 2025)ACL ARR 2024 December SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Abstract: Large Language Models (LLMs) perform well in tabular prediction tasks with limited data, using their ability to understand instructions and learn from examples. However, their reliance on training data can perpetuate social biases, leading to unfair outcomes and disproportionately impacting underprivileged groups. Addressing these biases is critical as LLMs see wider adoption in tabular data tasks. Traditional bias mitigation strategies in machine learning, such as balancing datasets or applying fairness constraints, are less effective with LLMs. Our research explores whether bias in LLMs for tabular data classification can be mitigated. Through extensive experiments, we found that using LLMs in a zero-shot setting introduces bias, and in-context learning slightly reduces the bias. Meanwhile, fine-tuning and retrieval augmented generation show limited effectiveness in bias mitigation. We introduced three instruction-based prompting strategies to enhance fairness: Fair Prompting, Generalised Prompting, and Descriptive Prompting. The results show that combining descriptive prompting with in-context learning, particularly the Equal Samples Across Demographics approach, significantly improved fairness metrics such as Statistical Parity Ratio and Equal Opportunity Ratio and yielded accuracy gains ranging from 3.27% to 15.05% across multiple datasets, underscoring its potential as a powerful strategy in the ongoing effort to mitigate bias in LLMs.
Paper Type: Long
Research Area: Ethics, Bias, and Fairness
Research Area Keywords: fairness and bias in NLP, model bias/fairness evaluation, model bias/unfairness mitigation
Contribution Types: Model analysis & interpretability, Data analysis
Languages Studied: English
Submission Number: 737
Loading