Wisdom from Diversity: Bias Mitigation Through Hybrid Human-LLM Crowds

Published: 15 Jun 2025, Last Modified: 07 Aug 2025AIA 2025EveryoneRevisionsBibTeXCC BY 4.0
Keywords: collective intelligence, large language models, wisdom of the crowd, bias mitigation, hybrid crowds, human-AI collaboration, ensemble methods
TL;DR: We show that hybrid human-LLM crowds outperform either humans or LLMs alone in a false headline identification task.
Abstract: Despite their performance, large language models (LLMs) can inadvertently perpetuate biases found in the data they are trained on. By analyzing LLM responses to bias-eliciting headlines, we find that these models often mirror human biases. To address this, we explore crowd-based strategies for mitigating bias through response aggregation. We first demonstrate that simply averaging responses from multiple LLMs, intended to leverage the "wisdom of the crowd", can exacerbate existing biases due to the limited diversity within LLM crowds. In contrast, we show that locally weighted aggregation methods more effectively leverage the wisdom of the LLM crowd, achieving both bias mitigation and improved accuracy. Finally, recognizing the complementary strengths of LLMs (accuracy) and humans (diversity), we demonstrate that hybrid crowds containing both significantly enhance performance and further reduce biases across ethnic and gender-related contexts.
Paper Type: Previously Published Paper
Venue For Previously Published Paper: IJCAI 2025
Submission Number: 22
Loading