Learning to Be Fair: Modeling Fairness Dynamics by Simulating Moral-Based Multi-Agent Resource Allocation
Track: Main Papers Track (6 to 9 pages)
Keywords: fairness, multi-agent simulation, consensus, llm-based agents, morality
TL;DR: We simulate moral-based agents in repeated resource-allocation games and show that social feedback drives convergence toward shared fairness norms, with trends that align with human judgments.
Abstract: Fairness is a foundational social construct for stable, resilient societies, yet its meaning is dynamic, context-dependent, and inherently subjective. This multifaceted nature reveals a gap between traditional social science and contemporary computational approaches: the former offers rich conceptual accounts but limited computational models, while the latter often relies on static objectives or purely data-driven criteria that overlook the subjective and communicative nature of fairness.
We address this gap through a computational framework and two resource-allocation scenarios in which large language model (LLM)–based cognitive agents operate with heterogeneous roles, relationships, and moral commitments. The framework supports agent reflection and negotiation via explicit, language-based feedback, enabling the study of norm evolution and consensus formation of fairness in multi-agent social systems.
Using standard objective metrics from resource allocation, we analyze how our approach captures the complexities of fairness, such as ambiguity, procedural justice, and subjective satisfaction—while remaining quantitatively evaluable. This work also offers insights for designing tractable AI systems that can navigate evolving social norms in dynamic, multi-stakeholder environments.
Submission Number: 34
Loading