Fake Player: Imitating Real Player to Distill Data for LLM-based NPC Training

ICLR 2026 Conference Submission17228 Authors

19 Sept 2025 (modified: 08 Oct 2025)ICLR 2026 Conference SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Keywords: LLM, NPC, Data Distillation
TL;DR: FakePlayer​​: Multi-Agent Simulation of Expression-Constrained Human Players for Automated High-Quality Data Distillation in LLM-based NPC Training.
Abstract: In the era of large language models (LLMs), games increasingly deploy LLM-based role-playing NPCs to replace traditional scripted NPCs, enabling more intelligent and dynamic interactions. To ensure persona consistency and output stability, these NPCs require fine-tuning for alignment, utilizing training data with dual dimensions: persona-aligned assistant responses and diverse, authentic user inputs reflecting real player behaviors. However, existing research prioritizes persona consistency in NPC responses while neglecting the diversity and authenticity of user-side inputs. This critical gap leads to NPC responses that are misaligned with genuine player interactions, significantly impairing player immersion and experience. Human annotation struggles to address this gap due to its inability to comprehensively cover the vast spectrum of player behaviors. Moreover, practical deployment constraints strongly favor small-parameter LLMs for NPCs, making data quality paramount. To bridge this gap, we propose: 1) $\textbf{Fake Player}$: A multi-agent LLM distillation framework where collaborative agents simulate expression-constrained human players to distill diverse, human-aligned dialogue data from large LLMs; 2) $\textbf{Distill Bench}$: A standardized benchmark for quantitatively assessing distilled data quality, bypassing costly NPC retraining. Extensive experiments validate our method’s effectiveness in generating diverse player interactions and the benchmark’s reliability for data evaluation.
Supplementary Material: zip
Primary Area: applications to computer vision, audio, language, and other modalities
Submission Number: 17228
Loading