Idiosyncrasies in Large Language Models

Published: 01 May 2025, Last Modified: 18 Jun 2025ICML 2025 posterEveryoneRevisionsBibTeXCC BY 4.0
TL;DR: We study idiosyncrasies in LLMs.
Abstract: In this work, we unveil and study idiosyncrasies in Large Language Models (LLMs) -- unique patterns in their outputs that can be used to distinguish the models. To do so, we consider a simple classification task: given a particular text output, the objective is to predict the source LLM that generates the text. We evaluate this synthetic task across various groups of LLMs and find that simply fine-tuning text embedding models on LLM-generated texts yields excellent classification accuracy. Notably, we achieve 97.1\% accuracy on held-out validation data in the five-way classification problem involving ChatGPT, Claude, Grok, Gemini, and DeepSeek. Our further investigation reveals that these idiosyncrasies are rooted in word-level distributions. These patterns persist even when the texts are rewritten, translated, or summarized by an external LLM, suggesting that they are also encoded in the semantic content. Additionally, we leverage LLM as judges to generate detailed, open-ended descriptions of each model's idiosyncrasies. Finally, we discuss the broader implications of our findings, including training on synthetic data, inferring model similarity, and robust evaluation of LLMs.
Lay Summary: We study the problem of distinguishing large language models (LLMs) based on their outputs. We find that neural networks can reliably distinguish the texts from different LLMs. Our results have several implications: first it can help people develop tools for detecting LLM generated texts; it could shed light on the relative uptake of different LLMs, beyond what is reported by individual companies, and on the nature of data used to build different models in the first place.
Link To Code: https://github.com/locuslab/llm-idiosyncrasies
Primary Area: Deep Learning->Large Language Models
Keywords: Large Language Models; Idiosyncrasies; Dataset bias; Synthetic data;
Submission Number: 2660
Loading