CIMemories: A Compositional Benchmark For Contextual Integrity In LLMs

ICLR 2026 Conference Submission22216 Authors

20 Sept 2025 (modified: 08 Oct 2025)ICLR 2026 Conference SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Contextual Integrity; Inference-time Privacy; Input-output flow
TL;DR: CIMemories is a dataset of synthetic user profiles paired with recipient–task contexts that simulates persistent,cross-session LLM “memory” to evaluate whether models use long-term context appropriately—sharing what’s needed while avoiding leaks.
Abstract: Large Language Models (LLMs) increasingly use persistent memory from past interactions to enhance personalization and task performance. However, this memory creates critical risks when sensitive information is revealed in inappropriate contexts. We present CIMemories, a benchmark for evaluating whether LLMs appropriately control information flow from memory based on task context. CIMemories uses synthetic user profiles with 100+ attributes per user, paired with various task contexts where each attribute may be essential for some tasks but inappropriate for others. For example, mental health details are necessary for booking therapy but inappropriate when requesting time off from work. This design enables two forms of compositionality: (1) flexible memory composition by varying which attributes are necessary versus inappropriate across different settings, and (2) multi-task composition per user, measuring cumulative information disclosure across sessions. Our evaluation reveals frontier models exhibit between 14%-69% attribute-level violations (leaking inappropriate information), and that higher task completeness (sharing necessary information) is accompanied by increased violations, highlighting critical gaps in integrity-aware memory systems.
Supplementary Material: zip
Primary Area: datasets and benchmarks
Submission Number: 22216
Loading