Hot PATE: Private Aggregation of Distributions for Diverse Tasks

ICLR 2026 Conference Submission13904 Authors

18 Sept 2025 (modified: 08 Oct 2025)ICLR 2026 Conference SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Differential Privacy, Sequential Text Generation, Coordinated Ensembles
TL;DR: Hot PATE is a PATE variant for generative tasks that preserves output diversity, provably transfers it without extra privacy cost, and yields orders-of-magnitude higher utility at the same privacy level.
Abstract: The Private Aggregation of Teacher Ensembles (PATE) framework enables privacy-preserving machine learning by aggregating responses from disjoint subsets of sensitive data. Adaptations of PATE to tasks with inherent output diversity such as text generation, where the desired output is a sample from a distribution, face a core tension: as diversity increases, samples from different teachers are less likely to agree, but lower agreement results in reduced utility for the same privacy requirements. Yet suppressing diversity to artificially increase agreement is undesirable, as it distorts the output of the underlying model, and thus reduces output quality. We propose Hot PATE, a variant of PATE designed for diverse generative settings. We formalize the notion of a \emph{diversity-preserving} \emph{ensemble sampler} and introduce an efficient sampler that provably transfers diversity without incurring additional privacy cost. Hot PATE requires only API access to proprietary models and can be used as a drop-in replacement for existing "cold" PATE samplers. Our empirical results corroborate the theoretical guarantees, showing that Hot PATE achieves orders-of-magnitude improvements in utility per privacy budget on in-context learning tasks.
Primary Area: alignment, fairness, safety, privacy, and societal considerations
Submission Number: 13904
Loading