Bound by semanticity: universal laws governing the generalization-identification tradeoff

ICLR 2026 Conference Submission16680 Authors

19 Sept 2025 (modified: 08 Oct 2025)ICLR 2026 Conference SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Keywords: generalization, multi-object reasoning, cognitive science
Abstract: Intelligent systems must deploy internal representations that are simultaneously structured—to support broad generalization—and selective—to preserve input identity. We expose a fundamental limit on this tradeoff. For any model whose representational similarity between inputs decays with finite semantic resolution, we derive closed‑form expressions that pin its probability of correct generalization $p_S$ and identification $p_I$ to a universal Pareto front independent of input space geometry. Extending the analysis to noisy, heterogeneous spaces and to inputs $n>2$ predicts a sharp $1/n$ collapse of multi-input processing capacity and a non‑monotonic optimum for $p_S$. A minimal ReLU network trained end‑to‑end reproduces these laws: during learning a resolution boundary self‑organizes and empirical $(p_S,p_I)$ trajectories closely follow theoretical curves for linearly decaying similarity. Finally, we demonstrate that the same limits persist in two markedly more complex settings—a convolutional neural network and state‑of‑the‑art vision–language models—confirming that finite‑resolution similarity is a fundamental emergent informational constraint, not merely a toy‑model artifact. Together, these results provide an exact theory of the generalization‑identification trade‑off and clarify how semantic resolution shapes the representational capacity of deep networks and brains alike.
Primary Area: applications to neuroscience & cognitive science
Submission Number: 16680
Loading