Structured In-Context Task Representations

Published: 23 Oct 2024, Last Modified: 24 Feb 2025NeurReps 2024 PosterEveryoneRevisionsBibTeXCC BY 4.0
Keywords: In-Context Learning, Representation Learning, Representation Geometry, Large Language Models, Task Representation
TL;DR: Large Language Models can form structural representations for tasks solely given in-context exemplars.
Abstract: Representation learning has been central to deep learning’s evolution. While interpretable structures have been observed in pre-trained models’ representations, an important question arises: Do networks develop such interpretable structures during in-context learning? Using synthetic sequence data derived from underlying geometrically structured graphs (e.g., grids, rings), we provide affirmative evidence that language models develop internal representations mirroring these geometric structures during in-context learning. Furthermore, we demonstrate how in-context examples can override semantic priors by constructing a representation in dimensions other than the one used by the prior. Overall, our study demonstrates that models can form meaningful representations solely from in-context exemplars.
Submission Number: 82
Loading

OpenReview is a long-term project to advance science through improved peer review with legal nonprofit status. We gratefully acknowledge the support of the OpenReview Sponsors. © 2025 OpenReview