Keywords: In-Context Learning, Representation Learning, Representation Geometry, Large Language Models, Task Representation
TL;DR: Large Language Models can form structural representations for tasks solely given in-context exemplars.
Abstract: Representation learning has been central to deep learning’s evolution. While interpretable structures have been observed in pre-trained models’ representations, an important question arises: Do networks develop such interpretable structures during in-context learning? Using synthetic sequence data derived from underlying geometrically structured graphs (e.g., grids, rings), we provide affirmative evidence that language models develop internal representations mirroring these geometric structures during in-context learning. Furthermore, we demonstrate how in-context examples can override semantic priors by constructing a representation in dimensions other than the one used by the prior. Overall, our study demonstrates that models can form meaningful representations solely from in-context exemplars.
Submission Number: 82
Loading