On the Power of Context Enhanced Learning in LLMs

Published: 06 Mar 2025, Last Modified: 08 Apr 2025ICLR 2025 DeLTa Workshop PosterEveryoneRevisionsBibTeXCC BY 4.0
Track: long paper (up to 8 pages)
Keywords: In-Context Learning, Sample Efficiency, Knowledge Internalization, Privileged Information, Multi-step reasoning
TL;DR: We formalize Context-Enhanced Learning, a paradigm where LLMs leverage context without direct gradient updates, achieving exponential sample efficiency gains over standard autoregressive training when the model has ICL capabilities.
Abstract: We formalize a new concept for LLMs, **context-enhanced learning**. It involves standard gradient-based learning on text except that the context is enhanced with additional data on which no auto-regressive gradients are computed. This setting is a gradient-based analog of usual in-context learning (ICL) and appears in some recent works. Using a multi-step reasoning task, we prove in a simplified setting that context-enhanced learning can be **exponentially more sample-efficient** than standard learning when the model is capable of ICL. At a mechanistic level, we find that the benefit of context-enhancement arises from a more accurate gradient learning signal. Our findings highlight the potential of context-enhanced learning to bridge gradient-based learning and in-context learning, offering new insights into their interplay.
Submission Number: 38
Loading