LLMs learn governing principles of dynamical systems, revealing an in-context neural scaling law

Published: 18 Jun 2024, Last Modified: 08 Jul 2024ICML 2024 Workshop ICL PosterEveryoneRevisionsBibTeXCC BY 4.0
Track: short paper (up to 4 pages)
Keywords: in-context learning, ICL, dynamical systems, time-series forecasting, neural scaling law, world model, physics, LLM
TL;DR: LLMs can accurately learn the transition rules of dynamical systems, and obey an in-context neural scaling law.
Abstract: Pretrained large language models (LLMs) are surprisingly effective at performing zero-shot tasks, including time-series forecasting. However, understanding the mechanisms behind such capabilities remains highly challenging due to the complexity of the models. We study LLMs' ability to extrapolate the behavior of dynamical systems whose evolution is governed by principles of physical interest. Our results show that LLaMA 2, a language model trained primarily on texts, achieves accurate predictions of dynamical system time series without fine-tuning or prompt engineering. Moreover, the accuracy of the learned physical rules increases with the length of the input context window, revealing an in-context version of neural scaling law. Along the way, we present a flexible and efficient algorithm for extracting probability density functions of multi-digit numbers directly from LLMs.
Submission Number: 30
Loading