In-Context Fine-Tuning for Neural Operators

Published: 03 Mar 2025, Last Modified: 09 Apr 2025AI4MAT-ICLR-2025 PosterEveryoneRevisionsBibTeXCC BY 4.0
Submission Track: Paper Track (Tiny Paper)
Submission Category: AI-Guided Design
Keywords: neural operators, in-context learning
TL;DR: In-context learning enables transformers to be rapidly fine-tuned for new tasks; we demonstrate a similar phenomenon holds for a generalization of transformers to function spaces that allow for rapid learning of the solution operator across PDEs.
Abstract:

Neural operator surrogate models are becoming increasingly popular for material design, where they are used to rapidly evaluate candidate designs. Such surrogate models, however, most commonly either amortize the flow map across initial conditions or the map from a fixed initial condition across design parameters. Neural operator surrogates are fundamentally unsuited for learning across both simultaneously, as the underlying flow map operator of interest varies across system parameters. Towards this end, recent interest has emerged in the use of PDE foundation models that can subsequently be fine-tuned for particular system parameters of interest. Such expensive fine-tuning, however, is rarely feasible in downstream engineering design use cases. To combat this, we instead propose to perform fine-tuning via a generalization of the in-context learning exhibited by transformers to operator spaces. In particular, we propose a generalization of the transformer that handles sequences of functions, dubbed the "OpFormer,"" and formally demonstrate such an architecture is capable of performing in-context learning by implementing a form of gradient descent in an operator RKHS.

Submission Number: 3
Loading