Path Independent Equilibrium Models Can Better Exploit Test-Time ComputationDownload PDF

Published: 31 Oct 2022, Last Modified: 16 Dec 2022NeurIPS 2022 AcceptReaders: Everyone
Keywords: equilibrium models, attractor networks, test time compute, out of distribution generalization
TL;DR: We find that equilibrium models are capable of achieving improved test performance with more test time compute, and identify a mechanism - named Path Independence - that facilitates this capability.
Abstract: Designing networks capable of attaining better performance with an increased inference budget is important to facilitate generalization to harder problem instances. Recent efforts have shown promising results in this direction by making use of depth-wise recurrent networks. In this work, we reproduce the performance of the prior art using a broader class of architectures called equilibrium models, and find that stronger generalization performance on harder examples (which require more iterations of inference to get correct) strongly correlates with the path independence of the system—its ability to converge to the same attractor (or limit cycle) regardless of initialization, given enough computation. Experimental interventions made to promote path independence result in improved generalization on harder (and thus more compute-hungry) problem instances, while those that penalize it degrade this ability. Path independence analyses are also useful on a per-example basis: for equilibrium models that have good in-distribution performance, path independence on out-of-distribution samples strongly correlates with accuracy. Thus, considering equilibrium models and path independence jointly leads to a valuable new viewpoint under which we can study the generalization performance of these networks on hard problem instances.
Supplementary Material: zip
24 Replies