Track: long paper (up to 10 pages)
Domain: neuroscience
Abstract: Sequence working memory (SWM) is the ability to remember a sequence of items and recall them in order. Recent neural recordings in the prefrontal cortex during SWM found “slot”-like representations, where each item in the sequence is stored in a separate low-dimensional subspace of population activity. However, it is unclear what circuit structure could give rise to such representations, and if such slot-like representation naturally arises from the constraints of the SWM task. Here, we trained recurrent neural networks (RNNs) on a SWM task using standard architectures and training procedures. Two types of networks emerged, relying on two types of representations: (1) is the brain-like fixed slot representation, (2) stores items in time-varying subspaces. They lead to different length generalization behaviors: (1) generalizes learned slots to novel sequence lengths, while (2) makes sequence order errors on novel sequence lengths. For (1), we delve into the network weights and identify a simple, interpretable circuit with modular components that implements slots. Our work bridges biological and artificial networks by demonstrating 1) more brain-like representations in artificial networks (which can emerge from standard architecture and training) may lead to improved generalization, and 2) how one can extract interpretable circuit structures from these brain-like artificial networks to serve as hypotheses for the brain.
Submission Number: 22
Loading