Local plasticity rules can learn deep representations using self-supervised contrastive predictionsDownload PDF

21 May 2021, 20:45 (modified: 25 Oct 2021, 09:48)NeurIPS 2021 PosterReaders: Everyone
Keywords: Synaptic plasticity, Hebbian learning, deep learning, self-supervised learning, contrastive predictive coding
TL;DR: We propose a new local learning rule, inspired by neuroscience and self-supervised deep learning, and demonstrate that it can train deep neural networks on images, speech and video.
Abstract: Learning in the brain is poorly understood and learning rules that respect biological constraints, yet yield deep hierarchical representations, are still unknown. Here, we propose a learning rule that takes inspiration from neuroscience and recent advances in self-supervised deep learning. Learning minimizes a simple layer-specific loss function and does not need to back-propagate error signals within or between layers. Instead, weight updates follow a local, Hebbian, learning rule that only depends on pre- and post-synaptic neuronal activity, predictive dendritic input and widely broadcasted modulation factors which are identical for large groups of neurons. The learning rule applies contrastive predictive learning to a causal, biological setting using saccades (i.e. rapid shifts in gaze direction). We find that networks trained with this self-supervised and local rule build deep hierarchical representations of images, speech and video.
Supplementary Material: pdf
Code Of Conduct: I certify that all co-authors of this work have read and commit to adhering to the NeurIPS Statement on Ethics, Fairness, Inclusivity, and Code of Conduct.
Code: https://github.com/EPFL-LCN/pub-illing2021-neurips
15 Replies