Sparse Attentive Backtracking: Long-Range Credit Assignment in Recurrent NetworksDownload PDF

15 Feb 2018 (modified: 13 Jul 2023)ICLR 2018 Conference Blind SubmissionReaders: Everyone
Abstract: A major drawback of backpropagation through time (BPTT) is the difficulty of learning long-term dependencies, coming from having to propagate credit information backwards through every single step of the forward computation. This makes BPTT both computationally impractical and biologically implausible. For this reason, full backpropagation through time is rarely used on long sequences, and truncated backpropagation through time is used as a heuristic. However, this usually leads to biased estimates of the gradient in which longer term dependencies are ignored. Addressing this issue, we propose an alternative algorithm, Sparse Attentive Backtracking, which might also be related to principles used by brains to learn long-term dependencies. Sparse Attentive Backtracking learns an attention mechanism over the hidden states of the past and selectively backpropagates through paths with high attention weights. This allows the model to learn long term dependencies while only backtracking for a small number of time steps, not just from the recent past but also from attended relevant past states.
TL;DR: Towards Efficient Credit Assignment in Recurrent Networks without Backpropagation Through Time
Keywords: recurrent neural networks, long-term dependencies, back-propagation through time, truncated back-propagation, biological inspiration, self-attention
11 Replies

Loading