Feed-Forward Networks with Attention Can Solve Some Long-Term Memory ProblemsDownload PDF

20 Apr 2024 (modified: 12 Feb 2016)ICLR 2016 workshop submissionReaders: Everyone
Abstract: We propose a simplified model of attention which is applicable to feed-forward neural networks and demonstrate that the resulting model can solve the synthetic "addition" and "multiplication" long-term memory problems for sequence lengths which are both longer and more widely varying than the best published results for these tasks.
Conflicts: columbia.edu, google.com
9 Replies

Loading