Associated Learning: an Alternative to End-to-End Backpropagation that Works on CNN, RNN, and Transformer
Keywords: pipeline training, parallel training, backpropagation, associated learning
Abstract: This paper studies Associate Learning (AL), an alternative methodology to the end-to-end backpropagation (BP). We introduce the workflow to convert a neural network into a proper structure such that AL can be used to learn the weights for various types of neural networks. We compared AL and BP on some of the most successful types of neural networks -- Convolutional Neural Network (CNN), Recurrent Neural Network (RNN), and Transformer. Experimental results show that AL consistently outperforms BP on various open datasets. We discuss possible reasons for AL's success and its limitations.
One-sentence Summary: This paper studies Associate Learning, an alternative methodology to the end-to-end backpropagation
Supplementary Material: zip