SBO-RNN: Reformulating Recurrent Neural Networks via Stochastic Bilevel OptimizationDownload PDF

21 May 2021, 20:44 (modified: 28 Jan 2022, 15:44)NeurIPS 2021 PosterReaders: Everyone
Keywords: recurrent neural networks, stochastic bilevel optimization, training stability
TL;DR: SBO-RNN: Reformulating Recurrent Neural Networks via Stochastic Bilevel Optimization
Abstract: In this paper we consider the training stability of recurrent neural networks (RNNs) and propose a family of RNNs, namely SBO-RNN, that can be formulated using stochastic bilevel optimization (SBO). With the help of stochastic gradient descent (SGD), we manage to convert the SBO problem into an RNN where the feedforward and backpropagation solve the lower and upper-level optimization for learning hidden states and their hyperparameters, respectively. We prove that under mild conditions there is no vanishing or exploding gradient in training SBO-RNN. Empirically we demonstrate our approach with superior performance on several benchmark datasets, with fewer parameters, less training data, and much faster convergence. Code is available at
Supplementary Material: pdf
Code Of Conduct: I certify that all co-authors of this work have read and commit to adhering to the NeurIPS Statement on Ethics, Fairness, Inclusivity, and Code of Conduct.
24 Replies