Connecting the Dots Between MLE and RL for Sequence GenerationDownload PDF

Published: 10 Apr 2019, Last Modified: 05 May 2023drlStructPred 2019Readers: Everyone
Keywords: sequence generation, maximum likelihood learning, reinforcement learning, policy optimization, text generation, reward augmented maximum likelihood, exposure bias
TL;DR: A unified perspective of various learning algorithms for sequence generation, such as MLE, RL, RAML, data noising, etc.
Abstract: Sequence generation models such as recurrent networks can be trained with a diverse set of learning algorithms. For example, maximum likelihood learning is simple and efficient, yet suffers from the exposure bias problem. Reinforcement learning like policy gradient addresses the problem but can have prohibitively poor exploration efficiency. A variety of other algorithms such as RAML, SPG, and data noising, have also been developed in different perspectives. This paper establishes a formal connection between these algorithms. We present a generalized entropy regularized policy optimization formulation, and show that the apparently divergent algorithms can all be reformulated as special instances of the framework, with the only difference being the configurations of reward function and a couple of hyperparameters. The unified interpretation offers a systematic view of the varying properties of exploration and learning efficiency. Besides, based on the framework, we present a new algorithm that dynamically interpolates among the existing algorithms for improved learning. Experiments on machine translation and text summarization demonstrate the superiority of the proposed algorithm.
6 Replies