LSGANs with Gradient Regularizers are Smooth High-dimensional InterpolatorsDownload PDF

06 Oct 2022 (modified: 05 May 2023)INTERPOLATE at NeurIPS 2022Readers: Everyone
Keywords: generative adversarial networks, Least-squares GAN, higher-order gradient penalty, variational calculus, polyharmonic function, radial basis function
TL;DR: We link the optimal discriminator in gradient-regularized LSGANs to interpolation. Enforcing interpolation-based image-space penalties on LSGAN results in an RBF interpolator on the class labels, with weights given by solving a system of equations.
Abstract: We consider the problem of discriminator optimization in least-squares generative adversarial networks (LSGANs) subject to higher-order gradient regularization enforced on the convex hull of all possible interpolation points between the target (real) and generated (fake) data. We analyze the proposed LSGAN cost within a variational framework, and show that the optimal discriminator solves a regularized least-squares problem, and can be represented through a polyharmonic radial basis function (RBF) interpolator. The optimal RBF discriminator can be implemented in closed-form, with the weights computed by solving a linear system of equations. We validate the proposed approach on synthetic Gaussian and standard image datasets. While the optimal LSGAN discriminator leads to a superior convergence on Gaussian data, the inherent low-dimensional manifold structure of images makes the implementation of the optimal discriminator ill-posed. Nevertheless, replacing the trainable discriminator network with a closed-form RBF interpolator results in superior convergence on 2-D Gaussian data, while overcoming pitfalls in GAN training, namely mode dropping and mode collapse.
0 Replies

Loading