Data-Efficient GAN Training Beyond (Just) Augmentations: A Lottery Ticket PerspectiveDownload PDF

21 May 2021, 20:47 (edited 22 Oct 2021)NeurIPS 2021 PosterReaders: Everyone
  • Keywords: Data-Efficient GAN Training, The Lottery Tickets Hypothesis
  • TL;DR: Treating the lottery ticket as an inductive prior, we provide a brand-new angle for the data-hungry GAN training, that is orthogonal to augmentation-based methods.
  • Abstract: Training generative adversarial networks (GANs) with limited real image data generally results in deteriorated performance and collapsed models. To conquer this challenge, we are inspired by the latest observation, that one can discover independently trainable and highly sparse subnetworks (a.k.a., lottery tickets) from GANs. Treating this as an inductive prior, we suggest a brand-new angle towards data-efficient GAN training: by first identifying the lottery ticket from the original GAN using the small training set of real images; and then focusing on training that sparse subnetwork by re-using the same set. We find our coordinated framework to offer orthogonal gains to existing real image data augmentation methods, and we additionally present a new feature-level augmentation that can be applied together with them. Comprehensive experiments endorse the effectiveness of our proposed framework, across various GAN architectures (SNGAN, BigGAN, and StyleGAN-V2) and diverse datasets (CIFAR-10, CIFAR-100, Tiny-ImageNet, ImageNet, and multiple few-shot generation datasets). Codes are available at:
  • Supplementary Material: pdf
  • Code Of Conduct: I certify that all co-authors of this work have read and commit to adhering to the NeurIPS Statement on Ethics, Fairness, Inclusivity, and Code of Conduct.
  • Code:
18 Replies