Keywords: Few-shot Image Generation, Generative Adversarial Networks, Meta Learning
TL;DR: We propose a new method for few-shot image generation that is based on learning relationships across datasets using an unsupervised dataset representation, while the distribution of each individual dataset is learned using implicit distributions.
Abstract: Recent generative models such as generative adversarial networks have achieved remarkable success in generating realistic images, but they require large training datasets and computational resources. The goal of few-shot image generation is to learn the distribution of a new dataset from only a handful of examples by transferring knowledge learned from structurally similar datasets. Towards achieving this goal, we propose the “Implicit Support Set Autoencoder” (ISSA) that adversarially learns the relationship across datasets using an unsupervised dataset representation, while the distribution of each individual dataset is learned using implicit distributions. Given a few examples from a new dataset, ISSA can generate new samples by inferring the representation of the underlying distribution using a single forward pass. We showcase significant gains from our method on generating high quality and diverse images for unseen classes in the Omniglot and CelebA datasets in few-shot image generation settings.
Contribution Process Agreement: Yes
Poster Session Selection: Poster session #2 (15:00 UTC), Poster session #3 (16:50 UTC)