Abstract: Audio signals are sampled at high temporal resolutions, and learning to synthesize audio requires capturing structure across a range of timescales. Generative adversarial networks (GANs) have seen wide success at generating images that are both locally and globally coherent, but they have seen little application to audio generation. In this paper we introduce WaveGAN, a first attempt at applying GANs to unsupervised synthesis of raw-waveform audio. WaveGAN is capable of synthesizing one second slices of audio waveforms with global coherence, suitable for sound effect generation. Our experiments demonstrate that—without labels—WaveGAN learns to produce intelligible words when trained on a small-vocabulary speech dataset, and can also synthesize audio from other domains such as drums, bird vocalizations, and piano. We compare WaveGAN to a method which applies GANs designed for image generation on image-like audio feature representations, finding both approaches to be promising.
Keywords: audio, waveform, spectrogram, GAN, adversarial, WaveGAN, SpecGAN
TL;DR: Learning to synthesize raw waveform audio with GANs
Code: [![github](/images/github_icon.svg) chrisdonahue/wavegan](https://github.com/chrisdonahue/wavegan) + [![Papers with Code](/images/pwc_icon.svg) 20 community implementations](https://paperswithcode.com/paper/?openreview=ByMVTsR5KQ)
Community Implementations: [![CatalyzeX](/images/catalyzex_icon.svg) 19 code implementations](https://www.catalyzex.com/paper/adversarial-audio-synthesis/code)
11 Replies
Loading