Adversarial training for predictive tasks: theoretical analysis and limitations in the deterministic case.Download PDF

Published: 09 Dec 2020, Last Modified: 05 May 2023ICBINB 2020 PosterReaders: Everyone
Keywords: Adversarial training, conditional GAN, Wasserstein, DNN, predictive network
TL;DR: We propose a theoretical analysis of conditional GAN applied to predictive tasks, and of its limitations in the deterministic case.
Abstract: To train a deep neural network to mimic the outcomes of processing sequences, a version of Conditional Generalized Adversarial Network (CGAN) can be used. It has been observed by others that CGAN can help to improve the results even for deterministic sequences, where only one output is associated with the processing of a given input. Surprisingly, our CGAN-based tests on deterministic geophysical processing sequences did not produce a real improvement compared to the use of an $L_p$ loss; we here propose a first theoretical explanation why. Our analysis goes from the non-deterministic case to the deterministic one. It led us to develop an adversarial way to train a content loss that gave better results on our data.
1 Reply

Loading