Keywords: continual deep learning, synthetic data generation, continual learning benchmarks, computer vision, catastrophic interference
TL;DR: We introduce a graphics simulator to flexibly compose datasets for deep continual learning.
Abstract: Several families of continual learning techniques have been proposed to alleviate catastrophic interference in deep neural network training on non-stationary data. However, a comprehensive comparison and analysis of limitations remains largely open due to the inaccessibility to suitable datasets. Empirical examination not only varies immensely between individual works, it further currently relies on contrived composition of benchmarks through subdivision and concatenation of various prevalent static vision datasets. In this work, our goal is to bridge this gap by introducing a computer graphics simulation framework that repeatedly renders only upcoming urban scene fragments in an endless real-time procedural world generation process. At its core lies a modular parametric generative model with adaptable generative factors. The latter can be used to flexibly compose data streams, which significantly facilitates a detailed analysis and allows for effortless investigation of various continual learning schemes.
Supplementary Material: zip
URL: See paper contributions - Simulator Standalone Executable: https://doi.org/10.5281/zenodo.4899294 ; Simulator Source Code: https://github.com/ccc-frankfurt/EndlessCL-Simulator-Source ; Generated Experimental Datasets: https://doi.org/10.5281/zenodo.4899267
Community Implementations: [![CatalyzeX](/images/catalyzex_icon.svg) 2 code implementations](https://www.catalyzex.com/paper/a-procedural-world-generation-framework-for/code)
10 Replies
Loading