Progressive Growing of Neural ODEsDownload PDF

Published: 27 Feb 2020, Last Modified: 05 May 2023ICLR 2020 Workshop ODE/PDE+DL PosterReaders: Everyone
Keywords: Neural ODEs, Curriculum Learning, Progressive Growing, Time Series
TL;DR: Progressively growing Neural ODE network and data complexity to learn complex time-series.
Abstract: Neural Ordinary Differential Equations (NODEs) have proven to be a powerful modeling tool for approximating (interpolation) and forecasting (extrapolation) irregularly sampled time series data. However, their performance degrades substantially when applied to real-world data, especially long-term data with complex behaviors (e.g., long-term trend across years, mid-term seasonality across months, and short-term local variation across days). To address the modeling of such complex data with different behaviors at different frequencies (time spans), we propose a novel progressive learning paradigm of NODEs for long-term time series forecasting. Specifically, following the principle of curriculum learning, we gradually increase the complexity of data and network capacity as training progresses. Our experiments with both synthetic data and real traffic data (PeMS Bay Area traffic data) show that our training methodology consistently improves the performance of vanilla NODEs by over 64%.
1 Reply

Loading