Consistency Matters: Neural ODE Parameters are Dependent on the Training Numerical Method

Published: 03 Mar 2024, Last Modified: 30 Apr 2024AI4DiffEqtnsInSci @ ICLR 2024 PosterEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Differential Equations; Numerical Methods; Neural ODEs; Optimization
TL;DR: We provide theoretical insights into how numerical methods of varying orders or with different step sizes influence the loss function of Neural ODEs, demonstrating that the choice of numerical method inherently impacts the parameters of the network.
Abstract: Neural Ordinary Differential Equations (Neural ODEs) are continuous-depth models that use an ordinary differential equation (ODE) to capture the dynamics of data. Due to their modelling capabilities several works on applications and novel architectures using Neural ODEs can be found in the literature. In this work, we call for the attention to the need of using the same numerical method for both training and making predictions with Neural ODEs since the numerical method employed influences the prediction process, thereby impacting the loss function and introducing variance into parameter optimisation. We provide theoretical insights into how numerical methods of varying orders or with different step sizes influence the loss function of the network. To validate our theoretical analysis, we conduct a series of simple preliminary numerical experiments employing a regression task, demonstrating how the training numerical method influences model performance for testing. Our findings underscore the need for consistency in numerical methods for training and prediction, a consideration not previously emphasised or documented in the literature.
Submission Number: 70
Loading