Neural Operator: Graph Kernel Network for Partial Differential EquationsDownload PDF

Published: 27 Feb 2020, Last Modified: 22 Oct 2023ICLR 2020 Workshop ODE/PDE+DL PosterReaders: Everyone
Keywords: Graph Neural Networks, Partial Differential Equations, Kernel Methods, Infinite Space Mappings
TL;DR: Graph kernel networks use message passing to do kernel integration, so that it can learn an operator from the input functions to the solution functions, among different resolution and discretization.
Abstract: The classical development of neural networks has been primarily for mappings between a finite-dimensional Euclidean space and a set of classes, or between two finite-dimensional Euclidean spaces. The purpose of this work is to generalize neural networks so that they can learn mappings between infinite-dimensional spaces (operators). We formulate approximation of the infinite-dimensional mapping by composing nonlinear activation functions and a class of integral operators. The kernel integration is computed by message passing on graph networks. This approach has substantial practical consequences which we will illustrate in the context of mappings between input data to partial differential equations (PDEs) and their solutions. In this context, such learned networks can generalize among different approximation methods for the PDE (such as finite difference or finite element methods) and among approximations corresponding to different underlying levels of resolution and discretization. Experiments confirm the purposed graph kernel network does have the desired properties and show competitive performance compared to the stat of the art solvers.
Community Implementations: [![CatalyzeX](/images/catalyzex_icon.svg) 3 code implementations](https://www.catalyzex.com/paper/arxiv:2003.03485/code)
1 Reply

Loading