Two-layer neural network on infinite dimensional data: global optimization guarantee in the mean-field regimeDownload PDF

Published: 31 Oct 2022, Last Modified: 06 Jan 2023NeurIPS 2022 AcceptReaders: Everyone
Keywords: mean-field regime, neural network, optimization, functional data analysis
TL;DR: We develop a new mean-field analysis of two-layer neural network in an infinite-dimensional parameter space and propose two optimization algorithms which converge to the (regularized) global optimal solution.
Abstract: Analysis of neural network optimization in the mean-field regime is important as the setting allows for feature learning. Existing theory has been developed mainly for neural networks in finite dimensions, i.e., each neuron has a finite-dimensional parameter. However, the setting of infinite-dimensional input naturally arises in machine learning problems such as nonparametric functional data analysis and graph classification. In this paper, we develop a new mean-field analysis of two-layer neural network in an infinite-dimensional parameter space. We first give a generalization error bound, which shows that the regularized empirical risk minimizer properly generalizes when the data size is sufficiently large, despite the neurons being infinite-dimensional. Next, we present two gradient-based optimization algorithms for infinite-dimensional mean-field networks, by extending the recently developed particle optimization framework to the infinite-dimensional setting. We show that the proposed algorithms converge to the (regularized) global optimal solution, and moreover, their rates of convergence are of polynomial order in the online setting and exponential order in the finite sample setting, respectively. To our knowledge this is the first quantitative global optimization guarantee of neural network on infinite-dimensional input and in the presence of feature learning.
Supplementary Material: zip
8 Replies

Loading