Improving Linear System Solvers for Hyperparameter Optimisation in Iterative Gaussian Processes

Published: 25 Sept 2024, Last Modified: 06 Nov 2024NeurIPS 2024 posterEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Gaussian process, marginal likelihood, iterative linear system solver, pathwise, warm start, conjugate gradient, alternating projection, stochastic gradient descent
TL;DR: We accelerate iterative linear system solvers in the context of marginal likelihood optimisation for Gaussian processes.
Abstract: Scaling hyperparameter optimisation to very large datasets remains an open problem in the Gaussian process community. This paper focuses on iterative methods, which use linear system solvers, like conjugate gradients, alternating projections or stochastic gradient descent, to construct an estimate of the marginal likelihood gradient. We discuss three key improvements which are applicable across solvers: (i) a pathwise gradient estimator, which reduces the required number of solver iterations and amortises the computational cost of making predictions, (ii) warm starting linear system solvers with the solution from the previous step, which leads to faster solver convergence at the cost of negligible bias, (iii) early stopping linear system solvers after a limited computational budget, which synergises with warm starting, allowing solver progress to accumulate over multiple marginal likelihood steps. These techniques provide speed-ups of up to $72\times$ when solving to tolerance, and decrease the average residual norm by up to $7\times$ when stopping early.
Primary Area: Probabilistic methods (for example: variational inference, Gaussian processes)
Submission Number: 19482
Loading