Scalable Second-Order Optimization Algorithms for Minimizing Low-rank Functions

Published: 10 Oct 2024, Last Modified: 07 Dec 2024NeurIPS 2024 WorkshopEveryoneRevisionsBibTeXCC BY 4.0
Keywords: random subspace methods, cubic regularization, global rate of convergence, low rank functions
TL;DR: We present a random-subspace variant of cubic regularization algorithm that chooses the size of the subspace adaptively, and is particularly efficient for low-rank functions.
Abstract: We present a random-subspace variant of cubic regularization algorithm that chooses the size of the subspace adaptively, based on the rank of the projected second derivative matrix. Iteratively, our variant only requires access to (small-dimensional) projections of first- and second-order problem derivatives and calculates a reduced step inexpensively. The ensuing method maintains the optimal global rate of convergence of (full-dimensional) cubic regularization, while showing improved scalability both theoretically and numerically, particularly when applied to low-rank functions. When applied to the latter, our algorithm naturally adapts the subspace size to the true rank of the function, without knowing it a priori.
Submission Number: 64
Loading