On Linear Convergence in Smooth Convex-Concave Bilinearly-Coupled Saddle-Point Optimization: Lower Bounds and Optimal Algorithms
Abstract: We revisit the smooth convex-concave bilinearly-coupled saddle-point problem of the form $\min_x\max_y f(x) + \langle y,\mathbf{B} x\rangle - g(y)$. In the highly specific case where function $f(x)$ is strongly convex and function $g(y)$ is affine, or both functions are affine, there exist lower bounds on the number of gradient evaluations and matrix-vector multiplications required to solve the problem, as well as matching optimal algorithms. A notable aspect of these algorithms is that they are able to attain linear convergence, i.e., the number of iterations required to solve the problem is proportional to $\log(1/\epsilon)$. However, the class of bilinearly-coupled saddle-point problems for which linear convergence is possible is much wider and can involve general smooth non-strongly convex functions $f(x)$ and $g(y)$. Therefore, *we develop the first lower complexity bounds and matching optimal linearly converging algorithms for this problem class*. Our lower complexity bounds are much more general, but they cover and unify the existing results in the literature. On the other hand, our algorithm implements the separation of complexities, which, for the first time, enables the simultaneous achievement of both optimal gradient evaluation and matrix-vector multiplication complexities, resulting in the best theoretical performance to date.
Lay Summary: In this paper, we revisit iterative first-order, i.e., gradient algorithms for solving a class of minimax optimization problems of the form $\min_x \max_y f(x) + \langle y, \mathbf{B} x\rangle - g(y)$. We investigate the following question: "How many computations of the gradients $\nabla f(x)$, $\nabla g(y)$, and matrix-vector multiplications with the matrix $\mathbf{B}$ are necessary to find an approximate solution to the problem up to a given precision?" We provide an exhaustive answer to this question: (1) we establish mathematical lower bounds on these numbers; (2) we develop a state-of-the-art algorithm for solving the problem, which, for the first time in the literature, can simultaneously match these lower bounds.
Primary Area: Optimization->Convex
Keywords: convex optimization, saddle-point problems, gradient methods, minimax optimization
Submission Number: 12624
Loading