## Simple steps are all you need: Frank-Wolfe and generalized self-concordant functions

21 May 2021, 20:44 (modified: 21 Dec 2021, 18:49)NeurIPS 2021 PosterReaders: Everyone
Keywords: Convex Optimization, Machine Learning
TL;DR: A simple Frank-Wolfe variant for generalized self-concordant functions that achieves the same guarantees/performance as more involved state-of-the-art Frank-Wolfe algorithms.
Abstract: Generalized self-concordance is a key property present in the objective function of many important learning problems. We establish the convergence rate of a simple Frank-Wolfe variant that uses the open-loop step size strategy $\gamma_t = 2/(t+2)$, obtaining a $\mathcal{O}(1/t)$ convergence rate for this class of functions in terms of primal gap and Frank-Wolfe gap, where $t$ is the iteration count. This avoids the use of second-order information or the need to estimate local smoothness parameters of previous work. We also show improved convergence rates for various common cases, e.g., when the feasible region under consideration is uniformly convex or polyhedral.
Supplementary Material: pdf
Code Of Conduct: I certify that all co-authors of this work have read and commit to adhering to the NeurIPS Statement on Ethics, Fairness, Inclusivity, and Code of Conduct.
Code: https://github.com/ZIB-IOL/fw-generalized-selfconcordant
15 Replies