Keywords: Non-negative weighted regret, Nash equilibrium, Harmonic games, Zero-sum games
TL;DR: Optimistic variants of classical no-regret learning algorithms converge to an \epsilon-approximate Nash equilibrium in games with conflicting interests.
Abstract: In this work, we introduce the concept of non-negative weighted regret, an extension of non-negative regret [APFS22] in games. Investigating games with non-negative weighted regret helps us to understand games with conflicting interests, including harmonic games and important classes of zero-sum games. We show that optimistic variants of classical no-regret learning algorithms, namely optimistic mirror descent (OMD) and optimistic follow the regularized leader (OFTRL), converge to an $\epsilon$-approximate Nash equilibrium at a rate of $O(1/\epsilon^2)$. Consequently, they guarantee pointwise convergence to a Nash equilibrium if there are only finitely many Nash equilibria in the game. These algorithms are robust in the sense the convergence holds even if the players deviate from prescribed strategies, as long as the corruption level remains finite. Our theoretical findings are supported by empirical evaluations of OMD and OFTRL on the game of matching pennies and harmonic game instances.
Supplementary Material: pdf
Primary Area: learning theory
Submission Number: 15442
Loading