TL;DR: We search subspace of regret minimization algorithms for those which converge close to a Nash equilibrium.
Abstract: Nash equilibrium is perhaps the best-known solution concept in game theory.
Such a solution assigns a strategy to each player which offers no incentive to unilaterally deviate.
While a Nash equilibrium is guaranteed to always exist, the problem of finding one in general-sum games is PPAD-complete, generally considered intractable.
Regret minimization is an efficient framework for approximating Nash equilibria in two-player zero-sum games.
However, in general-sum games, such algorithms are only guaranteed to converge to a coarse-correlated equilibrium (CCE), a solution concept where player can correlate their strategies.
In this work, we use meta-learning to minimize the correlations in strategies produced by a regret minimizer. This encourages the regret minimizer to find strategies that are closer to a Nash equilibrium.
The meta-learned regret minimizer is still guaranteed to converge to a CCE, but we give a bound on the distance to Nash equilibrium in terms of our meta-loss.
We evaluate our approach in general-sum imperfect information games.
Our algorithms provide significantly better approximations of Nash equilibria than state-of-the-art regret minimization techniques.
Primary Area: Theory->Game Theory
Keywords: online learning, Nash equilibrium, general-sum games
Submission Number: 4012
Loading