Keywords: Learning in games, Saddle-point optimization
TL;DR: This paper proposes an asymmetric payoff perturbation technique that achieves last-iterate convergence without requiring parameter adjustments.
Abstract: This paper proposes an asymmetric perturbation technique for solving bilinear saddle-point optimization problems, commonly arising in minimax problems, game theory, and constrained optimization.
Perturbing payoffs or values is known to be effective in stabilizing learning dynamics and equilibrium computation.
However, it requires careful adjustment of the perturbation magnitude; otherwise, learning dynamics converge to only an approximate equilibrium.
To overcome this, we introduce an asymmetric perturbation approach, where only one player's payoff function is perturbed.
Exploiting the near-linear structure of bilinear problems, we show that, for a sufficiently small perturbation, the equilibrium strategy of the asymmetrically perturbed game coincides with an equilibrium strategy of the original game.
This property yields a perturbation-based learning algorithm that achieves convergence to an equilibrium strategy in the original game without requiring parameter adjustments.
Furthermore, we empirically demonstrate fast convergence toward equilibria in both normal-form and extensive-form games.
Supplementary Material: zip
Primary Area: optimization
Submission Number: 15327
Loading