Status-quo policy gradient in Multi-Agent Reinforcement LearningDownload PDF

12 Oct 2021, 19:37 (modified: 23 Nov 2021, 06:32)Deep RL Workshop NeurIPS 2021Readers: Everyone
Keywords: multi-agent rl, reinforcement learning, social dilemma, policy gradient, game theory
TL;DR: We address the problem of evolving cooperation in sequential social dilemmas by introducing two new algorithms.
Abstract: Individual rationality, which involves maximizing expected individual return, does not always lead to optimal individual or group outcomes in multi-agent problems. For instance, in social dilemma situations, Reinforcement Learning (RL) agents trained to maximize individual rewards converge to mutual defection that is individually and socially sub-optimal. In contrast, humans evolve individual and socially optimal strategies in such social dilemmas. Inspired by ideas from human psychology that attribute this behavior in humans to the status-quo bias, we present a status-quo loss (SQLoss) and the corresponding policy gradient algorithm that incorporates this bias in an RL agent. We demonstrate that agents trained with SQLoss learn individually as well as socially optimal behavior in several social dilemma matrix games. To apply SQLoss to games where cooperation and defection are determined by a sequence of non-trivial actions, we present GameDistill, an algorithm that reduces a multi-step game with visual input to a matrix game. We empirically show how agents trained with SQLoss on GameDistill reduced version of Coin Game and StagHunt evolve optimal policies. Finally, we show that SQLoss extends to a 4-agent setting by demonstrating the emergence of cooperative behavior in the popular Braess' paradox.
Supplementary Material: zip
0 Replies