Large Language Models Can Design Game-Theoretic Objectives for Multi-Agent Planning

23 Sept 2023 (modified: 25 Mar 2024)ICLR 2024 Conference Withdrawn SubmissionEveryoneRevisionsBibTeX
Keywords: Multi-Agent Planning, Game-theoretic Objectives
TL;DR: This work proposes a multi-agent planning architecture that relies on a large language model (LLM) for the design of the underlying game's objectives.
Abstract: Game theory is a powerful paradigm to describe the interplay between participants in interactive multi-agent scenarios, and relies on the knowledge of player objectives or payoff structures for game optimal decision making. However, designing such objectives for games is challenging as it requires evaluating the impact of an agent’s actions on the behavior of others, and understanding the effect of changes in one's policy on the behavior of others. Indeed, aligning objective representations with a desired multi-agent behavior is achieved via tedious and impractical heuristics or human trial-and-error. This work aims to ease this process and proposes a multi-agent planning architecture that relies on a large language model (LLM) as the game formulation designer. First, we exhibit the zero-shot proficiency of the more capable LLMs (such as GPT-4) in tuning continuous objective function parameters in accordance with a specified high-level goal for autonomous driving examples. We then develop a planner which uses an LLM as a matrix game designer, for scenarios with discrete and finite action spaces. Given a scene history, the actions available to each agent, and high-level objectives (expressed in natural language), the LLM evaluates the payoffs associated with each combination of actions. From the game structure obtained, agents execute Nash optimal actions, the scene is re-evaluated, and the process is repeated. We evaluate our approach on a heterogeneous robot planning task inspired by wildlife conservation, as well as a household multi-humanoid transport task, and show the superiority of our LLM-based approach to other baselines
Primary Area: applications to robotics, autonomy, planning
Code Of Ethics: I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics.
Submission Guidelines: I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2024/AuthorGuide.
Anonymous Url: I certify that there is no URL (e.g., github page) that could be used to find authors' identity.
No Acknowledgement Section: I certify that there is no acknowledgement section in this submission for double blind review.
Submission Number: 8283
Loading