A Game Theoretic Approach to Meta-Learning: Nash Model-Agnostic Meta-Learning

22 Sept 2023 (modified: 25 Mar 2024)ICLR 2024 Conference Withdrawn SubmissionEveryoneRevisionsBibTeX
Keywords: Meta learning, Game Theory, Generalized Stackelberg Equilibrium
TL;DR: This paper proposes a new modeling framework for model-agnostic meta-learning and an algorithm that the convergence to the generalized Stackelberg equilibrium is proved.
Abstract: Meta-learning, or learning to learn, aims to develop algorithms that can quickly adapt to new tasks and environments. Model-agnostic meta-learning (MAML), proposed as a bi-level optimization problem, is widely used as a baseline for gradient-based meta-learning algorithms that learn meta-parameters. In MAML, task-specific parameters are adapted independently in the inner-loop. After learning the task-specific parameters, the meta-parameters are learned in the outer-loop by minimizing the average task loss. After MAML, some gradient-based meta-learning research has explored objectives beyond average task losses, such as minimizing worst-case task losses for risk management and improving zero-shot performance in unadaptable environments. However, if the purpose of learning meta-parameters changes, the inner-loop formulation must change accordingly. Therefore, we propose a novel gradient-based meta-learning framework that imposes joint strategy sets and utility functions among tasks, making each task affected by other tasks. To solve this complex problem, we first show the proposed framework can be formulated as a generalized Stackelberg game. After that, we propose the NashMAML algorithm to compute the generalized Stackelberg equilibrium of this model and theoretically prove its convergence. We validate our approach on sinusoidal regression and few-shot image classification tasks. The results demonstrate that our approach outperforms previous methods in handling few-shot learning problems.
Supplementary Material: zip
Primary Area: transfer learning, meta learning, and lifelong learning
Code Of Ethics: I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics.
Submission Guidelines: I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2024/AuthorGuide.
Anonymous Url: I certify that there is no URL (e.g., github page) that could be used to find authors' identity.
No Acknowledgement Section: I certify that there is no acknowledgement section in this submission for double blind review.
Submission Number: 5455
Loading