Keywords: Federated Learning, Out-of-Distribution Generalization, Causal Inference, Invariant Learning
TL;DR: FL Games - a game-theoretic framework for federated learning that learns causal features that are invariant across clients
Abstract: Federated learning aims to train predictive models for data that is distributed across clients, under the orchestration of a server.
However, participating clients typically each hold data from a different distribution, which can yield to catastrophic generalization on data from a different client, which represents a new domain. In this work, we argue that in order to generalize better across non-i.i.d. clients, it is imperative only to learn correlations that are stable and invariant across domains. We propose FL Games, a game-theoretic framework for federated learning that learns causal features that are invariant across clients. While training to achieve the Nash equilibrium, the traditional best response strategy suffers from high-frequency oscillations. We demonstrate that FL Games effectively resolves this challenge and exhibits smooth performance curves. Further, FL Games scales well in the number of clients, requires significantly fewer communication rounds, and is agnostic to device heterogeneity. Through empirical evaluation, we demonstrate that \flgames achieves high out-of-distribution performance on various benchmarks.
Is Student: Yes
Community Implementations: [![CatalyzeX](/images/catalyzex_icon.svg) 1 code implementation](https://www.catalyzex.com/paper/fl-games-a-federated-learning-framework-for/code)
5 Replies
Loading