Towards Federated Learning with on-device Training and Communication in 8-bit Floating Point

Published: 26 Aug 2024, Last Modified: 26 Aug 2024FedKDD 2024 PosterEveryoneRevisionsBibTeXCC BY 4.0
Keywords: federated learning, quantization, FP8
TL;DR: We present a method for federated learning in FP8 format, provide convergence analysis and experiments for several models and datasets.
Abstract: Recent work has shown that 8-bit floating point (FP8) can be used for efficiently training neural networks with reduced computational overhead compared to training in FP32/FP16. In this work, we investigate the use of FP8 training in a federated learning context. This brings not only the usual benefits of FP8 which are desirable for on-device training at the edge, but also reduces client-server communication costs due to significant weight compression. We present a novel method for combining FP8 client training while maintaining a global FP32 server model and provide convergence analysis. Experiments with various machine learning models and datasets show that our method consistently yields communication reductions of at least 2.9x across a variety of tasks and models compared to an FP32 baseline.
Submission Number: 6
Loading

OpenReview is a long-term project to advance science through improved peer review with legal nonprofit status. We gratefully acknowledge the support of the OpenReview Sponsors. © 2025 OpenReview