Abstract: Batch Normalisation (BN) is widely used in conventional deep neural network training to harmonise the input-output distributions for each batch of data.
However, federated learning, a distributed learning paradigm, faces the challenge of dealing with non-independent and identically distributed data among the client nodes.
Due to the lack of a coherent methodology for updating BN statistical parameters, standard BN degrades the federated learning performance.
To this end, it is urgent to explore an alternative normalisation solution for federated learning.
In this work, we resolve the dilemma of the BN layer in federated learning by developing a customised normalisation approach, Hybrid Batch Normalisation (HBN).
HBN separates the update of statistical parameters (*i.e.*, means and variances used for evaluation) from that of learnable parameters (*i.e.*, parameters that require gradient updates), obtaining unbiased estimates of global statistical parameters in distributed scenarios.
In contrast with the existing solutions, we emphasise the supportive power of global statistics for federated learning.
The HBN layer introduces a learnable hybrid distribution factor, allowing each computing node to adaptively mix the statistical parameters of the current batch with the global statistics.
Our HBN can serve as a powerful plugin to advance federated learning performance.
It reflects promising merits across a wide range of federated learning settings, especially for small batch sizes and heterogeneous data.
Code is available at https://github.com/Hongyao-Chen/HybridBN.
Lay Summary: We propose a hybrid batch normalisation approach in federated learning. We achieve more effective normalisation in data heterogeneous distribution federated learning by adaptively mixing historical unbiased global statistics with local batch statistics. Our hybrid batch normalistion layer can serve as a powerful plugin to advance federated learning performance.
Link To Code: https://github.com/Hongyao-Chen/HybridBN
Primary Area: Optimization->Large Scale, Parallel and Distributed
Keywords: Federated Learning, Batch Normalisation, Distributed Learning, Data Heterogeneity
Submission Number: 9218
Loading