FedMDP: A Federated Learning Framework to Handle System and Model Heterogeneity in Resource-Constrained Environments

Published: 31 Aug 2023, Last Modified: 30 Sept 20243rd Annual AAAI Workshop on AI to Accelerate Science and Engineering (AI2ASE)EveryoneCC0 1.0
Abstract: The advancement of technology, improvement of network infrastructures, and wide availability of internet open up the door of new opportunities to perform on-device inference. Realizing the potential of such advancement, Federated Learning (FL) was invented that facilitates the formation of a powerful model without exposing user data. While successful, it does not consider the combinational case, where the selected FL agents independently craft their local model with heterogeneous architecture and perform computational tasks based on their available resources. In the original FL model, all agents need to agree on a uniform model architecture and are assigned a uniform computational task. However, in a real-life resource-constrained FL setting, agents may not be interested to share their local model architecture details due to privacy and security concerns. Also, the heterogeneous local model architectures cannot be aggregated together on the FL server following the traditional approaches. Moving forward, we may observe straggler agents due to resource-constrained environments, such that any FL agent may find a task as computationally challenging that can prolong the model convergence. To address the above-mentioned challenges regarding agent’s local model and resource heterogeneity, we propose an FL framework, FedMDP that can effectively handle federated agents possessing nonidentical local model structure as well as variant local resources using knowledge distillation and dynamic local task allocation techniques. We tested our framework on MNIST and CIFAR100 dataset and observed significant improvement in accuracy in a highly heterogeneous environment. By considering 10 uniquely designed model of the agents, we achieved 15% gain on average compared to the accuracy of the traditional learning methods and observed a few percent lower accuracy compared to the case if the agents’ local datasets were pooled and made available for all the network agents.
Loading