A Performance Analysis for Confidential Federated Learning

Published: 01 Jan 2024, Last Modified: 06 Feb 2025SP (Workshops) 2024EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: Federated Learning (FL) has emerged as a solution to preserve data privacy by keeping the data locally on each participant’s device. However, FL alone is still vulnerable to attacks that can cause privacy leaks. Therefore, additional security measures, at the cost of increasing runtimes, become necessary. The Trusted Execution Environment (TEE) approach offers the highest degree of security during execution. However, TEEs suffer from memory limits which prevent safe end-to-end FL training of modern deep models. Stateof-the-art approaches limit secure training to selected layers, failing to avert the full spectrum of attacks or adopt layerwise training affecting model performance. We benchmark the usage of a library OS (LibOS) to run the full, unmodified endto-end FL training inside the TEE. We extensively evaluate and model the overhead of the different security mechanisms needed to protect the data and model during computation (TEE), communication (TLS), and storage (disk encryption). The obtained results across three datasets and two models demonstrate that LibOSes are a viable way to seamlessly inject security into FL with limited overhead (at most $2 x$), offering valuable guidance for researchers and developers aiming to apply FL in data-security-focused contexts.
Loading