Train your cake and eat it too! Repurposing collaborative training to tailor LLMs to private data without sharing

Published: 21 Jun 2024, Last Modified: 26 Jul 2024ES-FoMo-II 2024 PosterEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Split Learning, Collaborative Training, Distributed Training, LLM Training, Resource Efficiency
TL;DR: Evaluation of split learning and federated learning algorithms for LLM fine-tuning without revealing private data.
Abstract: In the emerging field of large language models (LLMs), a significant challenge arises when organizations with vast datasets lack the computational resources to independently train and fine-tune models. This issue stems from privacy, compliance, and resource constraints: organizations cannot share their sensitive data but still need external computational assistance for model training. In this paper, we implement, enhance, and empirically compare several methods, including Split Learning (SL) and select Federated Learning (FL) methods, which enable data-rich yet compute-poor clients to offload LLM training without sharing raw data. Our study evaluates these methods across multiple dimensions, including model quality and training time.
Submission Number: 42
Loading