Towards On-Device Personalization: Cloud-device Collaborative Data Augmentation for Efficient On-device Language Model
Abstract: With the advancement of large language models (LLMs), significant progress has been achieved in various Natural Language Processing (NLP) tasks.
However, existing LLMs still face two major challenges that hinder their broader adoption: (1) their responses tend to be generic and lack personalization tailored to individual users, and (2) they rely heavily on cloud infrastructure due to intensive computational requirements, leading to stable network dependency and response delay.
Recent research has predominantly focused on either developing cloud-based personalized LLMs or exploring the on-device deployment of general-purpose LLMs. However, few studies have addressed both limitations simultaneously by investigating personalized on-device language models.
To bridge this gap, we propose CDCDA-PLM, a framework for deploying personalized on-device language models on user devices with support from a powerful cloud-based LLM.
Specifically, CDCDA-PLM leverages the server-side LLM's strong generalization capabilities to augment users' limited personal data, mitigating the issue of data scarcity.
Using both real and synthetic data, A personalized on-device language models (LMs) is fine-tuned via parameter-efficient fine-tuning (PEFT) modules and deployed on users' local devices, enabling them to process queries without depending on cloud-based LLMs.
This approach eliminates reliance on network stability and ensures high response speeds.
Experimental results across six tasks in a widely used personalization benchmark demonstrate the effectiveness of CDCDA-PLM.
Paper Type: Long
Research Area: Information Retrieval and Text Mining
Research Area Keywords: Large Language Model, Personalization, On-device LLM
Contribution Types: Approaches to low-resource settings
Languages Studied: English
Submission Number: 2496
Loading