Flexible Multitask Learning with Factorized Diffusion Policy

Published: 12 Jun 2025, Last Modified: 20 Jun 2025RobotEvaluation@RSS 2025 PosterEveryoneRevisionsBibTeXCC BY 4.0
Type: An approach-centric paper (introducing new robot systems and approaches with a strong emphasis on real-world applicability and evaluation)
Keywords: Multitask Imitation Learning, Compositional Diffusion, Robotic Manipulation
TL;DR: We propose a factorized diffusion policy that composes specialized diffusion models for robot control, enabling strong multitask performance and efficient adaptation to new tasks.
Abstract: In recent years, large-scale behavioral cloning has emerged as a promising paradigm for training general-purpose robot policies. However, effectively fitting policies to complex task distributions is often challenging, and existing models often underfit the action distribution. In this paper, we present a novel modular diffusion policy framework that factorizes modeling the complex action distributions as a composition of specialized diffusion models, each capturing a distinct sub-mode of the multimodal behavior space. This factorization enables each composed model to specialize and capture a subset of the task distribution, allowing the overall task distribution to be more effectively represented. In addition, this modular structure enables flexible policy adaptation to new tasks by simply fine-tuning a subset of components or adding new ones for novel tasks, while inherently mitigating catastrophic forgetting. Empirically, across both simulation and real-world robotic manipulation settings, we illustrate how our method consistently outperforms strong modular and monolithic baselines, achieving a 24% average relative improvement in multitask learning and a 34% improvement in task adaptation across all settings.
Submission Number: 13
Loading