Learning Adaptive Diffusion Policies for Hybrid Dynamical Systems

Published: 15 Jun 2025, Last Modified: 07 Aug 2025AIA 2025EveryoneRevisionsBibTeXCC BY 4.0
Keywords: Hybrid Systems, Reinforcement Learning, Mixture-of-Experts, Diffusion, Contrastive Learning
TL;DR: We propose a novel model-free, energy-diffusion based online RL algorithm that learns policies which can adapt to unseen scenarios in hybrid dynamical system control.
Abstract: Hybrid dynamical systems result from the combination of a set of continuous-variable with a discrete-event system that can be influenced by the environment and encompasses systems like legged robots and aircrafts. Model-based control of these systems is challenging, particularly when the systems are complex (high-dimensional non-linear dynamics) and the switching conditions are not perfectly known. We propose a model-free actor-critic learning framework, MoE-Diff, that leverages energy-diffusion for multi-modal action generation with contrastive loss enabling scalable, compositional, and robust decision-making across complex tasks. We use MoE-Diff for control of hybrid systems in apriori unknown environments. Qualitative and ablation studies attribute this improvement to MoE-Diff's ability to learn interpolatable representations and distinct behavior modes across experts enabling compositional generalization and adaptation to unseen scenarios.
Paper Type: New Full Paper
Submission Number: 4
Loading