GENFLOWRL: Generative Object-Centric Flow Matching for Reward Shaping in Visual Reinforcement Learning

Published: 18 Apr 2025, Last Modified: 07 May 2025ICRA 2025 FMNS SpotlightEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Object-centric Representation, Reinforcement Learning, Diffusion Model
TL;DR: The framework "GenFlowRL" utilizes the foundation generative model “Stable Diffusion” for generating object-centric flow and uses delta-flow matching to shape dense reward in reinforcement learning.
Abstract: Recent advances have demonstrated the potential of video generation models to guide robot learning by deriving effective robot actions through inverse dynamics. However, these methods heavily depend on the quality of generated data and struggle with fine-grained manipulation due to the lack of environment feedback. While video-based reinforcement learning improves policy robustness, it remains constrained by the artifacts of generated video and the challenges of collecting large-scale in-domain robot datasets for training diffusion models. Motivated by the above, we propose GenFlowRL, which derives shaped rewards from generated flow trained from cross-embodiment datasets. This enables learning generalizable and robust policies from expert demostrations using low-dimensional, object-centric features. Experiments on 10 manipulation tasks, both in simulation and real-world cross-embodiment evaluations, demonstrate that GenFlowRL effectively leverages manipulation features extracted from generated object-centric flow, consistently achieving superior performance across diverse and challenging scenarios.
Submission Number: 24
Loading

OpenReview is a long-term project to advance science through improved peer review with legal nonprofit status. We gratefully acknowledge the support of the OpenReview Sponsors. © 2025 OpenReview