MENTOR: Mixture-of-Experts Network with Task-Oriented Perturbation for Visual Reinforcement Learning
TL;DR: MENTOR is a highly efficient visual RL algorithm that excels in both simulation and real-world complex robotic learning tasks.
Abstract: Visual deep reinforcement learning (RL) enables robots to acquire skills from visual input for unstructured tasks. However, current algorithms suffer from low sample efficiency, limiting their practical applicability. In this work, we present MENTOR, a method that improves both the *architecture* and *optimization* of RL agents. Specifically, MENTOR replaces the standard multi-layer perceptron (MLP) with a mixture-of-experts (MoE) backbone and introduces a task-oriented perturbation mechanism. MENTOR outperforms state-of-the-art methods across three simulation benchmarks and achieves an average of 83\% success rate on three challenging real-world robotic manipulation tasks, significantly surpassing the 32% success rate of the strongest existing model-free visual RL algorithm. These results underscore the importance of sample efficiency in advancing visual RL for real-world robotics. Experimental videos are available at https://suninghuang19.github.io/mentor_page/.
Lay Summary: (1) Teaching robots skills from raw camera images is painfully slow and data-hungry. (2) We introduce MENTOR, an AI mentor that assigns specialist mini-brains to each situation and nudges them using lessons from past successes. (3) Robots can learn complex manipulation in real world, taking a step toward adaptable helpers in workplaces and everyday homes.
Link To Code: https://github.com/suninghuang19/mentor
Primary Area: Reinforcement Learning
Keywords: Visual Reinforcement Learning, Robotics, Mixture-of-Experts
Submission Number: 8464
Loading