Object Empowerment-Driven Tool Selection for Exploration in Reinforcement Learning

Published: 12 Jun 2025, Last Modified: 03 Aug 2025CoLLAs 2025 - Workshop TrackEveryoneRevisionsBibTeXCC BY 4.0
Keywords: intrinsic motivation, reinforcement learning, empowerment
TL;DR: This paper integrates object-tool interactions into empowerment based intrinsic motivation for improving exploration in reinforcement learning
Abstract: Tool use enables humans to solve complex physical tasks beyond their immediate capabilities. However, discovering tool use remains a major challenge for reinforcement learning (RL) agents, as it requires mastering long-horizon behaviors with sparse, delayed feedback—resulting in poor exploration and sample efficiency. While classic intrinsic motivation (IM) improves exploration, its lack of bias toward object-tool interactions leads agents to explore irrelevant states, resulting in many costly real-world interactions. In this paper, we investigate how RL agents can efficiently learn to use tools by optimizing object-centric intrinsic motivations — specifically, object empowerment, which quantifies the agent’s potential influence over specific objects in the environment. We extend this intrinsic motivation to multi-tool, multi-object environments that better reflect real-world lifelong learning challenges. Our method enables agents to identify meaningful tool-object relationships, learn when and how to use tools, and understand their lasting effects. Experiments in grid-based Minihack environments demonstrate that agents guided by object empowerment explore more effectively, generalize to new object configurations, and outperform PPO under sparse reward conditions.
Submission Number: 10
Loading