Self-Efficacy Update in Reinforcement Learning: Impact on Goal Selection for Q-learning Agents

Published: 09 Oct 2024, Last Modified: 02 Dec 2024NeurIPS 2024 Workshop IMOL asTinyPaperPosterEveryoneRevisionsBibTeXCC BY 4.0
Track: Tiny paper track
Keywords: Interplay between intrinsic and extrinsic motivations, Goal selection, Self-efficacy
TL;DR: We model self-efficacy update in Q-learning agents and examine its impact on goal selection behavior.
Abstract: We introduce a dynamic self-efficacy learning rule and examine its impact on multi-goal selection in a grid-world. We model the Q-learning agent's self-efficacy as the integral of reward prediction errors (RPEs), allowing it to modulate the agent's expectation of achieving the best possible future outcome. Initial simulation results suggest that faster self-efficacy updates lead to higher overall reward accumulation, but with increased variability in reaching the optimal goal. These findings indicate that an optimal self-efficacy update rate, which can be learned through experience, may strike a balance between maximizing performance and maintaining stability.
Submission Number: 18
Loading