Resolving Multiple-Dynamic Model Uncertainty in Hypothesis-Driven Belief-MDPs

Published: 2025, Last Modified: 13 Feb 2026AAMAS 2025EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: This work explores the set of planning problems we define as hypothesis-driven planning. In these problems, a human with potentially no access to the system's underlying planning model (e.g., the reward or transition functions) is interested in answering a set of questions (hypotheses) that are not reflected in the underlying planning problem. To reason about the possible hypotheses and autonomously decide which one is most likely, we develop a new multiple-dynamic hypothesis belief Markov decision process (MDH-BMDP). This enables adding multiple hypotheses for existing planning problems and reasoning over arbitrary belief shapes using existing sparse tree search solvers. Lastly, we suggest a reward function that supports balancing the objectives of determining the correct hypothesis in time and performing well in the underlying problem; and test the framework's applicability and performance of the new reward function in simulations.
Loading