Vejde: A Framework for Inductive Deep Reinforcement Learning Based on Factor Graph Color Refinement

TMLR Paper5416 Authors

18 Jul 2025 (modified: 19 Dec 2025)Decision pending for TMLREveryoneRevisionsBibTeXCC BY 4.0
Abstract: We present and evaluate Vejde; a framework which combines data abstraction, graph learning and reinforcement learning to produce inductive policy functions for decision problems with richly structured states, such as object classes and relations. Markov decision process states are represented as data bases of facts about entities, and Vejde converts each state to a bipartite graph, which is mapped to latent states through neural message passing. The factored representation of both states and actions allows Vejde agents to handle problems of varying size and structure. We tested Vejde agents on eight problem domains defined in RDDL, with ten problem instances each, where policies were trained using both supervised and reinforcement learning. To test policy generalization, we separate problem instances in two sets, one for training and the other solely for testing. Test results on unseen instances for the Vejde agents were compared to MLP agents trained on each problem instance, as well as the online planning algorithm Prost. Our results show that Vejde policies in average generalize to the test instances without a significant loss in score. Additionally, the inductive agents received scores on unseen test instances that on average were close to the instance-specific MLP agents.
Submission Length: Long submission (more than 12 pages of main content)
Changes Since Last Submission: - Added an description and evaluation of a graph attention agent to the Appendix - Shortened the main contents of the paper by moving certain sections to the Appendix - Shortening various sentences for clarity and space. - Move certain formulas inline. - Added some additional notes on Navigation to the Appendix
Assigned Action Editor: ~Dennis_J._N._J._Soemers1
Submission Number: 5416
Loading