DyCodeExplainer: Explainable Dynamic Graph Attention for Multi-Agent Reinforcement Learning in Collaborative Coding

ICLR 2026 Conference Submission25517 Authors

20 Sept 2025 (modified: 08 Oct 2025)ICLR 2026 Conference SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Collaborative Coding
Abstract: We propose \textbf{DyCodeExplainer}, a novel multi-agent reinforcement learning (MARL) framework that integrates dynamic graph attention with explainability techniques to improve collaborative coding. Existing MARL systems typically depend on static communication protocols which are not flexible and transparent in performing coding tasks that are more complicated. The above method suffers from this limitation by treating the interaction of agents in the form of a time-evolving graph in which the nodes represent coding agents, and edges indicate messages exchanged between them. A dynamic graph attention network (DGAT) dynamically prioritizes the messages considering contextually relevant message, whereas hard attention gate eliminates noises and helps improve decision-making efficiency. Furthermore, the framework includes gradient-based attention attribution and rule-based post-hoc explanations to explain message prioritization for providing interpretable budgetary information about the collaborative process. The policy and critic networks use Transformer-XL and graph neural networks respectively for managing the long-range dependencies and assessing the memory argument of the joint state values. Experiments show DyCodeExplainer to be more accurate in terms of code correctness and collaborative efficiency than traditional MARL baselines. The novelty of the system is the simultaneous optimization of thresholds for dynamic attention and explainability rules to bridge an important gap in transparent multi-agent coding systems. This work will move the field forward by providing a scalable and interpretable solution for collaborative software development.
Primary Area: transfer learning, meta learning, and lifelong learning
Submission Number: 25517
Loading