Towards Interpretable Visual Decoding with Attention to Brain Representations

Published: 26 Jan 2026, Last Modified: 11 Apr 2026ICLR 2026 PosterEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Brain Decoding, Visual Reconstruction, Interpretability, fMRI, Latent Diffusion Model
TL;DR: We introduce NeuroAdapter, an end-to-end fMRI-to-image decoding method that directly conditions diffusion models on fMRI signals, achieving competitive reconstructions while enabling interpretable analysis of how brain regions guide generation.
Abstract: Recent work has demonstrated that complex visual stimuli can be decoded from human brain activity using deep generative models, offering new ways to probe how the brain represents real-world scenes. However, many existing approaches first map brain signals into intermediate image or text feature spaces before guiding the generative process, which obscures the contributions of different brain areas to the final reconstruction output. In this work, we propose NeuroAdapter, a visual decoding framework that directly conditions a latent diffusion model on brain representations, bypassing the need for intermediate feature spaces. Our method demonstrates competitive visual reconstruction quality on public fMRI datasets compared to prior work, while providing greater transparency into how brain signals drive visual reconstruction. To this end, we introduce an Image–Brain BI-directional interpretability framework (IBBI) that analyzes cross-attention patterns across diffusion denoising steps to reveal how different cortical areas influence the unfolding generative trajectory. Our work highlights the potential of end-to-end brain-to-image reconstruction and establishes a path for interpretable neural decoding.
Primary Area: applications to neuroscience & cognitive science
Submission Number: 16079
Loading