Revisiting CHAMPAGNE: Sparse Bayesian Learning as Reweighted Sparse Coding

Published: 25 Mar 2025, Last Modified: 20 May 2025SampTA 2025 PosterEveryoneRevisionsBibTeXCC BY-SA 4.0
Session: General
Keywords: Inverse problem, Sparse Bayesian Learning, reweighted (group)-lasso
Abstract: This paper revisits the CHAMPAGNE algorithm within the Sparse Bayesian Learning (SBL) framework and establishes its connection to reweighted sparse coding. We demonstrate that the SBL objective can be reformulated as a reweighted $\ell_{21}$-minimization problem, providing a more straightforward interpretation of the sparsity mechanism and enabling the design of an efficient iterative algorithm. Additionally, we analyze the behavior of this reformulation in the low signal-to-noise ratio (SNR) regime, showing that it simplifies to a weighted $\ell_{21}$-regularized least squares problem. Numerical experiments validate the proposed approach, highlighting its improved computational efficiency and ability to produce exact sparse solutions, particularly in simulated MEG source localization tasks.
Submission Number: 26
Loading