The Persistence of Neural Collapse Despite Low-Rank Bias

Published: 18 Sept 2025, Last Modified: 29 Oct 2025NeurIPS 2025 posterEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Neural collapse, low-rank bias, unconstrained feature models, loss landscape analysis, deep learning theory, optimization dynamics
TL;DR: We analytically show that deep neural collapse is suboptimal in deep cross-entropy unconstrained feature models and explain why it persists empirically despite this.
Abstract: Neural collapse (NC) and its multi-layer variant, deep neural collapse (DNC), describe a structured geometry that occurs in the features and weights of trained deep networks. Recent theoretical work by Sukenik et al. using a deep unconstrained feature model (UFM) suggests that DNC is suboptimal under mean squared error (MSE) loss. They heuristically argue that this is due to low-rank bias induced by L2 regularization. In this work, we extend this result to deep UFMs trained with cross-entropy loss, showing that high-rank structures—including DNC—are not generally optimal. We characterize the associated low-rank bias, proving a fixed bound on the number of non-negligible singular values at global minima as network depth increases. We further analyze the loss surface, demonstrating that DNC is more prevalent in the landscape than other critical configurations, which we argue explains its frequent empirical appearance. Our results are validated through experiments in deep UFMs and deep neural networks.
Primary Area: Theory (e.g., control theory, learning theory, algorithmic game theory)
Submission Number: 20866
Loading