Convergence Analysis of Two-Layer Neural Networks under Gaussian Input Masking

15 Feb 2026 (modified: 21 Mar 2026)Under review for TMLREveryoneRevisionsBibTeXCC BY 4.0
Abstract: We investigate the convergence guarantee of two-layer neural network training with Gaussian randomly masked inputs. This scenario corresponds to Gaussian dropout at the input level, or noisy input training common in sensor networks, privacy-preserving training, and federated learning, where each user may have access to partial or corrupted features. Using a Neural Tangent Kernel (NTK) analysis, we demonstrate that training a two-layer ReLU network with Gaussian randomly masked inputs achieves linear convergence up to an error region proportional to the mask's variance. A key technical contribution is resolving the randomness within the non-linear activation, a problem of independent interest.
Submission Type: Long submission (more than 12 pages of main content)
Assigned Action Editor: ~Yiming_Ying1
Submission Number: 7520
Loading