Why do Larger Models Generalize Better? A Theoretical Perspective via the XOR ProblemDownload PDFOpen Website

2019 (modified: 11 Nov 2022)ICML 2019Readers: Everyone
Abstract: Empirical evidence suggests that neural networks with ReLU activations generalize better with over-parameterization. However, there is currently no theoretical analysis that explains this observati...
0 Replies

Loading