A Generalization Bound for Nearly-Linear Networks

TMLR Paper2981 Authors

09 Jul 2024 (modified: 25 Sept 2024)Under review for TMLREveryoneRevisionsBibTeXCC BY 4.0
Abstract: We consider nonlinear networks as perturbations of linear ones. Based on this approach, we present novel generalization bounds that become non-vacuous for networks that are close to being linear. The main advantage over the previous works which propose non-vacuous generalization bounds is that our bounds are a-priori: performing the actual training is not required for evaluating the bounds. To the best of our knowledge, they are the first non-vacuous generalization bounds for neural nets possessing this property.
Submission Length: Long submission (more than 12 pages of main content)
Assigned Action Editor: ~Jeffrey_Pennington1
Submission Number: 2981
Loading