Persistent Homology Captures the Generalization of Neural Networks Without A Validation SetDownload PDF

21 May 2021 (modified: 20 Oct 2024)NeurIPS 2021 SubmittedReaders: Everyone
Keywords: Neural Networks, Topological Data Analysis, learning, evolution, Persistent Homology
Abstract: The training of neural networks is usually monitored with a validation (holdout) set to estimate the generalization of the model. This is done instead of measuring intrinsic properties of the model to determine whether it is learning appropriately. In this work, we suggest studying the training of neural networks with Algebraic Topology, specifically Persistent Homology (PH). Using simplicial complex representations of neural networks, we study the PH diagram distance evolution on the neural network learning process with different architectures and several datasets. Results show that the PH diagram distance between consecutive neural network states correlates with the validation accuracy, implying that the generalization error of a neural network could be intrinsically estimated without any holdout set.
Code Of Conduct: I certify that all co-authors of this work have read and commit to adhering to the NeurIPS Statement on Ethics, Fairness, Inclusivity, and Code of Conduct.
TL;DR: We provide a method to monitor the generalization of Neural Networks without validation data
Supplementary Material: zip
Community Implementations: [![CatalyzeX](/images/catalyzex_icon.svg) 1 code implementation](https://www.catalyzex.com/paper/persistent-homology-captures-the/code)
17 Replies

Loading