Well-tuned Simple Nets Excel on Tabular DatasetsDownload PDF

21 May 2021, 20:43 (modified: 24 Jan 2022, 17:49)NeurIPS 2021 PosterReaders: Everyone
Keywords: tabular datasets, neural networks, multilayer perceptrons, deep learning, regularization cocktails, combined algorithm selection and hyperparameter optimization.
TL;DR: We demonstrate that the right combination of modern regularization methods allows plain neural networks to outperform the best traditional ML techniques on tabular data.
Abstract: Tabular datasets are the last "unconquered castle" for deep learning, with traditional ML methods like Gradient-Boosted Decision Trees still performing strongly even against recent specialized neural architectures. In this paper, we hypothesize that the key to boosting the performance of neural networks lies in rethinking the joint and simultaneous application of a large set of modern regularization techniques. As a result, we propose regularizing plain Multilayer Perceptron (MLP) networks by searching for the optimal combination/cocktail of 13 regularization techniques for each dataset using a joint optimization over the decision on which regularizers to apply and their subsidiary hyperparameters. We empirically assess the impact of these regularization cocktails for MLPs in a large-scale empirical study comprising 40 tabular datasets and demonstrate that (i) well-regularized plain MLPs significantly outperform recent state-of-the-art specialized neural network architectures, and (ii) they even outperform strong traditional ML methods, such as XGBoost.
Supplementary Material: pdf
Code Of Conduct: I certify that all co-authors of this work have read and commit to adhering to the NeurIPS Statement on Ethics, Fairness, Inclusivity, and Code of Conduct.
Code: https://github.com/releaunifreiburg/WellTunedSimpleNets
13 Replies