Rare Gems: Finding Lottery Tickets at InitializationDownload PDF

Published: 31 Oct 2022, 18:00, Last Modified: 21 Jan 2023, 07:15NeurIPS 2022 AcceptReaders: Everyone
TL;DR: We come up with the first algorithm for pruning at initialization which passes all known sanity checks and competes with the performance of algorithms with warmup like IMP.
Abstract: Large neural networks can be pruned to a small fraction of their original size, with little loss in accuracy, by following a time-consuming "train, prune, re-train" approach. Frankle & Carbin conjecture that we can avoid this by training lottery tickets, i.e., special sparse subnetworks found at initialization, that can be trained to high accuracy. However, a subsequent line of work presents concrete evidence that current algorithms for finding trainable networks at initialization, fail simple baseline comparisons, e.g., against training random sparse subnetworks. Finding lottery tickets that train to better accuracy compared to simple baselines remains an open problem. In this work, we resolve this open problem by proposing Gem-Miner which finds lottery tickets at initialization that beat current baselines. Gem-Miner finds lottery tickets trainable to accuracy competitive or better than Iterative Magnitude Pruning (IMP), and does so up to $19\times$ faster.
Supplementary Material: pdf
15 Replies