Weightless: Lossy Weight Encoding For Deep Neural Network CompressionDownload PDF

15 Feb 2018 (modified: 16 Jun 2024)ICLR 2018 Conference Blind SubmissionReaders: Everyone
Abstract: The large memory requirements of deep neural networks strain the capabilities of many devices, limiting their deployment and adoption. Model compression methods effectively reduce the memory requirements of these models, usually through applying transformations such as weight pruning or quantization. In this paper, we present a novel scheme for lossy weight encoding which complements conventional compression techniques. The encoding is based on the Bloomier filter, a probabilistic data structure that can save space at the cost of introducing random errors. Leveraging the ability of neural networks to tolerate these imperfections and by re-training around the errors, the proposed technique, Weightless, can compress DNN weights by up to 496x; with the same model accuracy, this results in up to a 1.51x improvement over the state-of-the-art.
TL;DR: We propose a new way to compress neural networks using probabilistic data structures.
Keywords: Deep Neural Network, Compression, Sparsity
Community Implementations: [![CatalyzeX](/images/catalyzex_icon.svg) 2 code implementations](https://www.catalyzex.com/paper/arxiv:1711.04686/code)
7 Replies