Model Compression via Symmetries of the Parameter SpaceDownload PDF

Published: 28 Jan 2022, Last Modified: 13 Feb 2023ICLR 2022 SubmittedReaders: Everyone
Keywords: symmetry, orthogonal group, quiver representation, representation theory, model compression, parameter optimization, projected gradient descent
Abstract: We provide a theoretical framework for neural networks in terms of the representation theory of quivers, thus revealing symmetries of the parameter space of neural networks. An exploitation of these symmetries leads to a model compression algorithm for radial neural networks based on an analogue of the QR decomposition. The algorithm is lossless; the compressed model has the same feedforward function as the original model. If applied before training, optimization of the compressed model by gradient descent is equivalent to a projected version of gradient descent on the original model.
One-sentence Summary: The representation theory of quivers describes symmetry in the structure of the neural networks which can be used for model compression.
Supplementary Material: zip
12 Replies

Loading