Training Neural Networks with Fixed Sparse MasksDownload PDF

May 21, 2021 (edited Jan 21, 2022)NeurIPS 2021 PosterReaders: Everyone
  • Keywords: Fisher information, sparse updates, parameter-efficient transfer learning, distributed training, efficient checkpointing
  • TL;DR: We introduce a method for pre-computing a fixed mask that selects a subset of network parameters to update, alleviating storage and communication costs.
  • Abstract: During typical gradient-based training of deep neural networks, all of the model's parameters are updated at each iteration. Recent work has shown that it is possible to update only a small subset of the model's parameters during training, which can alleviate storage and communication requirements. In this paper, we show that it is possible to induce a fixed sparse mask on the model’s parameters that selects a subset to update over many iterations. Our method constructs the mask out of the $k$ parameters with the largest Fisher information as a simple approximation as to which parameters are most important for the task at hand. In experiments on parameter-efficient transfer learning and distributed training, we show that our approach matches or exceeds the performance of other methods for training with sparse updates while being more efficient in terms of memory usage and communication costs. We release our code publicly to promote further applications of our approach.
  • Supplementary Material: zip
  • Code Of Conduct: I certify that all co-authors of this work have read and commit to adhering to the NeurIPS Statement on Ethics, Fairness, Inclusivity, and Code of Conduct.
  • Code: https://github.com/varunnair18/FISH
12 Replies

Loading