Channel-Prioritized Convolutional Neural Networks for Sparsity and Multi-fidelityDownload PDF

12 Feb 2018 (modified: 05 May 2023)ICLR 2018 Workshop SubmissionReaders: Everyone
Abstract: We propose a novel convolutional neural networks (CNNs) training procedure to allow dynamically trade-offs between different resource and performance requirements. Our approach prioritizes the channels to enable structured sparsity and multi-fidelity approximations at inference time. We train the VGG network with our method on various benchmark datasets. The experiment results show that, on the CIFAR-10 dataset, a 63x parameters reduction and a 11x FLOPs reduction can be achieved, with only a 2% accuracy drop.
Keywords: filter/channel prioritization, network pruning, compression, computational efficiency, multi-fidelity inference
4 Replies

Loading