Meta Compression: Learning to Compress Pre-trained Deep Neural Networks

TMLR Paper5919 Authors

18 Sept 2025 (modified: 06 Oct 2025)Under review for TMLREveryoneRevisionsBibTeXCC BY 4.0
Abstract: State-of-the-art deep neural networks (DNN) have achieved outstanding results in a variety of tasks. Unfortunately, these DNN so large that cannot fit into the limited resources of edge servers or end devices such as smartphones and IoT sensors. Several approaches have been proposed to design compact yet efficient DNNs, however, the performance of the compressed model can be only characterized a posteriori. This work addresses this issue by introducing meta compression, a novel approach based on meta learning to simplify a pre-trained DNN into one that fulfills given constraints on size or accuracy. We leverage diffusion-based generative models to improve generalization performance of meta learning and extensively evaluate meta compression on an image classification task with popular pre-trained DNNs. The obtained results show that meta compression achieves a 92% top-5 recommendation accuracy and that the top-1 recommendation is only 1% far from the optimal compression method in terms of average accuracy loss.
Submission Length: Regular submission (no more than 12 pages of main content)
Assigned Action Editor: ~Yani_Ioannou1
Submission Number: 5919
Loading