Lightweighted Sparse Autoencoder based on Explainable Contribution

Published: 11 Jul 2023, Last Modified: 16 Jul 2023NCW ICML 2023EveryoneRevisionsBibTeX
Keywords: Autoencoder, Sparse Autoencoder, Shapley value, Feature importance, Lightweight, Pruning, Link importance, Unit importance
Abstract: As deep learning models become heavier, developing lightweight models with the least performance degradation is paramount. In this paper, we propose an algorithm, SHAP-SAE (SHapley Additive exPlanations based Sparse AutoEncoder), that can explicitly measure the contribution of units and links and selectively activate only important units and links, leading to a lightweight sparse autoencoder. This allows us to explain how and why the sparse autoencoder is structured. We show that the SHAP-SAE outperforms other algorithms including a dense autoencoder. It is also confirmed that the SHAP-SAE is robust against the harsh sparsity of the autoencoder, as it shows remarkably limited performance degradation even with high sparsity levels.
Submission Number: 8
Loading