Lightweighted Sparse Autoencoder Based on Explainable Contribution
Abstract
As deep learning models become heavier, developing lightweight models with the least performance degradation is paramount. In this paper, we propose an algorithm, SHAP-SAE (SHapley Additive exPlanations based Sparse AutoEncoder), that can explicitly measure the contribution of units and links and selectively activate only important units and links, leading to a lightweight sparse autoencoder. This allows us to explain how and why the sparse autoencoder is structured. We show that the SHAP-SAE outperforms other algorithms including a dense autoencoder. It is also confirmed that the SHAP-SAE is robust against the harsh sparsity of the autoencoder, as it shows remarkably limited performance degradation even with high sparsity levels.
Cite
Text
Rheey and Park. "Lightweighted Sparse Autoencoder Based on Explainable Contribution." ICML 2023 Workshops: NCW, 2023.Markdown
[Rheey and Park. "Lightweighted Sparse Autoencoder Based on Explainable Contribution." ICML 2023 Workshops: NCW, 2023.](https://mlanthology.org/icmlw/2023/rheey2023icmlw-lightweighted/)BibTeX
@inproceedings{rheey2023icmlw-lightweighted,
title = {{Lightweighted Sparse Autoencoder Based on Explainable Contribution}},
author = {Rheey, Joohong and Park, Hyunggon},
booktitle = {ICML 2023 Workshops: NCW},
year = {2023},
url = {https://mlanthology.org/icmlw/2023/rheey2023icmlw-lightweighted/}
}