Interpreting and Disentangling Feature Components of Various Complexity from DNNs
Abstract
This paper aims to define, visualize, and analyze the feature complexity that is learned by a DNN. We propose a generic definition for the feature complexity. Given the feature of a certain layer in the DNN, our method decomposes and visualizes feature components of different complexity orders from the feature. The feature decomposition enables us to evaluate the reliability, the effectiveness, and the significance of over-fitting of these feature components. Furthermore, such analysis helps to improve the performance of DNNs. As a generic method, the feature complexity also provides new insights into existing deep-learning techniques, such as network compression and knowledge distillation.
Cite
Text
Ren et al. "Interpreting and Disentangling Feature Components of Various Complexity from DNNs." International Conference on Machine Learning, 2021.Markdown
[Ren et al. "Interpreting and Disentangling Feature Components of Various Complexity from DNNs." International Conference on Machine Learning, 2021.](https://mlanthology.org/icml/2021/ren2021icml-interpreting/)BibTeX
@inproceedings{ren2021icml-interpreting,
title = {{Interpreting and Disentangling Feature Components of Various Complexity from DNNs}},
author = {Ren, Jie and Li, Mingjie and Liu, Zexu and Zhang, Quanshi},
booktitle = {International Conference on Machine Learning},
year = {2021},
pages = {8971-8981},
volume = {139},
url = {https://mlanthology.org/icml/2021/ren2021icml-interpreting/}
}