Towards the Dynamics of a DNN Learning Symbolic Interactions
Abstract
This study proves the two-phase dynamics of a deep neural network (DNN) learning interactions. Despite the long disappointing view of the faithfulness of post-hoc explanation of a DNN, a series of theorems have been proven [27] in recent years to show that for a given input sample, a small set of interactions between input variables can be considered as primitive inference patterns that faithfully represent a DNN's detailed inference logic on that sample. Particularly, Zhang et al. [41] have observed that various DNNs all learn interactions of different complexities in two distinct phases, and this two-phase dynamics well explains how a DNN changes from under-fitting to over-fitting. Therefore, in this study, we mathematically prove the two-phase dynamics of interactions, providing a theoretical mechanism for how the generalization power of a DNN changes during the training process. Experiments show that our theory well predicts the real dynamics of interactions on different DNNs trained for various tasks.
Cite
Text
Ren et al. "Towards the Dynamics of a DNN Learning Symbolic Interactions." Neural Information Processing Systems, 2024. doi:10.52202/079017-1602Markdown
[Ren et al. "Towards the Dynamics of a DNN Learning Symbolic Interactions." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/ren2024neurips-dynamics/) doi:10.52202/079017-1602BibTeX
@inproceedings{ren2024neurips-dynamics,
title = {{Towards the Dynamics of a DNN Learning Symbolic Interactions}},
author = {Ren, Qihan and Zhang, Junpeng and Xu, Yang and Xin, Yue and Liu, Dongrui and Zhang, Quanshi},
booktitle = {Neural Information Processing Systems},
year = {2024},
doi = {10.52202/079017-1602},
url = {https://mlanthology.org/neurips/2024/ren2024neurips-dynamics/}
}