Towards Interpreting Computer Vision Based on Transformation Invariant Optimization

Abstract

Interpreting how does deep neural networks (DNNs) make predictions is a vital field in artificial intelligence, which hinders wide applications of DNNs. Visualization of learned representations helps we humans understand the vision of DNNs. In this work, visualized images that can activate the neural network to the target classes are generated by back-propagation method. Here, rotation and scaling operations are applied to introduce the transformation invariance in the image generating process, which we find a significant improvement on visualization effect. Finally, we show some cases that such method can help us to gain insight into neural networks.

Cite

Text

Li et al. "Towards Interpreting Computer Vision Based on Transformation Invariant Optimization." European Conference on Computer Vision Workshops, 2022. doi:10.1007/978-3-031-25075-0_26

Markdown

[Li et al. "Towards Interpreting Computer Vision Based on Transformation Invariant Optimization." European Conference on Computer Vision Workshops, 2022.](https://mlanthology.org/eccvw/2022/li2022eccvw-interpreting/) doi:10.1007/978-3-031-25075-0_26

BibTeX

@inproceedings{li2022eccvw-interpreting,
  title     = {{Towards Interpreting Computer Vision Based on Transformation Invariant Optimization}},
  author    = {Li, Chen and Jiang, Jinzhe and Zhang, Xin and Zhang, Tonghuan and Zhao, Yaqian and Jiang, Dongdong and Li, Rengang},
  booktitle = {European Conference on Computer Vision Workshops},
  year      = {2022},
  pages     = {371-382},
  doi       = {10.1007/978-3-031-25075-0_26},
  url       = {https://mlanthology.org/eccvw/2022/li2022eccvw-interpreting/}
}