Hierarchical Knowledge Squeezed Adversarial Network Compression

Abstract

Deep network compression has been achieved notable progress via knowledge distillation, where a teacher-student learning manner is adopted by using predetermined loss. Recently, more focuses have been transferred to employ the adversarial training to minimize the discrepancy between distributions of output from two networks. However, they always emphasize on result-oriented learning while neglecting the scheme of process-oriented learning, leading to the loss of rich information contained in the whole network pipeline. Whereas in other (non GAN-based) process-oriented meth-ods, the knowledge have usually been transferred in a redundant manner. Observing that, the small network can not perfectly mimic a large one due to the huge gap of network scale, we propose a knowledge transfer method, involving effective intermediate supervision, under the adversarial training framework to learn the student network. Different from the other intermediate supervision methods, we design the knowledge representation in a compact form by introducing a task-driven attention mechanism. Meanwhile, to improve the representation capability of the attention-based method, a hierarchical structure is utilized so that powerful but highly squeezed knowledge is realized and the knowledge from teacher network could accommodate the size of student network. Extensive experimental results on three typical benchmark datasets, i.e., CIFAR-10, CIFAR-100, and ImageNet, demonstrate that our method achieves highly superior performances against state-of-the-art

Cite

Text

Li et al. "Hierarchical Knowledge Squeezed Adversarial Network Compression." AAAI Conference on Artificial Intelligence, 2020. doi:10.1609/AAAI.V34I07.6799

Markdown

[Li et al. "Hierarchical Knowledge Squeezed Adversarial Network Compression." AAAI Conference on Artificial Intelligence, 2020.](https://mlanthology.org/aaai/2020/li2020aaai-hierarchical/) doi:10.1609/AAAI.V34I07.6799

BibTeX

@inproceedings{li2020aaai-hierarchical,
  title     = {{Hierarchical Knowledge Squeezed Adversarial Network Compression}},
  author    = {Li, Peng and Shu, Chang and Xie, Yuan and Qu, Yan and Kong, Hui},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2020},
  pages     = {11370-11377},
  doi       = {10.1609/AAAI.V34I07.6799},
  url       = {https://mlanthology.org/aaai/2020/li2020aaai-hierarchical/}
}