Deep Transferring Quantization

Abstract

Network quantization is an effective method for network compression. Existing methods train a low-precision network by fine-tuning from a pre-trained model. However, training a low-precision network often requires large-scale labeled data to achieve superior performance. In many real-world scenarios, only limited labeled data are available due to expensive labeling costs or privacy protection. With limited training data, fine-tuning methods may suffer from the overfitting issue and substantial accuracy loss. To alleviate these issues, we introduce transfer learning into network quantization to obtain an accurate low-precision model. Specifically, we propose a method named deep transferring quantization (DTQ) to effectively exploit the knowledge in a pre-trained full-precision model. To this end, we propose a learnable attentive transfer module to identify the informative channels for alignment. In addition, we introduce the Kullback–Leibler (KL) divergence to further help train a low-precision model. Extensive experiments on both image classification and face recognition demonstrate the effectiveness of DTQ.

Cite

Text

Xie et al. "Deep Transferring Quantization." Proceedings of the European Conference on Computer Vision (ECCV), 2020. doi:10.1007/978-3-030-58598-3_37

Markdown

[Xie et al. "Deep Transferring Quantization." Proceedings of the European Conference on Computer Vision (ECCV), 2020.](https://mlanthology.org/eccv/2020/xie2020eccv-deep/) doi:10.1007/978-3-030-58598-3_37

BibTeX

@inproceedings{xie2020eccv-deep,
  title     = {{Deep Transferring Quantization}},
  author    = {Xie, Zheng and Wen, Zhiquan and Liu, Jing and Liu, Zhiqiang and Wu, Xixian and Tan, Mingkui},
  booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
  year      = {2020},
  doi       = {10.1007/978-3-030-58598-3_37},
  url       = {https://mlanthology.org/eccv/2020/xie2020eccv-deep/}
}