PROFIT: A Novel Training Method for Sub-4-Bit MobileNet Models
Abstract
4-bit and lower precision mobile models are required due to the ever-increasing demand for better energy efficiency in mobile devices. In this work, we report that the activation instability induced by weight quantization (AIWQ) is the key obstacle to sub-4-bit quantization of mobile networks. To alleviate the AIWQ problem, we propose a novel training method called PROgressive-Freezing Iterative Training (PROFIT), which attempts to freeze layers whose weights are affected by the instability problem stronger than the other layers. We also propose a differentiable and unified quantization method (DuQ) and a negative padding idea to support asymmetric activation functions such as h-swish. We evaluate the proposed methods by quantizing MobileNet-v1, v2, and v3 on ImageNet and report that 4-bit quantization offers comparable (within 1.48 % top-1 accuracy) accuracy to full precision baseline. In the ablation study of the 3-bit quantization of MobileNet-v3, our proposed method outperforms the state-of-the-art method by a large margin, 12.86 % of top-1 accuracy. The quantized model and source code is available at https://github.com/EunhyeokPark/PROFIT.
Cite
Text
Park and Yoo. "PROFIT: A Novel Training Method for Sub-4-Bit MobileNet Models." Proceedings of the European Conference on Computer Vision (ECCV), 2020. doi:10.1007/978-3-030-58539-6_26Markdown
[Park and Yoo. "PROFIT: A Novel Training Method for Sub-4-Bit MobileNet Models." Proceedings of the European Conference on Computer Vision (ECCV), 2020.](https://mlanthology.org/eccv/2020/park2020eccv-profit/) doi:10.1007/978-3-030-58539-6_26BibTeX
@inproceedings{park2020eccv-profit,
title = {{PROFIT: A Novel Training Method for Sub-4-Bit MobileNet Models}},
author = {Park, Eunhyeok and Yoo, Sungjoo},
booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
year = {2020},
doi = {10.1007/978-3-030-58539-6_26},
url = {https://mlanthology.org/eccv/2020/park2020eccv-profit/}
}