NIPQ: Noise Proxy-Based Integrated Pseudo-Quantization
Abstract
Straight-through estimator (STE), which enables the gradient flow over the non-differentiable function via approximation, has been favored in studies related to quantization-aware training (QAT). However, STE incurs unstable convergence during QAT, resulting in notable quality degradation in low-precision representation. Recently, pseudo-quantization training has been proposed as an alternative approach to updating the learnable parameters using the pseudo-quantization noise instead of STE. In this study, we propose a novel noise proxy-based integrated pseudo-quantization (NIPQ) that enables unified support of pseudo-quantization for both activation and weight with minimal error by integrating the idea of truncation on the pseudo-quantization framework. NIPQ updates all of the quantization parameters (e.g., bit-width and truncation boundary) as well as the network parameters via gradient descent without STE instability, resulting in greatly-simplified but reliable precision allocation without human intervention. Our extensive experiments show that NIPQ outperforms existing quantization algorithms in various vision and language applications by a large margin.
Cite
Text
Shin et al. "NIPQ: Noise Proxy-Based Integrated Pseudo-Quantization." Conference on Computer Vision and Pattern Recognition, 2023. doi:10.1109/CVPR52729.2023.00375Markdown
[Shin et al. "NIPQ: Noise Proxy-Based Integrated Pseudo-Quantization." Conference on Computer Vision and Pattern Recognition, 2023.](https://mlanthology.org/cvpr/2023/shin2023cvpr-nipq/) doi:10.1109/CVPR52729.2023.00375BibTeX
@inproceedings{shin2023cvpr-nipq,
title = {{NIPQ: Noise Proxy-Based Integrated Pseudo-Quantization}},
author = {Shin, Juncheol and So, Junhyuk and Park, Sein and Kang, Seungyeop and Yoo, Sungjoo and Park, Eunhyeok},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2023},
pages = {3852-3861},
doi = {10.1109/CVPR52729.2023.00375},
url = {https://mlanthology.org/cvpr/2023/shin2023cvpr-nipq/}
}