HLHLp: Quantized Neural Networks Training for Reaching Flat Minima in Loss Surface

Abstract

Quantization of deep neural networks is extremely essential for efficient implementations. Low-precision networks are typically designed to represent original floating-point counterparts with high fidelity, and several elaborate quantization algorithms have been developed. We propose a novel training scheme for quantized neural networks to reach flat minima in the loss surface with the aid of quantization noise. The proposed training scheme employs high-low-high-low precision in an alternating manner for network training. The learning rate is also abruptly changed at each stage for coarse- or fine-tuning. With the proposed training technique, we show quite good performance improvements for convolutional neural networks when compared to the previous fine-tuning based quantization scheme. We achieve the state-of-the-art results for recurrent neural network based language modeling with 2-bit weight and activation.

Cite

Text

Shin et al. "HLHLp: Quantized Neural Networks Training for Reaching Flat Minima in Loss Surface." AAAI Conference on Artificial Intelligence, 2020. doi:10.1609/AAAI.V34I04.6035

Markdown

[Shin et al. "HLHLp: Quantized Neural Networks Training for Reaching Flat Minima in Loss Surface." AAAI Conference on Artificial Intelligence, 2020.](https://mlanthology.org/aaai/2020/shin2020aaai-hlhlp/) doi:10.1609/AAAI.V34I04.6035

BibTeX

@inproceedings{shin2020aaai-hlhlp,
  title     = {{HLHLp: Quantized Neural Networks Training for Reaching Flat Minima in Loss Surface}},
  author    = {Shin, Sungho and Park, Jinhwan and Boo, Yoonho and Sung, Wonyong},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2020},
  pages     = {5784-5791},
  doi       = {10.1609/AAAI.V34I04.6035},
  url       = {https://mlanthology.org/aaai/2020/shin2020aaai-hlhlp/}
}