BiLLM: Pushing the Limit of Post-Training Quantization for LLMs

Abstract

Pretrained large language models (LLMs) exhibit exceptional general language processing capabilities but come with significant demands on memory and computational resources. As a powerful compression technology, binarization can extremely reduce model weights to a mere 1 bit, lowering the expensive computation and memory requirements. However, existing quantization techniques fall short of maintaining LLM performance under ultra-low bit-widths. In response to this challenge, we present BiLLM, a groundbreaking 1-bit post-training quantization scheme tailored for pretrained LLMs. Based on the weight distribution of LLMs, BiLLM first identifies and structurally selects salient weights, and minimizes the compression loss through an effective binary residual approximation strategy. Moreover, considering the bell-shaped distribution of the non-salient weights, we propose an optimal splitting search to group and binarize them accurately. BiLLM, for the first time, achieves high-accuracy inference (e.g. 8.41 perplexity on LLaMA2-70B) with only 1.08-bit weights across various LLM families and evaluation metrics, outperforms SOTA quantization methods of LLM by significant margins. Moreover, BiLLM enables the binarization process of a 7-billion LLM within 0.5 hours on a single GPU, demonstrating satisfactory time efficiency. Our code is available at https://github.com/Aaronhuang-778/BiLLM .

Cite

Text

Huang et al. "BiLLM: Pushing the Limit of Post-Training Quantization for LLMs." International Conference on Machine Learning, 2024.

Markdown

[Huang et al. "BiLLM: Pushing the Limit of Post-Training Quantization for LLMs." International Conference on Machine Learning, 2024.](https://mlanthology.org/icml/2024/huang2024icml-billm/)

BibTeX

@inproceedings{huang2024icml-billm,
  title     = {{BiLLM: Pushing the Limit of Post-Training Quantization for LLMs}},
  author    = {Huang, Wei and Liu, Yangdong and Qin, Haotong and Li, Ying and Zhang, Shiming and Liu, Xianglong and Magno, Michele and Qi, Xiaojuan},
  booktitle = {International Conference on Machine Learning},
  year      = {2024},
  pages     = {20023-20042},
  volume    = {235},
  url       = {https://mlanthology.org/icml/2024/huang2024icml-billm/}
}