Adaptive Low-Precision Training for Embeddings in Click-Through Rate Prediction
Abstract
Embedding tables are usually huge in click-through rate (CTR) prediction models. To train and deploy the CTR models efficiently and economically, it is necessary to compress their embedding tables. To this end, we formulate a novel quantization training paradigm to compress the embeddings from the training stage, termed low-precision training (LPT). Also, we provide theoretical analysis on its convergence. The results show that stochastic weight quantization has a faster convergence rate and a smaller convergence error than deterministic weight quantization in LPT. Further, to reduce accuracy degradation, we propose adaptive low-precision training (ALPT) which learns the step size (i.e., the quantization resolution). Experiments on two real-world datasets confirm our analysis and show that ALPT can significantly improve the prediction accuracy, especially at extremely low bit width. For the first time in CTR models, we successfully train 8-bit embeddings without sacrificing prediction accuracy.
Cite
Text
Li et al. "Adaptive Low-Precision Training for Embeddings in Click-Through Rate Prediction." AAAI Conference on Artificial Intelligence, 2023. doi:10.1609/AAAI.V37I4.25564Markdown
[Li et al. "Adaptive Low-Precision Training for Embeddings in Click-Through Rate Prediction." AAAI Conference on Artificial Intelligence, 2023.](https://mlanthology.org/aaai/2023/li2023aaai-adaptive/) doi:10.1609/AAAI.V37I4.25564BibTeX
@inproceedings{li2023aaai-adaptive,
title = {{Adaptive Low-Precision Training for Embeddings in Click-Through Rate Prediction}},
author = {Li, Shiwei and Guo, Huifeng and Hou, Lu and Zhang, Wei and Tang, Xing and Tang, Ruiming and Zhang, Rui and Li, Ruixuan},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2023},
pages = {4435-4443},
doi = {10.1609/AAAI.V37I4.25564},
url = {https://mlanthology.org/aaai/2023/li2023aaai-adaptive/}
}