Purifying Quantization-Conditioned Backdoors via Layer-Wise Activation Correction with Distribution Approximation
Abstract
Model quantization is a compression technique that converts a full-precision model to a more compact low-precision version for better storage. Despite the great success of quantization, recent studies revealed the feasibility of malicious exploiting model quantization via implanting quantization-conditioned backdoors (QCBs). These special backdoors remain dormant in full-precision models but are exposed upon quantization. Unfortunately, existing defenses have limited effects on mitigating QCBs. In this paper, we conduct an in-depth analysis of QCBs. We reveal an intriguing characteristic of QCBs, where activation of backdoor-related neurons on even benign samples enjoy a distribution drift after quantization, although this drift is more significant on poisoned samples. Motivated by this finding, we propose to purify the backdoor-exposed quantized model by aligning its layer-wise activation with its full-precision version. To further exploit the more pronounced activation drifts on poisoned samples, we design an additional module to layer-wisely approximate poisoned activation distribution based on batch normalization statistics of the full-precision model. Extensive experiments are conducted, verifying the effectiveness of our defense. Our code is publicly available.
Cite
Text
Li et al. "Purifying Quantization-Conditioned Backdoors via Layer-Wise Activation Correction with Distribution Approximation." International Conference on Machine Learning, 2024.Markdown
[Li et al. "Purifying Quantization-Conditioned Backdoors via Layer-Wise Activation Correction with Distribution Approximation." International Conference on Machine Learning, 2024.](https://mlanthology.org/icml/2024/li2024icml-purifying/)BibTeX
@inproceedings{li2024icml-purifying,
title = {{Purifying Quantization-Conditioned Backdoors via Layer-Wise Activation Correction with Distribution Approximation}},
author = {Li, Boheng and Cai, Yishuo and Cai, Jisong and Li, Yiming and Qiu, Han and Wang, Run and Zhang, Tianwei},
booktitle = {International Conference on Machine Learning},
year = {2024},
pages = {27439-27456},
volume = {235},
url = {https://mlanthology.org/icml/2024/li2024icml-purifying/}
}