Language Model Pre-Training on True Negatives
Abstract
Discriminative pre-trained language models (PrLMs) learn to predict original texts from intentionally corrupted ones. Taking the former text as positive and the latter as negative samples, the PrLM can be trained effectively for contextualized representation. However, the training of such a type of PrLMs highly relies on the quality of the automatically constructed samples. Existing PrLMs simply treat all corrupted texts as equal negative without any examination, which actually lets the resulting model inevitably suffer from the false negative issue where training is carried out on pseudo-negative data and leads to less efficiency and less robustness in the resulting PrLMs. In this work, on the basis of defining the false negative issue in discriminative PrLMs that has been ignored for a long time, we design enhanced pre-training methods to counteract false negative predictions and encourage pre-training language models on true negatives by correcting the harmful gradient updates subject to false negative predictions. Experimental results on GLUE and SQuAD benchmarks show that our counter-false-negative pre-training methods indeed bring about better performance together with stronger robustness.
Cite
Text
Zhang et al. "Language Model Pre-Training on True Negatives." AAAI Conference on Artificial Intelligence, 2023. doi:10.1609/AAAI.V37I11.26639Markdown
[Zhang et al. "Language Model Pre-Training on True Negatives." AAAI Conference on Artificial Intelligence, 2023.](https://mlanthology.org/aaai/2023/zhang2023aaai-language/) doi:10.1609/AAAI.V37I11.26639BibTeX
@inproceedings{zhang2023aaai-language,
title = {{Language Model Pre-Training on True Negatives}},
author = {Zhang, Zhuosheng and Zhao, Hai and Utiyama, Masao and Sumita, Eiichiro},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2023},
pages = {14002-14010},
doi = {10.1609/AAAI.V37I11.26639},
url = {https://mlanthology.org/aaai/2023/zhang2023aaai-language/}
}