XLM-K: Improving Cross-Lingual Language Model Pre-Training with Multilingual Knowledge
Abstract
Cross-lingual pre-training has achieved great successes using monolingual and bilingual plain text corpora. However, most pre-trained models neglect multilingual knowledge, which is language agnostic but comprises abundant cross-lingual structure alignment. In this paper, we propose XLM-K, a cross-lingual language model incorporating multilingual knowledge in pre-training. XLM-K augments existing multilingual pre-training with two knowledge tasks, namely Masked Entity Prediction Task and Object Entailment Task. We evaluate XLM-K on MLQA, NER and XNLI. Experimental results clearly demonstrate significant improvements over existing multilingual language models. The results on MLQA and NER exhibit the superiority of XLM-K in knowledge related tasks. The success in XNLI shows a better cross-lingual transferability obtained in XLM-K. What is more, we provide a detailed probing analysis to confirm the desired knowledge captured in our pre-training regimen. The code is available at https://github.com/microsoft/Unicoder/tree/master/pretraining/xlmk.
Cite
Text
Jiang et al. "XLM-K: Improving Cross-Lingual Language Model Pre-Training with Multilingual Knowledge." AAAI Conference on Artificial Intelligence, 2022. doi:10.1609/AAAI.V36I10.21330Markdown
[Jiang et al. "XLM-K: Improving Cross-Lingual Language Model Pre-Training with Multilingual Knowledge." AAAI Conference on Artificial Intelligence, 2022.](https://mlanthology.org/aaai/2022/jiang2022aaai-xlm/) doi:10.1609/AAAI.V36I10.21330BibTeX
@inproceedings{jiang2022aaai-xlm,
title = {{XLM-K: Improving Cross-Lingual Language Model Pre-Training with Multilingual Knowledge}},
author = {Jiang, Xiaoze and Liang, Yaobo and Chen, Weizhu and Duan, Nan},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2022},
pages = {10840-10848},
doi = {10.1609/AAAI.V36I10.21330},
url = {https://mlanthology.org/aaai/2022/jiang2022aaai-xlm/}
}