Unsupervised Domain Adaptation of a Pretrained Cross-Lingual Language Model

Abstract

Recent research indicates that pretraining cross-lingual language models on large-scale unlabeled texts yields significant performance improvements over various cross-lingual and low-resource tasks. Through training on one hundred languages and terabytes of texts, cross-lingual language models have proven to be effective in leveraging high-resource languages to enhance low-resource language processing and outperform monolingual models. In this paper, we further investigate the cross-lingual and cross-domain (CLCD) setting when a pretrained cross-lingual language model needs to adapt to new domains. Specifically, we propose a novel unsupervised feature decomposition method that can automatically extract domain-specific features and domain-invariant features from the entangled pretrained cross-lingual representations, given unlabeled raw texts in the source language. Our proposed model leverages mutual information estimation to decompose the representations computed by a cross-lingual model into domain-invariant and domain-specific parts. Experimental results show that our proposed method achieves significant performance improvements over the state-of-the-art pretrained cross-lingual language model in the CLCD setting.

Cite

Text

Li et al. "Unsupervised Domain Adaptation of a Pretrained Cross-Lingual Language Model." International Joint Conference on Artificial Intelligence, 2020. doi:10.24963/IJCAI.2020/508

Markdown

[Li et al. "Unsupervised Domain Adaptation of a Pretrained Cross-Lingual Language Model." International Joint Conference on Artificial Intelligence, 2020.](https://mlanthology.org/ijcai/2020/li2020ijcai-unsupervised/) doi:10.24963/IJCAI.2020/508

BibTeX

@inproceedings{li2020ijcai-unsupervised,
  title     = {{Unsupervised Domain Adaptation of a Pretrained Cross-Lingual Language Model}},
  author    = {Li, Juntao and He, Ruidan and Ye, Hai and Ng, Hwee Tou and Bing, Lidong and Yan, Rui},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2020},
  pages     = {3672-3678},
  doi       = {10.24963/IJCAI.2020/508},
  url       = {https://mlanthology.org/ijcai/2020/li2020ijcai-unsupervised/}
}