Finding Sparse Structures for Domain Specific Neural Machine Translation
Abstract
Neural machine translation often adopts the fine-tuning approach to adapt to specific domains. However, nonrestricted fine-tuning can easily degrade on the general domain and over-fit to the target domain. To mitigate the issue, we propose Prune-Tune, a novel domain adaptation method via gradual pruning. It learns tiny domain-specific sub-networks during fine-tuning on new domains. Prune-Tune alleviates the over-fitting and the degradation problem without model modification. Furthermore, Prune-Tune is able to sequentially learn a single network with multiple disjoint domain-specific sub-networks for multiple domains. Empirical experiment results show that Prune-Tune outperforms several strong competitors in the target domain test set without sacrificing the quality on the general domain in both single and multi-domain settings. The source code and data are available at https://github.com/ohlionel/Prune-Tune.
Cite
Text
Liang et al. "Finding Sparse Structures for Domain Specific Neural Machine Translation." AAAI Conference on Artificial Intelligence, 2021. doi:10.1609/AAAI.V35I15.17574Markdown
[Liang et al. "Finding Sparse Structures for Domain Specific Neural Machine Translation." AAAI Conference on Artificial Intelligence, 2021.](https://mlanthology.org/aaai/2021/liang2021aaai-finding/) doi:10.1609/AAAI.V35I15.17574BibTeX
@inproceedings{liang2021aaai-finding,
title = {{Finding Sparse Structures for Domain Specific Neural Machine Translation}},
author = {Liang, Jianze and Zhao, Chengqi and Wang, Mingxuan and Qiu, Xipeng and Li, Lei},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2021},
pages = {13333-13342},
doi = {10.1609/AAAI.V35I15.17574},
url = {https://mlanthology.org/aaai/2021/liang2021aaai-finding/}
}