Contrastive Preference Optimization: Pushing the Boundaries of LLM Performance in Machine Translation

Abstract

Moderate-sized large language models (LLMs) – those with 7B or 13B parameters – exhibit promising machine translation (MT) performance. However, they do not match the performance of state-of-the-art conventional encoder-decoder translation models or larger-scale LLMs such as GPT-4. In this study, we bridge this performance gap. We first assess the shortcomings of supervised fine-tuning for LLMs in the MT task, emphasizing the quality issues present in the reference data, despite being human-generated. Then, in contrast to supervised fine-tuning which mimics reference translations, we introduce Contrastive Preference Optimization (CPO), a novel approach that trains models to avoid generating adequate but not perfect translations. Applying CPO to ALMA models with only 22K parallel sentences and 0.1% parameters yields significant improvements. The resulting model, called ALMA-R, can match or exceed the performance of the WMT competition winners and GPT-4 on WMT’21, WMT’22 and WMT’23 test datasets.

Cite

Text

Xu et al. "Contrastive Preference Optimization: Pushing the Boundaries of LLM Performance in Machine Translation." International Conference on Machine Learning, 2024.

Markdown

[Xu et al. "Contrastive Preference Optimization: Pushing the Boundaries of LLM Performance in Machine Translation." International Conference on Machine Learning, 2024.](https://mlanthology.org/icml/2024/xu2024icml-contrastive/)

BibTeX

@inproceedings{xu2024icml-contrastive,
  title     = {{Contrastive Preference Optimization: Pushing the Boundaries of LLM Performance in Machine Translation}},
  author    = {Xu, Haoran and Sharaf, Amr and Chen, Yunmo and Tan, Weiting and Shen, Lingfeng and Van Durme, Benjamin and Murray, Kenton and Kim, Young Jin},
  booktitle = {International Conference on Machine Learning},
  year      = {2024},
  pages     = {55204-55224},
  volume    = {235},
  url       = {https://mlanthology.org/icml/2024/xu2024icml-contrastive/}
}