Ε Kú <MASK>: Integrating Yorùbá Cultural Greetings into Machine Translation
Abstract
This paper investigates the performance of massively multilingual neural machine translation (NMT) systems in translating Yorùbá greetings (ε kú <MASK>), which are a big part of Yorùbá language and culture, into English. To evaluate these models, we present IkiniYorùbá, a Yorùbá-English translation dataset containing some Yorùbá greetings, and sample use cases. We analysed the performance of different multilingual NMT systems including Google and NLLB and show that these models struggle to accurately translate Yorùbá greetings into English. In addition, we trained a Yorùbá-English model by finetuning an existing NMT model on the training split of IkiniYorùbá and this achieved better performance when compared to the pre-trained multilingual NMT models, although they were trained on a large volume of data.
Cite
Text
Akinade et al. "Ε Kú <MASK>: Integrating Yorùbá Cultural Greetings into Machine Translation." ICLR 2023 Workshops: AfricaNLP, 2023.Markdown
[Akinade et al. "Ε Kú <MASK>: Integrating Yorùbá Cultural Greetings into Machine Translation." ICLR 2023 Workshops: AfricaNLP, 2023.](https://mlanthology.org/iclrw/2023/akinade2023iclrw-ku/)BibTeX
@inproceedings{akinade2023iclrw-ku,
title = {{Ε Kú <MASK>: Integrating Yorùbá Cultural Greetings into Machine Translation}},
author = {Akinade, Idris and Alabi, Jesujoba Oluwadara and Adelani, David Ifeoluwa and Odoje, Clement Oyeleke and Klakow, Dietrich},
booktitle = {ICLR 2023 Workshops: AfricaNLP},
year = {2023},
url = {https://mlanthology.org/iclrw/2023/akinade2023iclrw-ku/}
}