Autores
Tonja Atnafu Lambebo
Kolesnikova Olga
Arif Muhammad
Gelbukh Alexander
Sidorov Grigori
Título Improving Neural Machine Translation for Low Resource Languages Using Mixed Training: The Case of Ethiopian Languages
Tipo Congreso
Sub-tipo Memoria
Descripción 21st Mexican International Conference on Artificial Intelligence, MICAI 2022
Resumen Neural Machine Translation (NMT) has shown improvement for high-resource languages, but there is still a problem with low-resource languages as NMT performs well on huge parallel data available for high-resource languages. In spite of many proposals to solve the problem of low-resource languages, it continues to be a difficult challenge. The issue becomes even more complicated when few resources cover only one domain. In our attempt to combat this issue, we propose a new approach to improve NMT for low-resource languages. The proposed approach using the transformer model shows 5.3, 5.0, and 3.7 BLEU score improvement for Gamo-English, Gofa-English, and Dawuro-English language pairs, respectively, where Gamo, Gofa, and Dawuro are related low-resource Ethiopian languages. We discuss our contributions and envisage future steps in this challenging research area. © 2022, The Author(s), under exclusive license to Springer Nature Switzerland AG.
Observaciones DOI 10.1007/978-3-031-19496-2_3 Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), v. 13613
Lugar Monterrey
País Mexico
No. de páginas 30-40
Vol. / Cap. LNCS 13613
Inicio 2022-10-24
Fin 2022-10-29
ISBN/ISSN 9783031194955