Autores
Angel Gil Jason Efrain
Meque Abdul Gafar Manuel
Sidorov Grigori
Gelbukh Alexander
Título Comparing Transformer-Based Machine Translation Models for Low-Resource Languages of Colombia and Mexico
Tipo Congreso
Sub-tipo Memoria
Descripción 22nd Mexican International Conference on Artificial Intelligence, MICAI 2023
Resumen This paper offers a comparative analysis of two state-of-the-art machine translation models for Spanish to Indigenous languages of Colombia and Mexico, with the aim of investigating their effectiveness and limitations under low-resource conditions. Our methodology involved aligning verse pairs text using the Bible for twelve Indigenous languages and constructing parallel datasets for evaluation using BLEU and ROUGE metrics. The results demonstrate that transformer-based models can deliver competitive performance in translating from Spanish to Indigenous languages with minimal configuration. In particular, we found the Opus-based model obtained the best performance in 11 of the languages in the test set but, the Fairseq model performs competitively in scenarios where training data is more scarce. Additionally, we provide a comprehensive analysis of the findings, including insights into the strengths and limitations of the models. Finally, we suggest potential directions for future research in low-resource language translation, specifically in the context of Latin American indigenous languages. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2024.
Observaciones 10.1007/978-3-031-47640-2_8 Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)m v, 14392
Lugar Yucatán
País Mexico
No. de páginas 95-105
Vol. / Cap. v. 14392 LNAI
Inicio 2023-11-13
Fin 2023-11-18
ISBN/ISSN 9783031476396