Machine translation (MT) involving Indigenous languages, including endangered ones, is challenging primarily due to lack of sufficient parallel data. We describe an approach exploiting bilingual and multilingual pretrained MT models in a transfer learning setting to translate from Spanish into ten South American Indigenous languages. Our models set new SOTA on five out of the ten language pairs we consider, even doubling performance on one of these five pairs. Unlike previous SOTA that perform data augmentation to enlarge the train sets, we retain the low-resource setting to test the effectiveness of our models under such a constraint. In spite of the rarity of linguistic information available about the Indigenous languages, we offer a number of quantitative and qualitative analyses (e.g., as to morphology, tokenization, and orthography) to contextualize our results.
Weijia XuYuwei YinShuming MaDongdong ZhangHaoyang Huang
Sudhansu Bala DasAtharv BiradarTapas Kumar MishraBidyut Kr. Patra
Renz Iver BaliberCharibeth ChengKristine Mae M. AdlaonVirgion H. Mamonong
Luyu GaoXinyi WangGraham Neubig