JOURNAL ARTICLE

Improving Low-Resource Languages in Pre-Trained Multilingual Language Models

Abstract

Pre-trained multilingual language models are the foundation of many NLP approaches, including cross-lingual transfer solutions. However, languages with small available monolingual corpora are often not well-supported by these models leading to poor performance. We propose an unsupervised approach to improve the cross-lingual representations of low-resource languages by bootstrapping word translation pairs from monolingual corpora and using them to improve language alignment in pre-trained language models. We perform experiments on nine languages, using contextual word retrieval and zero-shot named entity recognition to measure both intrinsic cross-lingual word representation quality and downstream task performance, showing improvements on both tasks. Our results show that it is possible to improve pre-trained multilingual language models by relying only on non-parallel resources.

Keywords:
Computer science Bootstrapping (finance) Natural language processing Artificial intelligence Machine translation Word (group theory) Task (project management) Language model Quality (philosophy) Transfer of learning Linguistics

Metrics

13
Cited By
2.55
FWCI (Field Weighted Citation Impact)
45
Refs
0.87
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Natural Language Processing Techniques
Physical Sciences →  Computer Science →  Artificial Intelligence
Topic Modeling
Physical Sciences →  Computer Science →  Artificial Intelligence
Text Readability and Simplification
Physical Sciences →  Computer Science →  Artificial Intelligence

Related Documents

JOURNAL ARTICLE

Low Resource Summarization using Pre-trained Language Models

Mubashir MunafHammad AfzalKhawir MahmoodNaima Iltaf

Journal:   ACM Transactions on Asian and Low-Resource Language Information Processing Year: 2024 Vol: 23 (10)Pages: 1-19
BOOK-CHAPTER

Improving Pre-trained Language Models

Gerhard PaaßSven Giesselbach

Artificial intelligence: foundations, theory, and algorithms/Artificial intelligence: Foundations, theory, and algorithms Year: 2023 Pages: 79-159
© 2026 ScienceGate Book Chapters — All rights reserved.