JOURNAL ARTICLE

Performance Evaluation of Transformer-Based Pre-Trained Language Models for Turkish Question-Answering

Mert İncidelenMurat Aydoğan

Year: 2025 Journal:   Black Sea Journal of Engineering and Science Vol: 8 (2)Pages: 323-329

Abstract

Natural language processing (NLP) has made significant progress with the introduction of Transformer-based architectures that have revolutionized tasks such as question-answering (QA). While English is a primary focus of NLP research due to its high resource datasets, low-resource languages such as Turkish present unique challenges such as linguistic complexity and limited data availability. This study evaluates the performance of Transformer-based pre-trained language models on QA tasks and provides insights into their strengths and limitations for future improvements. In the study, using the SQuAD-TR dataset, which is the machine-translated Turkish version of the SQuAD 2.0 dataset, variations of the mBERT, BERTurk, ConvBERTurk, DistilBERTurk, and ELECTRA Turkish pre-trained models were fine-tuned. The performance of these fine-tuned models was tested using the XQuAD-TR dataset. The models were evaluated using Exact Match (EM) Rate and F1 Score metrics. Among the tested models, the ConvBERTurk Base (cased) model performed the best, achieving an EM Rate of 57.81512% and an F1 Score of 71.58769%. In contrast, the DistilBERTurk Base (cased) and ELECTRA TR Small (cased) models performed poorly due to their smaller size and fewer parameters. The results indicate that case-sensitive models generally perform better than case-insensitive models. The ability of case-sensitive models to discriminate proper names and abbreviations more effectively improved their performance. Moreover, models specifically adapted for Turkish performed better on QA tasks compared to the multilingual mBERT model.

Keywords:
Turkish Question answering Transformer Computer science Natural language processing Artificial intelligence Linguistics Engineering Electrical engineering Philosophy

Metrics

0
Cited By
0.00
FWCI (Field Weighted Citation Impact)
20
Refs
0.02
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Topics

Topic Modeling
Physical Sciences →  Computer Science →  Artificial Intelligence
Natural Language Processing Techniques
Physical Sciences →  Computer Science →  Artificial Intelligence
Text and Document Classification Technologies
Physical Sciences →  Computer Science →  Artificial Intelligence

Related Documents

JOURNAL ARTICLE

EXPLORING THE EFFECTIVENESS OF PRE-TRAINED TRANSFORMER MODELS FOR TURKISH QUESTION ANSWERING

Abdullah Talha Kabakuş

Journal:   Kahramanmaraş Sütçü İmam Üniversitesi Mühendislik Bilimleri Dergisi Year: 2025 Vol: 28 (2)Pages: 975-993
BOOK-CHAPTER

Question Answering Systems Based on Pre-trained Language Models: Recent Progress

Xudong LuoYing LuoBinxia Yang

IFIP advances in information and communication technology Year: 2024 Pages: 173-189
JOURNAL ARTICLE

Comparison of Transformer-Based Turkish Models for Question-Answering Task

Mehmet ArzuMurat Aydoğan

Journal:   Balkan Journal of Electrical and Computer Engineering Year: 2025 Vol: 12 (4)Pages: 387-393
BOOK-CHAPTER

Pre-trained Language Model for Biomedical Question Answering

Wonjin YoonJinhyuk LeeDonghyeon KimMinbyul JeongJaewoo Kang

Communications in computer and information science Year: 2020 Pages: 727-740
JOURNAL ARTICLE

Pre-trained transformer-based language models for Sundanese

Wilson WongsoHenry LuckyDerwin Suhartono

Journal:   Journal Of Big Data Year: 2022 Vol: 9 (1)
© 2026 ScienceGate Book Chapters — All rights reserved.