JOURNAL ARTICLE

Gradual unfreezing transformer-based language models for biomedical question answering

Abstract

Pretrained transformer-based language models have achieved state-of-the-art results on various Natural Language Processing (NLP) tasks. These models can be fine-tuned on a range of downstream tasks with minimalistic modifications. However, fine-tuning a language model may result in the problem of catastrophic forgetting and tend to overfit on smaller training datasets. Therefore, gradually unfreezing the pretrained weights is a possible approach to avoid catastrophic forgetting of the knowledge learnt from the source task. Multi-task fine-tuning is an intermediate step on a high-resource dataset that yields good results for low-resource tasks. In this project, we will be investigating the strategies of multi-task fine-tuning and gradual unfreezing on DistilBERT, which have not yet been applied for biomedical domain. First, we explore whether DistilBERT improves the accuracy of a low-resource dataset, BioASQ, with question answering (QA) task as our NLP use-case. Second, we investigate the effect that gradual unfreezing has on the performance of DistilBERT. We observe that despite being 40% smaller and without any domain-specific pretraining, DistilBERT achieves comparable results to a larger model, BERT on smaller BioASQ dataset. However, we observed that gradually unfreezing DistilBERT has no significant impact on the results of our QA task in comparison to standard non-gradual fine-tuning.

Keywords:
Question answering Overfitting Language model Forgetting Task (project management) Language understanding Generalization

Metrics

0
Cited By
0.00
FWCI (Field Weighted Citation Impact)
0
Refs
0.31
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Topics

Topic Modeling
Physical Sciences →  Computer Science →  Artificial Intelligence
Machine Learning in Healthcare
Physical Sciences →  Computer Science →  Artificial Intelligence
Artificial Intelligence in Healthcare and Education
Health Sciences →  Medicine →  Health Informatics

Related Documents

JOURNAL ARTICLE

Gradual unfreezing transformer-based language models for biomedical question answering

Khanna, Urvashi

Journal:   OPAL (Open@LaTrobe) (La Trobe University) Year: 2022
JOURNAL ARTICLE

Transformer-based Language Models for Factoid Question Answering at BioASQ9b

Urvashi KhannaDiego Mollá

Journal:   arXiv (Cornell University) Year: 2021 Pages: 247-257
JOURNAL ARTICLE

Enhancing Biomedical Question Answering with Large Language Models

Hua YangShilong LiTeresa Gonçalves

Journal:   Information Year: 2024 Vol: 15 (8)Pages: 494-494
JOURNAL ARTICLE

QUESTION ANSWERING SYSTEM FOR HOSPITALITY DOMAIN USING TRANSFORMER-BASED LANGUAGE MODELS

Sathish Sathish Dhanasegar

Journal:   International Research Journal of Computer Science Year: 2022 Vol: 9 (5)Pages: 110-134
© 2026 ScienceGate Book Chapters — All rights reserved.