JOURNAL ARTICLE

Augmenting Pre-trained Language Models with QA-Memory for Open-Domain Question Answering

Abstract

Existing state-of-the-art methods for open-domain question-answering (ODQA) use an open book approach in which information is first retrieved from a large text corpus or knowledge base (KB) and then reasoned over to produce an answer. A recent alternative is to retrieve from a collection of previously-generated question-answer pairs; this has several practical advantages including being more memory and compute-efficient. Question-answer pairs are also appealing in that they can be viewed as an intermediate between text and KB triples: like KB triples, they often concisely express a single relationship, but like text, have much higher coverage than traditional KBs. In this work, we describe a new QA system that augments a text-to-text model with a large memory of question-answer pairs, and a new pre-training task for the latent step of question retrieval. The pre-training task substantially simplifies training and greatly improves performance on smaller QA benchmarks. Unlike prior systems of this sort, our QA system can also answer multi-hop questions that do not explicitly appear in the collection of stored question-answer pairs.

Keywords:
Question answering Computer science sort Task (project management) Open domain Information retrieval Language model Natural language processing Artificial intelligence Knowledge base Domain (mathematical analysis)

Metrics

12
Cited By
1.79
FWCI (Field Weighted Citation Impact)
42
Refs
0.84
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Topic Modeling
Physical Sciences →  Computer Science →  Artificial Intelligence
Natural Language Processing Techniques
Physical Sciences →  Computer Science →  Artificial Intelligence
Multimodal Machine Learning Applications
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition

Related Documents

BOOK-CHAPTER

L2R-QA: An Open-Domain Question Answering Framework

Tieke HeLi YuZhipeng ZouQing Wu

Lecture notes in computer science Year: 2019 Pages: 151-162
BOOK-CHAPTER

Open-Domain Question Answering over Tables with Large Language Models

Xinyi LiangRui HuYu LiuKonglin Zhu

Lecture notes in computer science Year: 2024 Pages: 347-358
JOURNAL ARTICLE

Audio-Aware Spoken Multiple-Choice Question Answering With Pre-Trained Language Models

Chia-Chih KuoKuan-Yu ChenShang-Bao Luo

Journal:   IEEE/ACM Transactions on Audio Speech and Language Processing Year: 2021 Vol: 29 Pages: 3170-3179
© 2026 ScienceGate Book Chapters — All rights reserved.