Abstract

Conversational Question Answering (CQA) aims to answer questions contained within dialogues, which are not easily interpretable without context. Developing a model to rewrite conversational questions into self-contained ones is an emerging solution in industry settings as it allows using existing single-turn QA systems to avoid training a CQA model from scratch. Previous work trains rewriting models using human rewrites as supervision. However, such objectives are disconnected with QA models and therefore more human-like rewrites do not guarantee better QA performance. In this paper we propose using QA feedback to supervise the rewriting model with reinforcement learning. Experiments show that our approach can effectively improve QA performance over baselines for both extractive and retrieval QA. Furthermore, human evaluation shows that our method can generate more accurate and detailed rewrites when compared to human annotations.

Keywords:
Rewriting Computer science Question answering Context (archaeology) Reinforcement learning Natural language processing Information retrieval Artificial intelligence Programming language

Metrics

23
Cited By
4.50
FWCI (Field Weighted Citation Impact)
23
Refs
0.93
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Topic Modeling
Physical Sciences →  Computer Science →  Artificial Intelligence
Natural Language Processing Techniques
Physical Sciences →  Computer Science →  Artificial Intelligence
Speech and dialogue systems
Physical Sciences →  Computer Science →  Artificial Intelligence

Related Documents

BOOK-CHAPTER

Conversational Question Answering

Rishiraj Saha RoyAvishek Anand

Synthesis lectures on information concepts, retrieval, and services Year: 2022 Pages: 129-136
BOOK-CHAPTER

Conversational Question Answering

Rishiraj Saha RoyAvishek Anand

Synthesis lectures on information concepts, retrieval, and services Year: 2022 Pages: 65-76
© 2026 ScienceGate Book Chapters — All rights reserved.