JOURNAL ARTICLE

Enhancing End-to-End Speech Synthesis by Modeling Interrogative Sentences with Speaker Adaptation

Abstract

Despite end-to-end text-to-speech (TTS) synthesizers producing human-like speech, they might still need more intuitive user control over prosody. Modeling interrogative sentence prosody is challenging due to the significant variation in question types. Synthesized intonation frequently requires more accuracy, richness, and detail when only a small amount of adaptation data from particular sentence types are available. This paper uses speaker adaptation to enhance the modeling of interrogative sentence prosody in speech synthesis, tested on an English dataset. The adaptation data were selected based on the occurrence of interrogative sentences. The first dataset consisted of sentences with frequent interrogative sentences, whereas the second dataset contained declarative sentences. Two target speakers (male and female) were adapted. Objective and subjective evaluations show that the proposed model achieves remarkable performance in intonation. The MUSHRA subjective listening test has shown better intonation patterns using the interrogative dataset than the declarative one. The potential application of this model is for the vision impaired and chatbots/voice bots.

Keywords:
Interrogative Intonation (linguistics) Prosody Computer science Speech recognition Sentence Natural language processing Adaptation (eye) Active listening Interrogative word Artificial intelligence Speech synthesis Linguistics Psychology Communication

Metrics

2
Cited By
0.51
FWCI (Field Weighted Citation Impact)
40
Refs
0.67
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Speech Recognition and Synthesis
Physical Sciences →  Computer Science →  Artificial Intelligence
Speech and dialogue systems
Physical Sciences →  Computer Science →  Artificial Intelligence
Natural Language Processing Techniques
Physical Sciences →  Computer Science →  Artificial Intelligence
© 2026 ScienceGate Book Chapters — All rights reserved.