JOURNAL ARTICLE

Text-to-visual speech synthesis based on parameter generation from HMM

Abstract

This paper presents a new technique for synthesizing visual speech from arbitrarily given text. The technique is based on an algorithm for parameter generation from HMM with dynamic features, which has been successfully applied to text-to-speech synthesis. In the training phase, syllable HMMs are trained with visual speech parameter sequences that represent lip movements. In the synthesis phase, a sentence HMM is constructed by concatenating syllable HMMs corresponding to the phonetic transcription for the input text. Then an optimum visual speech parameter sequence is generated from the sentence HMM in an ML sense. The proposed technique can generate synchronized lip movements with speech in a unified framework. Furthermore, coarticulation is implicitly incorporated into the generated mouth shapes. As a result, synthetic lip motion becomes smooth and realistic.

Keywords:
Coarticulation Hidden Markov model Computer science Speech synthesis Speech recognition Sentence Syllable Artificial intelligence Transcription (linguistics) Natural language processing Linguistics

Metrics

72
Cited By
5.43
FWCI (Field Weighted Citation Impact)
26
Refs
0.97
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Speech and Audio Processing
Physical Sciences →  Computer Science →  Signal Processing
Face recognition and analysis
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Social Robot Interaction and HRI
Social Sciences →  Psychology →  Social Psychology
© 2026 ScienceGate Book Chapters — All rights reserved.