JOURNAL ARTICLE

Specification and realisation of multimodal output in dialogue systems

Abstract

We present a high level formalism for specifying verbal and nonverbal output from a multimodal dialogue system. The output specification is XML-based and provides information about communicative functions of the output without detailing the realisation of these functions. The specification can be used to control an animated character that uses speech and gestures. We give examples from an implementation in a multimodal spoken dialogue system, and describe how facial gestures are implemented in a 3Danimated talking agent within this system.

Keywords:
Realisation Gesture Computer science Formalism (music) Multimodality XML Nonverbal communication Human–computer interaction Speech synthesis Multimodal interaction Artificial intelligence Natural language processing Communication Psychology World Wide Web

Metrics

6
Cited By
0.74
FWCI (Field Weighted Citation Impact)
14
Refs
0.72
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Topics

Speech and dialogue systems
Physical Sciences →  Computer Science →  Artificial Intelligence
Natural Language Processing Techniques
Physical Sciences →  Computer Science →  Artificial Intelligence
Social Robot Interaction and HRI
Social Sciences →  Psychology →  Social Psychology
© 2026 ScienceGate Book Chapters — All rights reserved.