Abstract

Speech is the most natural form of communication for human beings and is often described as a unimodal communication channel. However, it is well known that speech is multimodal in nature and includes the auditive, visual, and tactile modalities. Other less natural modalities such as electromyographic signal, invisible articulator display, or brain electrical activity or electromagnetic activity can also be considered. Therefore, in situations where audio speech is not available or is corrupted because of disability or adverse environmental condition, people may resort to alternative methods such as augmented speech. In several automatic speech recognition systems, visual information from lips/mouth and facial movements has been used in combination with audio signals. In such cases, visual information is used to complement the audio information to improve the system’s robustness against acoustic noise (Potamianos et al., 2003). For the orally educated deaf or hearing-impaired people, lip reading remains a crucial speech modality, though it is not sufficient to achieve full communication. Therefore, in 1967, Cornett developed the Cued Speech system as a supplement to lip reading (O.Cornett, 1967). Recently, studies have been presented on automatic Cued Speech recognition using hand gestures in combination with lip/mouth information (Heracleous et al., 2009). Several other studies have been introduced that deal with the problem of alternative speech communication based on speech modalities other than audio speech. A method for communication based on inaudible speech received through body tissues has been introduced using the Non-Audible Murmur (NAM) microphone. NAM microphones have been used for receiving and automatically recognizing sounds of speech-impaired people, for ensuring privacy in communication, and for achieving robustness against noise (Heracleous et al., 2007; Nakamura et al., 2008). Aside from automatic recognition of NAM speech, silicon NAM microphones were used for NAM-to-speech conversion (Toda & Shikano, 2005; Tran et al., 2008). 15

Keywords:
Cued speech Speech recognition Gesture Computer science Modalities Modality (human–computer interaction) Voice activity detection Microphone Speech processing Natural (archaeology) Motor theory of speech perception Speech perception Psychology Human–computer interaction Artificial intelligence Perception Cognitive psychology

Metrics

0
Cited By
0.00
FWCI (Field Weighted Citation Impact)
34
Refs
0.15
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Topics

Hand Gesture Recognition Systems
Physical Sciences →  Computer Science →  Human-Computer Interaction
Hearing Impairment and Communication
Social Sciences →  Psychology →  Developmental and Educational Psychology
Robotics and Automated Systems
Physical Sciences →  Engineering →  Control and Systems Engineering

Related Documents

JOURNAL ARTICLE

Towards Personalized Speech Synthesis for Augmentative and Alternative Communication

Timothy MillsH. Timothy BunnellRupal Patel

Journal:   Augmentative and Alternative Communication Year: 2014 Vol: 30 (3)Pages: 226-236
JOURNAL ARTICLE

Attitudes of speech therapists towards the use of augmentative and alternative communication

Milica VasićSanela Slavković

Journal:   PONS - medicinski casopis Year: 2023 Vol: 20 (2)Pages: 59-65
JOURNAL ARTICLE

Book Review: Communication without speech. Augmentative communication in practice: Scotland

Margaret H. Freeman

Journal:   Child Language Teaching and Therapy Year: 1998 Vol: 14 (3)Pages: 314-315
JOURNAL ARTICLE

Towards a narrative‐based augmentative communication system

Annalu WallerAlan F. Newell

Journal:   International Journal of Language & Communication Disorders Year: 1997 Vol: 32 (S3)Pages: 289-306
JOURNAL ARTICLE

Towards a narrative-based augmentative communication system

Annalu WallerAlan C. Newell

Journal:   International Journal of Language & Communication Disorders Year: 1997 Vol: 32 Pages: 289-306
© 2026 ScienceGate Book Chapters — All rights reserved.