JOURNAL ARTICLE

1P1-K06 Audio-Visual Speaker Detection in Human-Robot Interaction

Thatsaphan SuwannathatJun‐ichi ImaiMasahide Kaneko

Year: 2007 Journal:   The Proceedings of JSME annual Conference on Robotics and Mechatronics (Robomec) Vol: 2007 (0)Pages: _1P1-K06_1   Publisher: Japan Society Mechanical Engineers

Abstract

Tracking humans' position is a useful skill for the coming generation of mobile robot. It is a challenging problem of planning and control in dynamic environment. We propose the omni-directional estimation method of speaker's position using the combination of audio and visual information. Estimation of the position of the sound is carried out to calculate the difference of arrival time from sound source to multi-channel microphones. The robust human template matching on the omni-directional image is employed to combine the result of sound source estimation to realize a highly accurate estimation of speaker's location. In our experiments, the systems were implemented and tested on an omni-directional robot at our laboratory. The results show that we are able to reliably detect and track moving objects in natural environment.

Keywords:
Computer science Computer vision Artificial intelligence Robot Position (finance) Track (disk drive) Sound (geography) Acoustic source localization Matching (statistics) Speech recognition Acoustics Mathematics

Metrics

0
Cited By
0.00
FWCI (Field Weighted Citation Impact)
0
Refs
0.03
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Topics

Speech and Audio Processing
Physical Sciences →  Computer Science →  Signal Processing
Video Surveillance and Tracking Methods
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Video Analysis and Summarization
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
© 2026 ScienceGate Book Chapters — All rights reserved.