DocumentCode :
1669325
Title :
Audio-visual data fusion for tracking the direction of multiple speakers
Author :
Nguyen, Quang ; Choi, JongSuk
Author_Institution :
Center for Cognitive Robot. Res., Korea Inst. of Sci. & Technol., Seoul, South Korea
fYear :
2010
Firstpage :
1626
Lastpage :
1630
Abstract :
This paper presents a multi-speakers tracking algorithm using audio-visual data fusion. The audio information is the direction of speakers and the visual information is the direction of detected faces. These observations are used as inputs of the tracking algorithm, which employed the framework of particle filter. For multi-target tracking, we present a flexible data association and data fusion, which can deal with the appearance or absent of any information during tracking process. The experimental results on data collected from a robot platform in a conventional office room confirm a potential application for human-robot interaction.
Keywords :
face recognition; particle filtering (numerical methods); sensor fusion; speaker recognition; target tracking; audio information; audio-visual data fusion; data association; detected face; multiple speaker tracking; multitarget tracking; particle filter; robot platform; speaker direction; Arrays; Cameras; Face detection; Microphones; Particle filters; Target tracking; Visualization; audio-visual data fusion; particle filter; sound source localization; speaker tracking;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Control Automation and Systems (ICCAS), 2010 International Conference on
Conference_Location :
Gyeonggi-do
Print_ISBN :
978-1-4244-7453-0
Electronic_ISBN :
978-89-93215-02-1
Type :
conf
Filename :
5669639
Link To Document :
بازگشت