Title :
Audio-visual data fusion for tracking the direction of multiple speakers
Author :
Nguyen, Quang ; Choi, JongSuk
Author_Institution :
Center for Cognitive Robot. Res., Korea Inst. of Sci. & Technol., Seoul, South Korea
Abstract :
This paper presents a multi-speakers tracking algorithm using audio-visual data fusion. The audio information is the direction of speakers and the visual information is the direction of detected faces. These observations are used as inputs of the tracking algorithm, which employed the framework of particle filter. For multi-target tracking, we present a flexible data association and data fusion, which can deal with the appearance or absent of any information during tracking process. The experimental results on data collected from a robot platform in a conventional office room confirm a potential application for human-robot interaction.
Keywords :
face recognition; particle filtering (numerical methods); sensor fusion; speaker recognition; target tracking; audio information; audio-visual data fusion; data association; detected face; multiple speaker tracking; multitarget tracking; particle filter; robot platform; speaker direction; Arrays; Cameras; Face detection; Microphones; Particle filters; Target tracking; Visualization; audio-visual data fusion; particle filter; sound source localization; speaker tracking;
Conference_Titel :
Control Automation and Systems (ICCAS), 2010 International Conference on
Conference_Location :
Gyeonggi-do
Print_ISBN :
978-1-4244-7453-0
Electronic_ISBN :
978-89-93215-02-1