Title :
Audio scene segmentation using multiple features, models and time scales
Author :
Sundaram, Hari ; Chang, Shih-Fu
Author_Institution :
Dept. of Electr. Eng., Columbia Univ., New York, NY, USA
Abstract :
We present an algorithm for audio scene segmentation. An audio scene is a semantically consistent sound segment that is characterized by a few dominant sources of sound. A scene change occurs when a majority of the sources present in the data change. Our segmentation framework has three parts: a definition of an audio scene; multiple feature models that characterize the dominant sources; and a simple, causal listener model, which mimics human audition using multiple time-scales. We define a correlation function that determines correlation with past data to determine segmentation boundaries. The algorithm was tested on a difficult data set, a 1 hour audio segment of a film, with impressive results. It achieves an audio scene change detection accuracy of 97%
Keywords :
audio signal processing; hearing; multimedia systems; audio scene change detection; audio scene definition; audio scene segmentation; causal listener model; correlation function; data set; human audition; multimedia; multiple feature models; multiple time-scales; Data mining; Detection algorithms; Feature extraction; Layout; Music; Organizing; Speech; Testing;
Conference_Titel :
Acoustics, Speech, and Signal Processing, 2000. ICASSP '00. Proceedings. 2000 IEEE International Conference on
Conference_Location :
Istanbul
Print_ISBN :
0-7803-6293-4
DOI :
10.1109/ICASSP.2000.859335