DocumentCode :
3427269
Title :
Audio-based unsupervised segmentation of multiparty dialogue
Author :
Hsueh, Pei-Yun
Author_Institution :
Sch. of Inf., Edinburgh Univ., Edinburgh
fYear :
2008
fDate :
March 31 2008-April 4 2008
Firstpage :
5049
Lastpage :
5052
Abstract :
In this paper, we explore a novel way to leverage audio information for unsupervised segmentation of multiparty dialogue. Our system which segments directly on patterns derived from audio sources is evaluated with previous work that segments on lexical patterns found in transcripts. We examine the effectiveness of both systems on recovering a two-layer structure of meeting dialogue. We demonstrate that the audio-based system performs significantly better than the word-based system on this task. In particular, it effectively recover segments of off-topic discussion. Results are encouraging as the audio information used in the system can be obtained in near real time and with absence of manual and ASR transcripts. It is particularly desirable when a system has to be operated online, or in unfamiliar domains and languages.
Keywords :
audio signal processing; interactive systems; speech recognition; audio sources; audio-based unsupervised segmentation; automatic speech recognition; leverage audio information; multiparty dialogue; word-based system; Acoustic signal processing; Audio recording; Automatic speech recognition; Availability; Clustering methods; Informatics; Natural languages; Real time systems; Speech recognition; Vocabulary; acoustic signal processing; clustering methods; meetings;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Acoustics, Speech and Signal Processing, 2008. ICASSP 2008. IEEE International Conference on
Conference_Location :
Las Vegas, NV
ISSN :
1520-6149
Print_ISBN :
978-1-4244-1483-3
Electronic_ISBN :
1520-6149
Type :
conf
DOI :
10.1109/ICASSP.2008.4518793
Filename :
4518793
Link To Document :
بازگشت