DocumentCode
2174882
Title
An approach to sequential grouping in cochannel speech
Author
Hu, Ke ; DeLiang Wang
Author_Institution
Dept. of Comput. Sci. & Eng., Ohio State Univ., Columbus, OH, USA
fYear
2011
fDate
22-27 May 2011
Firstpage
4636
Lastpage
4639
Abstract
Model-based methods for sequential organization in cochannel speech require pretrained speaker models and often prior knowledge of participating speakers. We propose an unsupervised approach to sequential organization of cochannel speech. Based on cepstral features, we first cluster voiced speech into two speaker groups by maximizing the ratio of between- and within-group distances penalized by within-group concurrent pitches. To group unvoiced speech, we employ an onset/offset based analysis to generate time-frequency segments. Unvoiced segments are then labeled by the complementary portions of segregated voiced speech. Our method does not require any pretrained model and is computationally simple. Evaluations and comparisons show that the proposed method outperforms a model-based method in terms of speech segregation.
Keywords
speech processing; cochannel speech; sequential grouping; sequential organization; time-frequency segments; voiced speech clustering; voiced speech segregation; Computational modeling; Hidden Markov models; Organizations; Signal to noise ratio; Speech; Speech recognition; Time frequency analysis; clustering; cochannel speech separation; sequential grouping; unvoiced speech;
fLanguage
English
Publisher
ieee
Conference_Titel
Acoustics, Speech and Signal Processing (ICASSP), 2011 IEEE International Conference on
Conference_Location
Prague
ISSN
1520-6149
Print_ISBN
978-1-4577-0538-0
Electronic_ISBN
1520-6149
Type
conf
DOI
10.1109/ICASSP.2011.5947388
Filename
5947388
Link To Document