• DocumentCode
    2174882
  • Title

    An approach to sequential grouping in cochannel speech

  • Author

    Hu, Ke ; DeLiang Wang

  • Author_Institution
    Dept. of Comput. Sci. & Eng., Ohio State Univ., Columbus, OH, USA
  • fYear
    2011
  • fDate
    22-27 May 2011
  • Firstpage
    4636
  • Lastpage
    4639
  • Abstract
    Model-based methods for sequential organization in cochannel speech require pretrained speaker models and often prior knowledge of participating speakers. We propose an unsupervised approach to sequential organization of cochannel speech. Based on cepstral features, we first cluster voiced speech into two speaker groups by maximizing the ratio of between- and within-group distances penalized by within-group concurrent pitches. To group unvoiced speech, we employ an onset/offset based analysis to generate time-frequency segments. Unvoiced segments are then labeled by the complementary portions of segregated voiced speech. Our method does not require any pretrained model and is computationally simple. Evaluations and comparisons show that the proposed method outperforms a model-based method in terms of speech segregation.
  • Keywords
    speech processing; cochannel speech; sequential grouping; sequential organization; time-frequency segments; voiced speech clustering; voiced speech segregation; Computational modeling; Hidden Markov models; Organizations; Signal to noise ratio; Speech; Speech recognition; Time frequency analysis; clustering; cochannel speech separation; sequential grouping; unvoiced speech;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Acoustics, Speech and Signal Processing (ICASSP), 2011 IEEE International Conference on
  • Conference_Location
    Prague
  • ISSN
    1520-6149
  • Print_ISBN
    978-1-4577-0538-0
  • Electronic_ISBN
    1520-6149
  • Type

    conf

  • DOI
    10.1109/ICASSP.2011.5947388
  • Filename
    5947388