• DocumentCode
    2789753
  • Title

    Learning with synthesized speech for automatic emotion recognition

  • Author

    Schuller, Björn ; Burkhardt, Felix

  • Author_Institution
    Inst. for Human-Machine Commun., Tech. Univ. Munchen, München, Germany
  • fYear
    2010
  • fDate
    14-19 March 2010
  • Firstpage
    5150
  • Lastpage
    5153
  • Abstract
    Data sparseness is an ever dominating problem in automatic emotion recognition. Using artificially generated speech for training or adapting models could potentially ease this: though less natural than human speech, one could synthesize the exact spoken content in different emotional nuances - of many speakers and even in different languages. To investigate chances, the phonemisation components Txt2Pho and openMary are used with Emofilt and Mbrola for emotional speech synthesis. Analysis is realized with our Munich open Emotion and Affect Recognition toolkit. As test set we gently limit to the acted Berlin and eNTERFACE databases for the moment. In the result synthesized speech can indeed be used for the recognition of human emotional speech.
  • Keywords
    emotion recognition; speech processing; speech synthesis; affect recognition toolkit; automatic emotion recognition; data sparseness; eNTERFACE database; emotiona lspeech synthesis; human emotional speech recognition; munich open emotion; speech synthesis; Automatic speech recognition; Databases; Emotion recognition; Humans; Natural language processing; Natural languages; Speech analysis; Speech processing; Speech recognition; Speech synthesis; Affective Computing; Emotion Recognition; Speech Analysis; Speech Synthesis;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Acoustics Speech and Signal Processing (ICASSP), 2010 IEEE International Conference on
  • Conference_Location
    Dallas, TX
  • ISSN
    1520-6149
  • Print_ISBN
    978-1-4244-4295-9
  • Electronic_ISBN
    1520-6149
  • Type

    conf

  • DOI
    10.1109/ICASSP.2010.5495017
  • Filename
    5495017