DocumentCode :
2652486
Title :
Emotion recognition in speech using inter-sentence Glottal statistics
Author :
Iliev, Alexander I. ; Scordilis, Michael S.
Author_Institution :
Dept. of Electr. & Comput. Eng., Univ. of Miami Coral Gables, Coral Gables, FL
fYear :
2008
fDate :
25-28 June 2008
Firstpage :
465
Lastpage :
468
Abstract :
This study deals with the recognition of three emotional states in speech, namely: happiness, anger, and sadness. The corpus included speech from six subjects (3M and 3F) speaking ten sentences. Glottal inverse filtering was first performed on the spoken utterances. Then parameters for computing the glottal symmetry were collected and computed to create a final matrix of features. A combined with all emotions across the different subjects was formed and used to train a Gaussian mixture model (GMM) classifier. Training on 80% of all combined utterances for each emotion was performed. Testing was administered on the remaining 20%. The system shows confidence that glottal information may be used for determining the correct emotion in speech. The recognition performance varied between 48.96% and 82.29%.
Keywords :
Gaussian processes; emotion recognition; pattern classification; speech recognition; Gaussian mixture model classifier; emotion recognition; glottal information; glottal inverse filtering; inter-sentence glottal statistics; speech recognition; spoken utterances; Audio recording; Data mining; Emotion recognition; Focusing; Frequency; Lips; Microphones; Pulse modulation; Speech recognition; Statistics; Emotion Recognition; GMM; Glottal Symmetry; Glottal waveform; Pattern classification; Speech;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Systems, Signals and Image Processing, 2008. IWSSIP 2008. 15th International Conference on
Conference_Location :
Bratislava
Print_ISBN :
978-80-227-2856-0
Electronic_ISBN :
978-80-227-2880-5
Type :
conf
DOI :
10.1109/IWSSIP.2008.4604467
Filename :
4604467
Link To Document :
بازگشت