DocumentCode
2437185
Title
Relative Speech Emotion Recognition Based Artificial Neural Network
Author
Fu, Liqin ; Mao, Xia ; Chen, Lijiang
Author_Institution
Sch. of Electron. & Inf. Eng., Beihang Univ., Beijing
Volume
2
fYear
2008
fDate
19-20 Dec. 2008
Firstpage
140
Lastpage
144
Abstract
Artificial neural network (ANN) models based on static features vector as well as normalized temporal features vector, were used to recognize emotion state from speech. Moreover, relative features obtained by computing the changes of acoustic features of emotional speech relative to those of neutral speech were adopted to weaken the influence from the individual difference. The methods to relativize static features and temporal features were introduced individually and experiments based Germany database and Mandarin database were implemented. The results show that the performance of relative features excels that of absolute features for emotion recognition as a whole. When speaker is independent, the hybrid of relative static features vector and relative temporal features normalized vector achieves the best results.
Keywords
emotion recognition; feature extraction; neural nets; speech recognition; artificial neural network; normalized temporal features vector; speech emotion recognition; static features vector; Acoustic distortion; Artificial neural networks; Automatic speech recognition; Emotion recognition; Hidden Markov models; Humans; Loudspeakers; Shape; Spatial databases; Speech recognition; ANN; relative features; speech emotion recognition;
fLanguage
English
Publisher
ieee
Conference_Titel
Computational Intelligence and Industrial Application, 2008. PACIIA '08. Pacific-Asia Workshop on
Conference_Location
Wuhan
Print_ISBN
978-0-7695-3490-9
Type
conf
DOI
10.1109/PACIIA.2008.355
Filename
4756752
Link To Document