DocumentCode :
1149967
Title :
Sensitive Talking Heads [Applications Corner]
Author :
Huang, Thomas S. ; Hasegawa-Johnson, Mark A. ; Chu, Stephen M. ; Zeng, Zhihong ; Tang, Hao
Volume :
26
Issue :
4
fYear :
2009
fDate :
7/1/2009 12:00:00 AM
Firstpage :
67
Lastpage :
72
Abstract :
Spoken language user interfaces can dramatically speed up computer use. Unfortunately, if the speech user interface interferes too often, the user turns it off. Users are unforgiving: a technology that impairs productivity just once may never get a second chance. To give the user interface a fighting chance, why not endow it with a certain amount of emotional sensitivity? Users respond better to an avatar that displays appropriate emotional nuance; conversely, if the avatar detects extreme frustration on the part of the user, it can hide in the corner of the monitor until the frustration has passed. A hidden avatar is still present and can continue to be of service to the user upon request. This paper describes experiments in emotive spoken language user interfaces. We find that both recognition accuracy and synthesis quality are improved when one takes advantage of multimodal, synthesizing, and recognizing information in both the audio and video modalities.
Keywords :
avatars; emotion recognition; speech recognition; speech synthesis; speech-based user interfaces; audio modality; avatar; emotional nuance; emotional sensitivity; emotive spoken language user interface; sensitive talking heads; speech recognition; speech synthesis quality; speech user interface; user frustration; video modality; Automatic speech recognition; Avatars; Humans; Lips; Mouth; Shape measurement; Signal synthesis; Speech recognition; Speech synthesis; User interfaces;
fLanguage :
English
Journal_Title :
Signal Processing Magazine, IEEE
Publisher :
ieee
ISSN :
1053-5888
Type :
jour
DOI :
10.1109/MSP.2009.932562
Filename :
5174498
Link To Document :
بازگشت