Title :
Modality tracking in the multimodal Bell Labs Communicator
Author :
Potamianos, Alexandros ; Ammicht, Egbert ; Fosler-Lussie, Eric
Author_Institution :
Dept. of Elec. & Comp. Eng., Tech. Univ. Crete, Chania, Greece
fDate :
30 Nov.-3 Dec. 2003
Abstract :
We describe our efforts in designing and building the multimodal Bell Labs Communicator system. Several innovations are necessary for generalizing the speech-only user interface and the semantic/pragmatic algorithms to handle both speech and visual modalities, including changes to the parser, semantic and pragmatic modules to integrate better the user input from different modalities within and across dialogue turns, as well as extending the concept of e-forms to include visual modality. The paper also describes the design of a natural, consistent and efficient multimodal user interface. We introduce the concept of "dominant modality": the system adaptively tracks the most probable interaction mode and suggests that modality to the user. Modality tracking provides an elegant solution to the "turn-taking" issue in multimodal spoken dialogue systems.
Keywords :
graphical user interfaces; human computer interaction; interactive systems; natural language interfaces; speech recognition; speech synthesis; speech-based user interfaces; GUI; dominant modality; human-machine interaction; modality tracking; multimodal Bell Labs Communicator system; multimodal spoken dialogue systems; multimodal user interface; pragmatic algorithm; semantic algorithm; speech recognizer; speech-only user interface; text-to-speech system; turn-taking; visual modality; Information science; Keyboards; Natural languages; Navigation; Ontologies; Prototypes; Speech; Technological innovation; User interfaces; Visual databases;
Conference_Titel :
Automatic Speech Recognition and Understanding, 2003. ASRU '03. 2003 IEEE Workshop on
Print_ISBN :
0-7803-7980-2
DOI :
10.1109/ASRU.2003.1318427