• DocumentCode
    1349947
  • Title

    Toward multimodal human-computer interface

  • Author

    Sharma, Rejeev ; Pavlovic, Vladimir I. ; Huang, Thomas S.

  • Author_Institution
    Dept. of Comput. Sci. & Eng., Pennsylvania State Univ., University Park, PA, USA
  • Volume
    86
  • Issue
    5
  • fYear
    1998
  • fDate
    5/1/1998 12:00:00 AM
  • Firstpage
    853
  • Lastpage
    869
  • Abstract
    Recent advances in various signal processing technologies, coupled with an explosion in the available computing power, have given rise to a number of novel human-computer interaction (HCI) modalities: speech, vision-based gesture recognition, eye tracking, electroencephalograph, etc. Successful embodiment of these modalities into an interface has the potential of easing the HCI bottleneck that has become noticeable with the advances in computing and communication. It has also become increasingly evident that the difficulties encountered in the analysis and interpretation of individual sensing modalities may be overcome by integrating them into a multimodal human-computer interface. We examine several promising directions toward achieving multimodal HCI. We consider some of the emerging novel input modalities for HCI and the fundamental issues in integrating them at various levels, from early signal level to intermediate feature level to late decision level. We discuss the different computational approaches that may be applied at the different levels of modality integration. We also briefly review several demonstrated multimodal HCI systems and applications. Despite all the recent developments, it is clear that further research is needed for interpreting and fitting multiple sensing modalities in the context of HCI. This research can benefit from many disparate fields of study that increase our understanding of the different human communication modalities and their potential role in HCI
  • Keywords
    human factors; image recognition; speech recognition; user interfaces; HCI bottleneck; computational approaches; electroencephalograph; eye tracking; gesture recognition; input modalities; modality integration; multimodal human-computer interface; research; signal processing; speech; Artificial neural networks; Automatic speech recognition; Computer displays; Computer interfaces; Electroencephalography; Hidden Markov models; Human computer interaction; Maximum likelihood estimation; Personal digital assistants; Virtual reality;
  • fLanguage
    English
  • Journal_Title
    Proceedings of the IEEE
  • Publisher
    ieee
  • ISSN
    0018-9219
  • Type

    jour

  • DOI
    10.1109/5.664275
  • Filename
    664275