DocumentCode :
19448
Title :
Latent Semantic Analysis for Multimodal User Input With Speech and Gestures
Author :
Pui-Yu Hui ; Meng, Hsiang-Yun
Author_Institution :
Human-Comput. Commun. Lab., Chinese Univ. of Hong Kong, Hong Kong, China
Volume :
22
Issue :
2
fYear :
2014
fDate :
Feb. 2014
Firstpage :
417
Lastpage :
429
Abstract :
This paper describes our work in semantic interpretation of a “multimodal language” with speech and gestures using latent semantic analysis (LSA). Our aim is to infer the domain-specific informational goal of multimodal inputs. The informational goal is characterized by lexical terms used in the spoken modality, partial semantics of gestures in the pen modality, as well as term co-occurrence patterns across modalities, leading to “multimodal terms.” We designed and collected a multimodal corpus of navigational inquiries. We also obtained perfect (i.e. manual) and imperfect (i.e. automatic via recognition) transcriptions for these. We automatically align parsed spoken locative references (SLRs) with their corresponding pen gesture(s) using the Viterbi alignment, according to their numeric and location type features. Then, we characterize each cross-modal integration pattern as a 3-tuple multimodal term with SLR, pen gesture type and their temporal relationship. We propose to use latent semantic analysis (LSA) to derive the latent semantics from manual (i.e. perfect) and automatic (i.e. imperfect) transcriptions of the collected multimodal inputs. In order to achieve this, both multimodal and lexical terms are used to compose an inquiry-term matrix, which is then factorized using singular value decomposition (SVD) to derive the latent semantics automatically. Informational goal inference based on the latent semantics shows that the informational goal inference accuracy of a disjoint test set is 99% and 84% when a perfect and imperfect projection model is used respectively, which performs significantly better than (at least 9.9% absolute) the baseline performance using vector-space model (VSM).
Keywords :
gesture recognition; singular value decomposition; speech recognition; user interfaces; 3-tuple multimodal term; LSA; SLR; SVD; VSM; Viterbi alignment; automatic transcription; collected multimodal inputs; cross-modal integration pattern; domain-specific informational goal; imperfect projection model; inquiry-term matrix; latent semantic analysis; lexical terms; manual transcription; multimodal corpus; multimodal language; multimodal terms; multimodal user input; navigational inquiries; parsed spoken locative references; partial gesture semantics; pen modality; perfect projection model; semantic interpretation; singular value decomposition; speech; spoken modality; temporal relationship; term co-occurrence patterns; Educational institutions; Manuals; Semantics; Speech; Speech processing; Speech recognition; Training; Multimodal user interfaces; gesture recognition; latent semantic analysis; speech recognition;
fLanguage :
English
Journal_Title :
Audio, Speech, and Language Processing, IEEE/ACM Transactions on
Publisher :
ieee
ISSN :
2329-9290
Type :
jour
DOI :
10.1109/TASLP.2013.2294586
Filename :
6680708
Link To Document :
بازگشت