• DocumentCode
    3428632
  • Title

    Learning to Predict Gaze in Egocentric Video

  • Author

    Yin Li ; Fathi, Alahoum ; Rehg, James M.

  • fYear
    2013
  • fDate
    1-8 Dec. 2013
  • Firstpage
    3216
  • Lastpage
    3223
  • Abstract
    We present a model for gaze prediction in egocentric video by leveraging the implicit cues that exist in camera wearer´s behaviors. Specifically, we compute the camera wearer´s head motion and hand location from the video and combine them to estimate where the eyes look. We further model the dynamic behavior of the gaze, in particular fixations, as latent variables to improve the gaze prediction. Our gaze prediction results outperform the state-of-the-art algorithms by a large margin on publicly available egocentric vision datasets. In addition, we demonstrate that we get a significant performance boost in recognizing daily actions and segmenting foreground objects by plugging in our gaze predictions into state-of-the-art methods.
  • Keywords
    gaze tracking; gesture recognition; image segmentation; video signal processing; camera wearer behaviors; camera wearer hand location; camera wearer head motion; daily action recognition; egocentric video; egocentric vision datasets; fixations; foreground object segmentation; gaze dynamic behavior; gaze prediction; implicit cues; latent variables; Computational modeling; Feature extraction; Graphical models; Head; Hidden Markov models; Predictive models; Vectors; Action Recognition; Egocentric Vision; Gaze Prediction; Object Segmentation;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Computer Vision (ICCV), 2013 IEEE International Conference on
  • Conference_Location
    Sydney, NSW
  • ISSN
    1550-5499
  • Type

    conf

  • DOI
    10.1109/ICCV.2013.399
  • Filename
    6751511