DocumentCode
613808
Title
Poster: Real time hand pose recognition with depth sensors for mixed reality interfaces
Author
Byungkyu Kang ; Rodrigue, Mathieu ; Hollerer, Tobias ; Hwasup Lim
Author_Institution
Univ. of California Santa Barbara, Santa Barbara, CA, USA
fYear
2013
fDate
16-17 March 2013
Firstpage
171
Lastpage
172
Abstract
We present a method for predicting articulated hand poses in real-time with a single depth camera, such as the Kinect or Xtion Pro, for the purpose of interaction in a Mixed Reality environment and for studying the effects of realistic and non-realistic articulated hand models in a Mixed Reality simulator. We demonstrate that employing a randomized decision forest for hand recognition benefits real-time applications without the typical tracking pitfalls such as reinitialization. This object recognition approach to predict hand poses results in relatively low computation, high prediction accuracy and sets the groundwork needed to utilize articulated hand movements for 3D tasks in Mixed Reality workspaces.
Keywords
decision trees; object recognition; pose estimation; virtual reality; 3D task; Kinect; Xtion Pro; articulated hand pose prediction; depth sensor; hand recognition; mixed reality environment; mixed reality interface; mixed reality simulator; mixed reality workspace; nonrealistic articulated hand model; object recognition; randomized decision forest; real time hand pose recognition; reinitialization; single depth camera; tracking pitfall; Computational modeling; Real-time systems; Sensors; Three-dimensional displays; Training; User interfaces; Virtual reality;
fLanguage
English
Publisher
ieee
Conference_Titel
3D User Interfaces (3DUI), 2013 IEEE Symposium on
Conference_Location
Orlando, FL
Print_ISBN
978-1-4673-6097-5
Type
conf
DOI
10.1109/3DUI.2013.6550233
Filename
6550233
Link To Document