DocumentCode
3028508
Title
Detection of divided planar object for augmented reality applications
Author
Nishizaka, Shinya ; Narumi, Takuji ; Tanikawa, Tomohiro ; Hirose, Michitaka
Author_Institution
Graduate School of Information Science and Technology, The University of Tokyo, Japan
fYear
2011
fDate
19-23 March 2011
Firstpage
231
Lastpage
232
Abstract
In this research study, we propose a divided planar-object detection method for augmented reality(AR) applications. There are mainly two types of camera-registration methods for AR applications: marker-based methods, and natural-feature-based methods. In addition, the latter methods are classified into visual SLAM and object detection methods. With respect to object detection methods, particularly for planar objects such as paper, methods for dealing with bending, folding, and occlusion are proposed. However, the division of objects has not been studied. Once an object is divided, a conventional object detection method cannot identify each of the pieces because the feature points of only a single piece are recognized as the target object, and the other feature points are regarded as outliers. The proposed system prepares a database of the target object´s natural features, and applies progressive sample consensus(PROSAC), which is a robust estimation method, for iterative homography calculation to achieve the multiple planar-object detection. Moreover, the proposed method can detect shapes of pieces by simultaneously using an occlusion detection method. We demonstrate that it is possible to interact with an arbitrarily divided planar object in real time by our method to implement some AR applications.
Keywords
Databases; Estimation; Feature extraction; Object detection; Real time systems; Robustness; Target tracking;
fLanguage
English
Publisher
ieee
Conference_Titel
Virtual Reality Conference (VR), 2011 IEEE
Conference_Location
Singapore, Singapore
ISSN
1087-8270
Print_ISBN
978-1-4577-0039-2
Electronic_ISBN
1087-8270
Type
conf
DOI
10.1109/VR.2011.5759483
Filename
5759483
Link To Document