DocumentCode
270748
Title
Improving robot vision models for object detection through interaction
Author
Leitner, Jürgen ; Förster, Alexander ; Schmidhuber, Jürgen
Author_Institution
Dalle Molle Inst. for Artificial Intell., Lugano, Switzerland
fYear
2014
fDate
6-11 July 2014
Firstpage
3355
Lastpage
3362
Abstract
We propose a method for learning specific object representations that can be applied (and reused) in visual detection and identification tasks. A machine learning technique called Cartesian Genetic Programming (CGP) is used to create these models based on a series of images. Our research investigates how manipulation actions might allow for the development of better visual models and therefore better robot vision. This paper describes how visual object representations can be learned and improved by performing object manipulation actions, such as, poke, push and pick-up with a humanoid robot. The improvement can be measured and allows for the robot to select and perform the `right´ action, i.e. the action with the best possible improvement of the detector.
Keywords
genetic algorithms; humanoid robots; image representation; learning (artificial intelligence); object detection; robot vision; CGP; Cartesian genetic programming; humanoid robot; machine learning technique; object detection; object manipulation actions; robot vision model; visual detection tasks; visual identification tasks; visual object representations; Detectors; Image segmentation; Robot kinematics; Sociology; Statistics; Visualization;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks (IJCNN), 2014 International Joint Conference on
Conference_Location
Beijing
Print_ISBN
978-1-4799-6627-1
Type
conf
DOI
10.1109/IJCNN.2014.6889556
Filename
6889556
Link To Document