DocumentCode :
270748
Title :
Improving robot vision models for object detection through interaction
Author :
Leitner, Jürgen ; Förster, Alexander ; Schmidhuber, Jürgen
Author_Institution :
Dalle Molle Inst. for Artificial Intell., Lugano, Switzerland
fYear :
2014
fDate :
6-11 July 2014
Firstpage :
3355
Lastpage :
3362
Abstract :
We propose a method for learning specific object representations that can be applied (and reused) in visual detection and identification tasks. A machine learning technique called Cartesian Genetic Programming (CGP) is used to create these models based on a series of images. Our research investigates how manipulation actions might allow for the development of better visual models and therefore better robot vision. This paper describes how visual object representations can be learned and improved by performing object manipulation actions, such as, poke, push and pick-up with a humanoid robot. The improvement can be measured and allows for the robot to select and perform the `right´ action, i.e. the action with the best possible improvement of the detector.
Keywords :
genetic algorithms; humanoid robots; image representation; learning (artificial intelligence); object detection; robot vision; CGP; Cartesian genetic programming; humanoid robot; machine learning technique; object detection; object manipulation actions; robot vision model; visual detection tasks; visual identification tasks; visual object representations; Detectors; Image segmentation; Robot kinematics; Sociology; Statistics; Visualization;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks (IJCNN), 2014 International Joint Conference on
Conference_Location :
Beijing
Print_ISBN :
978-1-4799-6627-1
Type :
conf
DOI :
10.1109/IJCNN.2014.6889556
Filename :
6889556
Link To Document :
بازگشت