DocumentCode
416669
Title
Acquisition of box pushing by direct-vision-based reinforcement learning
Author
Shibata, Katsunari ; Iida, Masaru
Author_Institution
Dept. of Electr. & Electron. Eng., Oita Univ., Japan
Volume
3
fYear
2003
fDate
4-6 Aug. 2003
Firstpage
2322
Abstract
In this paper, it was confirmed that a real mobile robot with a CCD camera could learn appropriate actions to reach and push a lying box only by direct-vision-based reinforcement learning (RL). In direct-vision-based RL raw visual sensor signals are the inputs of a layered neural network; the neural network is trained by backpropagation using the training signal that is generated based on reinforcement learning. In other words, no image processing, no control methods, and no task information are given at premise even if as many as 1536 monochrome visual signals and 4 infrared signals are the inputs. The box-pushing task is rather difficult than reaching task for the reason that not only the center of gravity, but also the direction, weight and sliding character of the box should be considered. Nevertheless, the robot could learn appropriate actions even if the reward was given only when the robot was pushing the box. It was also observed that the neural network obtained global representation of the box location through the learning.
Keywords
learning (artificial intelligence); mobile robots; neural nets; robot vision; CCD camera; backpropagation; box pushing acquisition; direct-vision-based reinforcement learning; layered neural network; real mobile robot;
fLanguage
English
Publisher
ieee
Conference_Titel
SICE 2003 Annual Conference
Conference_Location
Fukui, Japan
Print_ISBN
0-7803-8352-4
Type
conf
Filename
1323606
Link To Document