DocumentCode
422677
Title
Visual perception for a partner robot based on computational intelligence
Author
Kubota, Naoyuki
Volume
1
fYear
2004
fDate
25-29 July 2004
Firstpage
293
Abstract
This paper proposes a method for visual perception for a partner robot interacting with a human. A robot with a physical body should extract information by using prediction based on the dynamics of its environment, because the computational cost can be reduced, imitation is a powerful tool for gestural interaction between children and for teaching behaviors to children by parent. Furthermore, others´ action can be a hint for obtaining a new behavior that might not be the same as the original action. This paper proposes a visual perception method for a partner robot based on the interactive teaching mechanism of a human teacher. The proposed method is composed of a spiking neural network, a self-organizing map, a steady-state genetic algorithm, and softmax action selection strategy. Furthermore, we discuss the interactive learning of a human and a partner robot based on the proposed method through several experiment results.
Keywords
computer aided instruction; human computer interaction; interactive systems; robot vision; service robots; computational intelligence; gestural interaction; interactive teaching mechanism; partner robot; self-organizing map; spiking neural network; steady-state genetic algorithm; visual perception; Computational efficiency; Computational intelligence; Data mining; Education; Educational robots; Human robot interaction; Intelligent robots; Neural networks; Steady-state; Visual perception;
fLanguage
English
Publisher
ieee
Conference_Titel
Fuzzy Systems, 2004. Proceedings. 2004 IEEE International Conference on
ISSN
1098-7584
Print_ISBN
0-7803-8353-2
Type
conf
DOI
10.1109/FUZZY.2004.1375737
Filename
1375737
Link To Document