DocumentCode :
1889030
Title :
Multitask reinforcement learning on the distribution of MDPs
Author :
Tanaka, Fumihide ; Yamamura, Masayuki
Author_Institution :
Dept. of Comput. Intelligence & Syst. Sci., Tokyo Inst. of Technol., Japan
Volume :
3
fYear :
2003
fDate :
16-20 July 2003
Firstpage :
1108
Abstract :
In this paper we address a new problem in reinforcement learning. Here we consider an agent that faces multiple learning tasks within its lifetime. The agent´s objective is to maximize its total reward in the lifetime as well as a conventional return in each task. To realize this, it has to be endowed an important ability to keep its past learning experiences and utilize them for improving future learning performance. This time we try to phrase this problem formally. The central idea is to introduce an environmental class, BV-MDPs that is defined with the distribution of MDPs. As an approach to exploiting past learning experiences, we focus on statistics (mean and deviation) about the agent´s value tables. The mean can be used as initial values of the table when a new task is presented. The deviation can be viewed as measuring reliability of the mean, and we utilize it in calculating priority of simulated backups. We conduct experiments in computer simulation to evaluate the effectiveness.
Keywords :
Markov processes; decision theory; learning (artificial intelligence); Markov decision processes; multiple learning tasks; multitask reinforcement learning; past learning experiences; reliability measurement; Cats; Cleaning; Computational intelligence; Computational modeling; Computer errors; Computer simulation; Dogs; Learning; Robots; Uncertainty;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Computational Intelligence in Robotics and Automation, 2003. Proceedings. 2003 IEEE International Symposium on
Print_ISBN :
0-7803-7866-0
Type :
conf
DOI :
10.1109/CIRA.2003.1222152
Filename :
1222152
Link To Document :
بازگشت