Title :
Minimax-based reinforcement learning with state aggregation
Author :
Jiang, Guofei ; Wu, Cang-Pu ; Cybenko, George
Author_Institution :
Dept. of Autom. Control, Beijing Inst. of Technol., China
Abstract :
One of the most important issues in scaling up reinforcement learning for practical problems is how to represent and store cost-to-go functions with more compact representations than lookup tables. We address the issue of combining the simple function approximation method-state aggregation with minimax-based reinforcement learning algorithms and present the convergence theory for online Q-hat-learning with state aggregation. Some empirical results are also included
Keywords :
Markov processes; aggregation; convergence; decision theory; function approximation; learning (artificial intelligence); convergence theory; cost-to-go functions; function approximation method; minimax-based reinforcement learning; online Q-hat-learning; state aggregation; Approximation algorithms; Convergence; Costs; Dynamic programming; Educational institutions; Function approximation; Learning; Minimax techniques; Stochastic processes; Table lookup;
Conference_Titel :
Decision and Control, 1998. Proceedings of the 37th IEEE Conference on
Conference_Location :
Tampa, FL
Print_ISBN :
0-7803-4394-8
DOI :
10.1109/CDC.1998.758445