Title : 
Minimax-based reinforcement learning with state aggregation
         
        
            Author : 
Jiang, Guofei ; Wu, Cang-Pu ; Cybenko, George
         
        
            Author_Institution : 
Dept. of Autom. Control, Beijing Inst. of Technol., China
         
        
        
        
        
        
            Abstract : 
One of the most important issues in scaling up reinforcement learning for practical problems is how to represent and store cost-to-go functions with more compact representations than lookup tables. We address the issue of combining the simple function approximation method-state aggregation with minimax-based reinforcement learning algorithms and present the convergence theory for online Q-hat-learning with state aggregation. Some empirical results are also included
         
        
            Keywords : 
Markov processes; aggregation; convergence; decision theory; function approximation; learning (artificial intelligence); convergence theory; cost-to-go functions; function approximation method; minimax-based reinforcement learning; online Q-hat-learning; state aggregation; Approximation algorithms; Convergence; Costs; Dynamic programming; Educational institutions; Function approximation; Learning; Minimax techniques; Stochastic processes; Table lookup;
         
        
        
        
            Conference_Titel : 
Decision and Control, 1998. Proceedings of the 37th IEEE Conference on
         
        
            Conference_Location : 
Tampa, FL
         
        
        
            Print_ISBN : 
0-7803-4394-8
         
        
        
            DOI : 
10.1109/CDC.1998.758445