DocumentCode :
2708813
Title :
Using continuous action spaces to solve discrete problems
Author :
Van Hasselt, Hado ; Wiering, Marco A.
Author_Institution :
Intell. Syst. Group, Utrecht Univ., Utrecht, Netherlands
fYear :
2009
fDate :
14-19 June 2009
Firstpage :
1149
Lastpage :
1156
Abstract :
Real-world control problems are often modeled as Markov decision processes (MDPs) with discrete action spaces to facilitate the use of the many reinforcement learning algorithms that exist to find solutions for such MDPs. For many of these problems an underlying continuous action space can be assumed. We investigate the performance of the Cacla algorithm, which uses a continuous actor, on two such MDPs: the mountain car and the cart pole. We show that Cacla has clear advantages over discrete algorithms such as Q-learning and Sarsa, even though its continuous actions get rounded to actions in the same finite action space that may contain only a small number of actions. In particular, we show that Cacla retains much better performance when the action space is changed by removing some actions after some time of learning.
Keywords :
Markov processes; decision theory; function approximation; learning (artificial intelligence); neural nets; Cacla algorithm; MDP; Markov decision process; Q-learning algorithm; Sarsa algorithm; cart pole task; continuous action space; discrete algorithm; function approximation; mountain car task; neural network; real-world control problem; reinforcement learning algorithm; Gears; Learning automata; Neural networks; Robustness; State-space methods; Testing;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 2009. IJCNN 2009. International Joint Conference on
Conference_Location :
Atlanta, GA
ISSN :
1098-7576
Print_ISBN :
978-1-4244-3548-7
Electronic_ISBN :
1098-7576
Type :
conf
DOI :
10.1109/IJCNN.2009.5178745
Filename :
5178745
Link To Document :
بازگشت