DocumentCode
264341
Title
Reinforcement learning algorithms for dynamic power management
Author
Triki, M. ; Ammari, Ahmed C. ; Wang, Yannan ; Pedram, Massoud
Author_Institution
MMA Lab., Carthage Univ., Tunis, Tunisia
fYear
2014
fDate
18-20 Jan. 2014
Firstpage
1
Lastpage
6
Abstract
In this paper we present a dynamic power management (DPM) framework based on model-free reinforcement learning (RL) techniques. For the RL algorithms, we employ both temporal difference learning and Q-learning for semi-Markov decision process in a continuous-time manner. The proposed DPM is model-free and do not require any prior information of the workload characteristics. The power manager learns the optimal power management policy that significantly reduces energy consumption while maintaining an acceptable performance level. Moreover, power-latency tradeoffs can be precisely controlled based on a user-defined parameter. In addition, the temporal difference (TD) learning is compared with the Q-learning approach in terms of both performance and convergence speed. Experiments on network cards show that TD achieves better power saving without sacrificing any latency and has faster convergence speed compared to Q-learning.
Keywords
Markov processes; energy conservation; learning (artificial intelligence); power engineering computing; Q-Iearning; RL algorithms; TD; dynamic power management; energy consumption reduction; model-free reinforcement learning algorithms; network cards; optimal power management policy; power-latency tradeoffs; semiMarkov decision process; temporal difference learning; user-defined parameter; Convergence;
fLanguage
English
Publisher
ieee
Conference_Titel
Computer Applications & Research (WSCAR), 2014 World Symposium on
Conference_Location
Sousse
Print_ISBN
978-1-4799-2805-7
Type
conf
DOI
10.1109/WSCAR.2014.6916835
Filename
6916835
Link To Document