• DocumentCode
    264341
  • Title

    Reinforcement learning algorithms for dynamic power management

  • Author

    Triki, M. ; Ammari, Ahmed C. ; Wang, Yannan ; Pedram, Massoud

  • Author_Institution
    MMA Lab., Carthage Univ., Tunis, Tunisia
  • fYear
    2014
  • fDate
    18-20 Jan. 2014
  • Firstpage
    1
  • Lastpage
    6
  • Abstract
    In this paper we present a dynamic power management (DPM) framework based on model-free reinforcement learning (RL) techniques. For the RL algorithms, we employ both temporal difference learning and Q-learning for semi-Markov decision process in a continuous-time manner. The proposed DPM is model-free and do not require any prior information of the workload characteristics. The power manager learns the optimal power management policy that significantly reduces energy consumption while maintaining an acceptable performance level. Moreover, power-latency tradeoffs can be precisely controlled based on a user-defined parameter. In addition, the temporal difference (TD) learning is compared with the Q-learning approach in terms of both performance and convergence speed. Experiments on network cards show that TD achieves better power saving without sacrificing any latency and has faster convergence speed compared to Q-learning.
  • Keywords
    Markov processes; energy conservation; learning (artificial intelligence); power engineering computing; Q-Iearning; RL algorithms; TD; dynamic power management; energy consumption reduction; model-free reinforcement learning algorithms; network cards; optimal power management policy; power-latency tradeoffs; semiMarkov decision process; temporal difference learning; user-defined parameter; Convergence;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Computer Applications & Research (WSCAR), 2014 World Symposium on
  • Conference_Location
    Sousse
  • Print_ISBN
    978-1-4799-2805-7
  • Type

    conf

  • DOI
    10.1109/WSCAR.2014.6916835
  • Filename
    6916835