• DocumentCode
    549563
  • Title

    Deriving a near-optimal power management policy using model-free reinforcement learning and Bayesian classification

  • Author

    Wang, Yanzhi ; Xie, Qing ; Ammari, Ahmed ; Pedram, Massoud

  • Author_Institution
    Dept. of Electr. Eng., Univ. of Southern California, Los Angeles, CA, USA
  • fYear
    2011
  • fDate
    5-9 June 2011
  • Firstpage
    41
  • Lastpage
    46
  • Abstract
    To cope with the variations and uncertainties that emanate from hardware and application characteristics, dynamic power management (DPM) frameworks must be able to learn about the system inputs and environment and adjust the power management policy on the fly. In this paper we present an online adaptive DPM technique based on model-free reinforcement learning (RL), which is commonly used to control stochastic dynamical systems. In particular, we employ temporal difference learning for semi-Markov decision process (SMDP) for the model-free RL. In addition a novel workload predictor based on an online Bayes classifier is presented to provide effective estimates of the workload states for the RL algorithm. In this DPM framework, power and latency tradeoffs can be precisely controlled based on a user-defined parameter. Experiments show that amount of average power saving (without any increase in the latency) is up to 16.7% compared to a reference expert-based approach. Alternatively, the per-request latency reduction without any power consumption increase is up to 28.6% compared to the expert-based approach.
  • Keywords
    Bayes methods; Markov processes; learning (artificial intelligence); pattern classification; power aware computing; Bayesian classification; dynamic power management frameworks; model-free reinforcement learning; near-optimal power management policy; online Bayes classifier; semiMarkov decision process; stochastic dynamical systems; temporal difference learning; Classification algorithms; Learning; Markov processes; Power demand; Prediction algorithms; Strontium; Bayes Classification; Dynamic Power Management; Reinforcement Learning;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Design Automation Conference (DAC), 2011 48th ACM/EDAC/IEEE
  • Conference_Location
    New York, NY
  • ISSN
    0738-100x
  • Print_ISBN
    978-1-4503-0636-2
  • Type

    conf

  • Filename
    5981919