Title :
Optimal Demand Response Using Device-Based Reinforcement Learning
Author :
Zheng Wen ; O´Neill, Daniel ; Maei, Hamid
Author_Institution :
Yahoo Labs., Sunnyvale, CA, USA
Abstract :
Demand response (DR) for residential and small commercial buildings is estimated to account for as much as 65% of the total energy savings potential of DR, and previous work shows that a fully automated energy management system (EMS) is a necessary prerequisite to DR in these areas. In this paper, we propose a novel EMS formulation for DR problems in these sectors. Specifically, we formulate a fully automated EMS´s rescheduling problem as a reinforcement learning (RL) problem, and argue that this RL problem can be approximately solved by decomposing it over device clusters. Compared with existing formulations, our new formulation does not require explicitly modeling the user´s dissatisfaction on job rescheduling, enables the EMS to self-initiate jobs, allows the user to initiate more flexible requests, and has a computational complexity linear in the number of device clusters. We also demonstrate the simulation results of applying Q-learning, one of the most popular and classical RL algorithms, to a representative example.
Keywords :
demand side management; energy management systems; learning (artificial intelligence); power engineering computing; DR problems; EMS; Q-learning; RL problem; computational complexity; device clusters; device-based reinforcement learning; energy management system; job rescheduling; optimal demand response; Buildings; Clustering algorithms; Energy consumption; Energy management; Markov processes; Optimal scheduling; Real-time systems; Building and home automation; Markov decision process (MDP); demand response (DR); energy management system (EMS); reinforcement learning (RL);
Journal_Title :
Smart Grid, IEEE Transactions on
DOI :
10.1109/TSG.2015.2396993