DocumentCode :
2191955
Title :
Maximize user rewards in distributed generation environments using reinforcement learning
Author :
Li, Bei ; Gangadhar, Siddharth ; Cheng, Samuel ; Verma, Pramode K.
Author_Institution :
Telecommun. Eng. Program, Univ. of Oklahoma, Tulsa, OK, USA
fYear :
2011
fDate :
25-26 May 2011
Firstpage :
1
Lastpage :
6
Abstract :
In Smart Grid environments, with distributed generation, homes are encouraged to generate power and sell it back to utilities. Time of Use pricing techniques and the introduction of storage devices would greatly influence a user in deciding when to sell back power and how much to sell. Therefore, a study of sequential decision making algorithms that can optimize the total pay off for the user is necessary. In this paper, Reinforcement Learning is used to solve this optimization problem. The problem of determining when to sell back power is formulated as a Markov decision process and a near optimal strategy is chosen using policy iteration. The results show a significant increase of total rewards from selling back power with the proposed approach.
Keywords :
Markov processes; decision making; distributed power generation; learning (artificial intelligence); power distribution economics; power engineering computing; smart power grids; Markov decision process; distributed generation environment; optimization problem; reinforcement learning; sequential decision making algorithms; smart grid environments; storage devices; Decision making; Markov processes; Power generation; Pricing; Springs; Testing; Training; Distributed Generation; Machine Learning; Markov decision process; Maximize user rewards; Reinforcement Learning; Smart Grid; Time of Use pricing;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Energytech, 2011 IEEE
Conference_Location :
Cleveland, OH
Print_ISBN :
978-1-4577-0777-3
Electronic_ISBN :
978-1-4577-0775-9
Type :
conf
DOI :
10.1109/EnergyTech.2011.5948518
Filename :
5948518
Link To Document :
بازگشت