DocumentCode :
1441538
Title :
Stochastic Optimal Relaxed Automatic Generation Control in Non-Markov Environment Based on Multi-Step Q(\\lambda ) Learning
Author :
Yu, Tao ; Zhou, Bin ; Chan, Ka Wing ; Chen, Liang ; Yang, Bo
Author_Institution :
Coll. of Electr. Power, South China Univ. of Technol., Guangzhou, China
Volume :
26
Issue :
3
fYear :
2011
Firstpage :
1272
Lastpage :
1282
Abstract :
This paper proposes a stochastic optimal relaxed control methodology based on reinforcement learning (RL) for solving the automatic generation control (AGC) under NERC´s control performance standards (CPS). The multi-step Q(λ) learning algorithm is introduced to effectively tackle the long time-delay control loop for AGC thermal plants in non-Markov environment. The moving averages of CPS1/ACE are adopted as the state feedback input, and the CPS control and relaxed control objectives are formulated as multi-criteria reward function via linear weighted aggregate method. This optimal AGC strategy provides a customized platform for interactive self-learning rules to maximize the long-run discounted reward. Statistical experiments show that the RL theory based Q(λ) controllers can effectively enhance the robustness and dynamic performance of AGC systems, and reduce the number of pulses and pulse reversals while the CPS compliances are ensured. The novel AGC scheme also provides a convenient way of controlling the degree of CPS compliance and relaxation by online tuning relaxation factors to implement the desirable relaxed control.
Keywords :
delays; learning (artificial intelligence); optimal control; power generation control; power system stability; robust control; state feedback; statistical analysis; stochastic processes; thermal power stations; AGC thermal plant; CPS compliance; CPS control; NERC control performance standard; RL theory based controller; interactive self learning rule; linear weighted aggregate method; long run discounted reward; multicriteria reward function; multistep Q(λ) learning algorithm; nonMarkov environment; online tuning relaxation factors; optimal AGC strategy; pulse reversal; reinforcement learning; state feedback input; statistical experiment; stochastic optimal relaxed automatic generation control; time delay control loop; Aerospace electronics; Frequency control; Markov processes; Power grids; Power system dynamics; Standards; AGC; CPS; multi-step $Q(lambda)$ learning; non-Markov environment; relaxed control; stochastic optimization;
fLanguage :
English
Journal_Title :
Power Systems, IEEE Transactions on
Publisher :
ieee
ISSN :
0885-8950
Type :
jour
DOI :
10.1109/TPWRS.2010.2102372
Filename :
5706397
Link To Document :
بازگشت