DocumentCode :
1802211
Title :
Min-max approximate dynamic programming
Author :
O´Donoghue, Brendan ; Wang, Yang ; Boyd, Stephen
Author_Institution :
Stanford Univ., Stanford, CA, USA
fYear :
2011
fDate :
28-30 Sept. 2011
Firstpage :
424
Lastpage :
431
Abstract :
In this paper we describe an approximate dynamic programming policy for a discrete-time dynamical system perturbed by noise. The approximate value function is the pointwise supremum of a family of lower bounds on the value function of the stochastic control problem; evaluating the control policy involves the solution of a min-max or saddle-point problem. For a quadratically constrained linear quadratic control problem, evaluating the policy amounts to solving a semidefinite program at each time step. By evaluating the policy, we obtain a lower bound on the value function, which can be used to evaluate performance: When the lower bound and the achieved performance of the policy are close, we can conclude that the policy is nearly optimal. We describe several numerical examples where this is indeed the case.
Keywords :
approximation theory; discrete time systems; dynamic programming; minimax techniques; stochastic systems; discrete-time dynamical system; linear quadratic control problem; min max approximate dynamic programming; saddle point problem; stochastic control problem; Approximation methods; Dynamic programming; Investments; Minimization; Monte Carlo methods; Noise; Optimization;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Computer-Aided Control System Design (CACSD), 2011 IEEE International Symposium on
Conference_Location :
Denver, CO
Print_ISBN :
978-1-4577-1066-7
Electronic_ISBN :
978-1-4577-1067-4
Type :
conf
DOI :
10.1109/CACSD.2011.6044538
Filename :
6044538
Link To Document :
بازگشت