• DocumentCode
    2109030
  • Title

    On solving optimal policies for event-based dynamic programming

  • Author

    Jia Qing-Shan

  • Author_Institution
    Dept. of Autom., Tsinghua Univ., Beijing, China
  • fYear
    2010
  • fDate
    29-31 July 2010
  • Firstpage
    1511
  • Lastpage
    1516
  • Abstract
    Markov decision processes (MDPs) have provided general frameworks for many control, decision making, and optimization problems. However, solving the optimal policies for many such problems is computationally prohibitive due to the large state space and the large action space. Event-based dynamic programming (EDP) has been developed to formulate the event-based decision making processes. Since the number of events could increase only linearly with respect to (w.r.t.) the problem scale, EDP provides a computationally feasible way to many problems which are time-consuming to solve in the MDP framework. However, the event sequence is not Markov, the optimal event-based policy could depend on the entire history, which cannot be implemented in practice. In this paper, for EDP with discrete and finite state space we construct a completely observable MDP with both the belief distribution over the internal system state and the current observable event being the state. Then we show that solving the original EDP is equivalent to solving this belief-event dynamic programming (BEDP), the optimal policies of which can be found within Markov policies that can be implemented in practice. Then potential-based policy iteration algorithms for completely observable MDP can be applied. We also discuss extensions to finite-stage EDP.
  • Keywords
    Markov processes; decision making; discrete event systems; dynamic programming; iterative methods; state-space methods; Markov decision process; belief-event dynamic programming; event-based decision making process; event-based dynamic programming; finite state space; large action space; large state space; potential-based policy iteration algorithms; Aerospace electronics; Approximation methods; Decision making; Dynamic programming; History; Markov processes; Process control; Discrete Event Dynamic Systems; Event-based Dynamic Programming; Markov Decision Processes;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Control Conference (CCC), 2010 29th Chinese
  • Conference_Location
    Beijing
  • Print_ISBN
    978-1-4244-6263-6
  • Type

    conf

  • Filename
    5573479