• DocumentCode
    1097294
  • Title

    Decentralized Learning in Finite Markov Chains: Revisited

  • Author

    Chang, Hyeong Soo

  • Author_Institution
    Dept. of Comput. Sci. & Eng., Sogang Univ., Seoul, South Korea
  • Volume
    54
  • Issue
    7
  • fYear
    2009
  • fDate
    7/1/2009 12:00:00 AM
  • Firstpage
    1648
  • Lastpage
    1653
  • Abstract
    The convergence proof in the paper ldquoDecentralized learning in finite Markov chains,rdquo published in the IEEE Transactions on Automatic Control, vol. AC-31, no. 6, pp. 519-526, 1986, is incomplete. This note first provides a sufficient condition for the existence of a unique optimal policy for infinite-horizon average-cost Markov decision processes (MDPs), making the convergence result established by Wheeler and Narendra preserved with the condition. We then present a novel simulation-based decentralized algorithm, called ldquosampled joint-strategy fictitious play for MDPrdquo for average MDPs based on the recent study by Garcia of a decentralized approach to discrete optimization via fictitious play applied to games with identical payoffs. We establish a stronger almost-sure convergence result than Wheeler and Narendra´s, showing that the sequence of probability distributions over the policy space for a given MDP generated by the algorithm converges to a unique optimal policy with probability one.
  • Keywords
    Markov processes; adaptive control; decentralised control; discrete systems; learning systems; optimisation; average-cost Markov decision processes; decentralized learning; discrete optimization; finite Markov chains; infinite-horizon processes; simulation-based decentralized algorithm; Controlled Markov chain; Markov decision process; decentralized learning; fictitious play; learning automata;
  • fLanguage
    English
  • Journal_Title
    Automatic Control, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    0018-9286
  • Type

    jour

  • DOI
    10.1109/TAC.2009.2017977
  • Filename
    5109518