• DocumentCode
    463394
  • Title

    Learn to Coordinate with Generic Non-Stationary Opponents

  • Author

    Kaifu, Zhang

  • Author_Institution
    Dept. of Comput. Sci. & Technol., Tsinghua Univ., Beijing
  • Volume
    1
  • fYear
    2006
  • fDate
    17-19 July 2006
  • Firstpage
    558
  • Lastpage
    565
  • Abstract
    Learning to coordinate with non-stationary opponents is a major challenge for adaptive agents. Most previous research investigated only restricted classes of such dynamic opponents. The main contribution of this paper is twofold: (i) A class of generic non-stationary opponents is introduced. The opponents keep mixed strategies which change with less regularity. Its showed that the independent reinforcement learners (ILs), which have neither prior knowledge nor opponent models, cannot coordinate well with this type of opponent, (ii) A new exploration strategy, the DAE (detect and explore) mechanism, is tailored for the ILs in such coordination tasks. This mechanism allows the ILs dynamically detect changes in the opponents behavior and adjust their learning rate and exploration temperature. It´s showed that ILs using this strategy are still able to converge in self-play and are able to coordinate well with the non-stationary opponents
  • Keywords
    learning (artificial intelligence); multi-agent systems; adaptive agents; detect and explore mechanism; dynamic opponents; exploration strategy; generic nonstationary opponents; independent reinforcement learner; Cognitive informatics; Computer science; Decision making; Game theory; Learning; Multiagent systems; Temperature; Testing; Coordination game; Exploration strategy; Non-stationary opponent; Reinforcement learning;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Cognitive Informatics, 2006. ICCI 2006. 5th IEEE International Conference on
  • Conference_Location
    Beijing
  • Print_ISBN
    1-4244-0475-4
  • Type

    conf

  • DOI
    10.1109/COGINF.2006.365546
  • Filename
    4216463