• DocumentCode
    756772
  • Title

    Multiple response learning automata

  • Author

    Economides, Anastasios A.

  • Author_Institution
    Univ. of Macedonia, Thessaloniki, Greece
  • Volume
    26
  • Issue
    1
  • fYear
    1996
  • fDate
    2/1/1996 12:00:00 AM
  • Firstpage
    153
  • Lastpage
    156
  • Abstract
    Learning Automata update their action probabilites on the basis of the response they get from a random environment. They use a reward adaptation rate for a favorable environment´s response and a penalty adaptation rate for an unfavorable environment´s response. In this correspondence, we introduce Multiple Response learning automata by explicitly classifying the environment responses into a reward (favorable) set and a penalty (unfavorable) set. We derive a new reinforcement scheme which uses different reward or penalty rates for the corresponding reward (favorable) or penalty (unfavorable) responses. Well known learning automata, such as the LR-P;LR-I; LR-eP are special cases of these Multiple Response learning automata. These automata are feasible at each step, nonabsorbing (when the penalty functions are positive), and strictly distance diminishing. Finally, we provide conditions in order that they are ergodic and expedient
  • Keywords
    automata theory; learning (artificial intelligence); learning automata; Multiple Response; action probabilites; learning automata; penalty adaptation rate; reward adaptation rate; Learning automata; Learning systems; Mathematical model; Packet switching; Psychology; Routing; Stochastic processes; Sufficient conditions; Switching circuits; Systems engineering and theory;
  • fLanguage
    English
  • Journal_Title
    Systems, Man, and Cybernetics, Part B: Cybernetics, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    1083-4419
  • Type

    jour

  • DOI
    10.1109/3477.484448
  • Filename
    484448