• DocumentCode
    493379
  • Title

    Bounds of optimal learning

  • Author

    Belavkin, Roman V.

  • Author_Institution
    Sch. of Eng. & Inf. Sci., Middlesex Univ., London
  • fYear
    2009
  • fDate
    March 30 2009-April 2 2009
  • Firstpage
    199
  • Lastpage
    204
  • Abstract
    Learning is considered as a dynamic process described by a trajectory on a statistical manifold, and a topology is introduced defining trajectories continuous in information. The analysis generalises the application of Orlicz spaces in non-parametric information geometry to topological function spaces with asymmetric gauge functions (e.g. quasi-metric spaces defined in terms of KL divergence). Optimality conditions are formulated for dynamical constraints, and two main results are outlined: 1) Parametrisation of optimal learning trajectories from empirical constraints using generalised characteristic potentials; 2) A gradient theorem for the potentials defining optimal utility and information bounds of a learning system. These results not only generalise some known relations of statistical mechanics and variational methods in information theory, but also can be used for optimisation of the exploration-exploitation balance in online learning systems.
  • Keywords
    gradient methods; learning systems; statistical analysis; Orlicz spaces; gradient theorem; information theory; nonparametric information geometry; online learning systems; optimal learning trajectories; statistical manifold; Constraint optimization; Constraint theory; Entropy; Information analysis; Information geometry; Information theory; Learning systems; Optimal control; Topology; Utility theory;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Adaptive Dynamic Programming and Reinforcement Learning, 2009. ADPRL '09. IEEE Symposium on
  • Conference_Location
    Nashville, TN
  • Print_ISBN
    978-1-4244-2761-1
  • Type

    conf

  • DOI
    10.1109/ADPRL.2009.4927545
  • Filename
    4927545