DocumentCode
493379
Title
Bounds of optimal learning
Author
Belavkin, Roman V.
Author_Institution
Sch. of Eng. & Inf. Sci., Middlesex Univ., London
fYear
2009
fDate
March 30 2009-April 2 2009
Firstpage
199
Lastpage
204
Abstract
Learning is considered as a dynamic process described by a trajectory on a statistical manifold, and a topology is introduced defining trajectories continuous in information. The analysis generalises the application of Orlicz spaces in non-parametric information geometry to topological function spaces with asymmetric gauge functions (e.g. quasi-metric spaces defined in terms of KL divergence). Optimality conditions are formulated for dynamical constraints, and two main results are outlined: 1) Parametrisation of optimal learning trajectories from empirical constraints using generalised characteristic potentials; 2) A gradient theorem for the potentials defining optimal utility and information bounds of a learning system. These results not only generalise some known relations of statistical mechanics and variational methods in information theory, but also can be used for optimisation of the exploration-exploitation balance in online learning systems.
Keywords
gradient methods; learning systems; statistical analysis; Orlicz spaces; gradient theorem; information theory; nonparametric information geometry; online learning systems; optimal learning trajectories; statistical manifold; Constraint optimization; Constraint theory; Entropy; Information analysis; Information geometry; Information theory; Learning systems; Optimal control; Topology; Utility theory;
fLanguage
English
Publisher
ieee
Conference_Titel
Adaptive Dynamic Programming and Reinforcement Learning, 2009. ADPRL '09. IEEE Symposium on
Conference_Location
Nashville, TN
Print_ISBN
978-1-4244-2761-1
Type
conf
DOI
10.1109/ADPRL.2009.4927545
Filename
4927545
Link To Document