DocumentCode :
493379
Title :
Bounds of optimal learning
Author :
Belavkin, Roman V.
Author_Institution :
Sch. of Eng. & Inf. Sci., Middlesex Univ., London
fYear :
2009
fDate :
March 30 2009-April 2 2009
Firstpage :
199
Lastpage :
204
Abstract :
Learning is considered as a dynamic process described by a trajectory on a statistical manifold, and a topology is introduced defining trajectories continuous in information. The analysis generalises the application of Orlicz spaces in non-parametric information geometry to topological function spaces with asymmetric gauge functions (e.g. quasi-metric spaces defined in terms of KL divergence). Optimality conditions are formulated for dynamical constraints, and two main results are outlined: 1) Parametrisation of optimal learning trajectories from empirical constraints using generalised characteristic potentials; 2) A gradient theorem for the potentials defining optimal utility and information bounds of a learning system. These results not only generalise some known relations of statistical mechanics and variational methods in information theory, but also can be used for optimisation of the exploration-exploitation balance in online learning systems.
Keywords :
gradient methods; learning systems; statistical analysis; Orlicz spaces; gradient theorem; information theory; nonparametric information geometry; online learning systems; optimal learning trajectories; statistical manifold; Constraint optimization; Constraint theory; Entropy; Information analysis; Information geometry; Information theory; Learning systems; Optimal control; Topology; Utility theory;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Adaptive Dynamic Programming and Reinforcement Learning, 2009. ADPRL '09. IEEE Symposium on
Conference_Location :
Nashville, TN
Print_ISBN :
978-1-4244-2761-1
Type :
conf
DOI :
10.1109/ADPRL.2009.4927545
Filename :
4927545
Link To Document :
بازگشت