Title :
Enhancement of a gradient descent or ascent with a function weight for the minimum α-divergence
Author_Institution :
Sch. of Knowledge Sci., JAIST, Ishikawa, Japan
Abstract :
The natural gradient method with Riemannian metric shows good performance for a parameter estimation problem based on the minimization of a divergence between an observed data and the estimated functions or models. However, it requires much computational times for iterative calculations of the inverse matrix of the metric. In this paper, we consider such estimation problem on an α-affine manifold in a function space over finite discrete variables. By using the auto-parallelism, we derive efficient linear calculation methods for the (-α)-geodesic on the manifold, and for the α-projection onto it from an observed data. The methods can be applied to any initial, terminal, and the interpolated points on the geodesic. Further, we discuss an enhancement of gradient descent or ascent with a function weight by an extension of log, which is related to the α-divergence. The results suggests an adaptive selection of the value of α ∈ R for the family of parametric functions or models (as α-affine manifold) according to the data
Keywords :
gradient methods; interpolation; minimisation; parameter estimation; Riemannian metric; divergence; function weight; gradient ascent method; gradient descent method; interpolation; minimization; parameter estimation; Gradient methods; Information geometry; Linearity; Machine learning; Minimization methods; Multilayer perceptrons; Neural networks; Parameter estimation; Probability distribution; Stochastic processes;
Conference_Titel :
Systems, Man, and Cybernetics, 1999. IEEE SMC '99 Conference Proceedings. 1999 IEEE International Conference on
Conference_Location :
Tokyo
Print_ISBN :
0-7803-5731-0
DOI :
10.1109/ICSMC.1999.815528