DocumentCode
353261
Title
α-EM algorithm and α-ICA learning based upon extended logarithmic information measures
Author
Mtsuyama, Y. ; Nimoto, T. ; Katsumata, Naoto ; Suzuki, Yoshitaka ; Furukawa, Satoshi
Author_Institution
Dept. of Electr. Electron. & Comput. Eng., Waseda Univ., Tokyo, Japan
Volume
3
fYear
2000
fDate
2000
Firstpage
351
Abstract
The α-logarithm extends the logarithm as the special case of α=-1. Usage of α-related information measures based upon this extended logarithm is expected to be effective to speedup of convergence, i.e., on the improvement of learning aptitude. In this paper, two typical cases are investigated. One is the α-EM algorithm (α-expectation-maximization algorithm) which is derived from the α-log-likelihood ratio. The other is the α-ICA (α-independent component analysis) which is formulated as minimizing the α-mutual information. In the derivation of both algorithms, the α-divergence plays the main role. For the α-EM algorithm, the reason for the speedup is explained using Hessian and Jacobian matrices for learning. For the α-ICA learning, methods of exploiting the past and future information are presented. Examples are shown on single-loop α-EM and sample-based α-ICA. In all cases, effective speedups are observed. Thus, this paper´s examples together with formerly reported ones are evidences that the speed improvement by the α-logarithm is a general property beyond individual problems
Keywords
Hessian matrices; Jacobian matrices; learning (artificial intelligence); maximum likelihood estimation; neural nets; principal component analysis; α-EM algorithm; α-ICA learning; α-divergence; α-expectation-maximization algorithm; α-independent component analysis; α-log-likelihood ratio; α-mutual information; α-related information measures; Hessian matrix; Jacobian matrix; extended logarithmic information measures; neural nets; sample-based α-ICA; single-loop α-EM; Convergence; Entropy; Equations; Expectation-maximization algorithms; Independent component analysis; Information analysis; Jacobian matrices; Minimization methods; Mutual information; Velocity measurement;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks, 2000. IJCNN 2000, Proceedings of the IEEE-INNS-ENNS International Joint Conference on
Conference_Location
Como
ISSN
1098-7576
Print_ISBN
0-7695-0619-4
Type
conf
DOI
10.1109/IJCNN.2000.861329
Filename
861329
Link To Document