• DocumentCode
    353261
  • Title

    α-EM algorithm and α-ICA learning based upon extended logarithmic information measures

  • Author

    Mtsuyama, Y. ; Nimoto, T. ; Katsumata, Naoto ; Suzuki, Yoshitaka ; Furukawa, Satoshi

  • Author_Institution
    Dept. of Electr. Electron. & Comput. Eng., Waseda Univ., Tokyo, Japan
  • Volume
    3
  • fYear
    2000
  • fDate
    2000
  • Firstpage
    351
  • Abstract
    The α-logarithm extends the logarithm as the special case of α=-1. Usage of α-related information measures based upon this extended logarithm is expected to be effective to speedup of convergence, i.e., on the improvement of learning aptitude. In this paper, two typical cases are investigated. One is the α-EM algorithm (α-expectation-maximization algorithm) which is derived from the α-log-likelihood ratio. The other is the α-ICA (α-independent component analysis) which is formulated as minimizing the α-mutual information. In the derivation of both algorithms, the α-divergence plays the main role. For the α-EM algorithm, the reason for the speedup is explained using Hessian and Jacobian matrices for learning. For the α-ICA learning, methods of exploiting the past and future information are presented. Examples are shown on single-loop α-EM and sample-based α-ICA. In all cases, effective speedups are observed. Thus, this paper´s examples together with formerly reported ones are evidences that the speed improvement by the α-logarithm is a general property beyond individual problems
  • Keywords
    Hessian matrices; Jacobian matrices; learning (artificial intelligence); maximum likelihood estimation; neural nets; principal component analysis; α-EM algorithm; α-ICA learning; α-divergence; α-expectation-maximization algorithm; α-independent component analysis; α-log-likelihood ratio; α-mutual information; α-related information measures; Hessian matrix; Jacobian matrix; extended logarithmic information measures; neural nets; sample-based α-ICA; single-loop α-EM; Convergence; Entropy; Equations; Expectation-maximization algorithms; Independent component analysis; Information analysis; Jacobian matrices; Minimization methods; Mutual information; Velocity measurement;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Neural Networks, 2000. IJCNN 2000, Proceedings of the IEEE-INNS-ENNS International Joint Conference on
  • Conference_Location
    Como
  • ISSN
    1098-7576
  • Print_ISBN
    0-7695-0619-4
  • Type

    conf

  • DOI
    10.1109/IJCNN.2000.861329
  • Filename
    861329