• DocumentCode
    699134
  • Title

    The LMS, PNLMS, and exponentiated gradient algorithms

  • Author

    Benesty, Jacob ; Yiteng Huang

  • Author_Institution
    INRS-EMT, Univ. du Quebec, Montreal, QC, Canada
  • fYear
    2004
  • fDate
    6-10 Sept. 2004
  • Firstpage
    721
  • Lastpage
    724
  • Abstract
    Sparse impulse responses are encountered in many applications (network and acoustic echo cancellation, feedback cancellation in hearing aids, etc). Recently, a class of exponentiated gradient (EG) algorithms has been proposed. One of the algorithms belonging to this class, the so-called EG± algorithm, converges and tracks much better than the classical stochastic gradient, or LMS, algorithm for sparse impulse responses. In this paper, we show how to derive the different algorithms. We analyze the EG± algorithm and explain when to expect it to behave like the LMS algorithm. It is also shown that the proportionate normalized LMS (PNLMS) algorithm proposed by Duttweiler in the context of network echo cancellation is an approximation of the EG±.
  • Keywords
    gradient methods; least mean squares methods; EG algorithms; PNLMS algorithm; exponentiated gradient algorithms; network echo cancellation; proportionate normalized LMS; sparse impulse responses; stochastic gradient; Abstracts; Artificial neural networks; Equations; Least squares approximations; Zinc;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Signal Processing Conference, 2004 12th European
  • Conference_Location
    Vienna
  • Print_ISBN
    978-320-0001-65-7
  • Type

    conf

  • Filename
    7079664