• DocumentCode
    768372
  • Title

    Benefits of gain: speeded learning and minimal hidden layers in back-propagation networks

  • Author

    Kruschke, John K. ; Movellan, Javier R.

  • Author_Institution
    Dept. of Psychol., California Univ., Berkeley, CA, USA
  • Volume
    21
  • Issue
    1
  • fYear
    1991
  • Firstpage
    273
  • Lastpage
    280
  • Abstract
    The gain of a mode in a connectionist network is a multiplicative constant that amplifies or attenuates the net input to the node. The benefits of adaptive gains in back-propagation networks are explored. It is shown that gradient descent with respect to gain greatly increases learning speed by amplifying those directions in weight space that are successfully chosen by gradient descent on weights. Adaptive gains also allow normalization of weight vectors without loss of computational capacity, and the authors suggest a simple modification of the learning rule that automatically achieves weight normalization. A method for creating small hidden layers by making hidden node gains compete according to similarities between nodes in an effect to improve generalization performance is described. Simulations show that this competition method is more effective than the special case of gain decay
  • Keywords
    learning systems; neural nets; adaptive gains; back-propagation networks; connectionist network; gradient descent; minimal hidden layers; neural nets; speeded learning; Computational modeling; Feedforward systems; Helium; Intelligent networks; Nonhomogeneous media; Performance gain; Psychology;
  • fLanguage
    English
  • Journal_Title
    Systems, Man and Cybernetics, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    0018-9472
  • Type

    jour

  • DOI
    10.1109/21.101159
  • Filename
    101159