• DocumentCode
    1168045
  • Title

    Pruning recurrent neural networks for improved generalization performance

  • Author

    Giles, C. Lee ; Omlin, Christian W.

  • Volume
    5
  • Issue
    5
  • fYear
    1994
  • fDate
    9/1/1994 12:00:00 AM
  • Firstpage
    848
  • Lastpage
    851
  • Abstract
    Determining the architecture of a neural network is an important issue for any learning task. For recurrent neural networks no general methods exist that permit the estimation of the number of layers of hidden neurons, the size of layers or the number of weights. We present a simple pruning heuristic that significantly improves the generalization performance of trained recurrent networks. We illustrate this heuristic by training a fully recurrent neural network on positive and negative strings of a regular grammar. We also show that rules extracted from networks trained with this pruning heuristic are more consistent with the rules to be learned. This performance improvement is obtained by pruning and retraining the networks. Simulations are shown for training and pruning a recurrent neural net on strings generated by two regular grammars, a randomly-generated 10-state grammar and an 8-state, triple-parity grammar. Further simulations indicate that this pruning method can have generalization performance superior to that obtained by training with weight decay
  • Keywords
    generalisation (artificial intelligence); grammars; learning (artificial intelligence); recurrent neural nets; 8-state triple-parity grammar; generalization performance; hidden neurons; learning task; negative string; positive strings; pruning heuristic; recurrent neural network prunning; regular grammar; regular grammars; state grammar; weight decay; Clustering algorithms; Doped fiber amplifiers; Learning automata; National electric code; Neural networks; Neurons; Quantization; Recurrent neural networks; Space exploration; State-space methods;
  • fLanguage
    English
  • Journal_Title
    Neural Networks, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    1045-9227
  • Type

    jour

  • DOI
    10.1109/72.317740
  • Filename
    317740