• DocumentCode
    1906647
  • Title

    Constructive learning of recurrent neural networks

  • Author

    Chen, D. ; Giles, C.L. ; Sun, G.Z. ; Chen, H.H. ; Lee, Y.C. ; Goudreau, M.W.

  • Author_Institution
    Inst. for Adv. Comput. Studies, Maryland Univ., College Park, MD, USA
  • fYear
    1993
  • fDate
    1993
  • Firstpage
    1196
  • Abstract
    It is difficult to determine the minimal neural network structure for a particular automaton. A large recurrent network in practice is very difficult to train. Constructive or destructive recurrent methods might offer a solution to this problem. It is proved that one current method, recurrent cascade correlation, has fundamental limitations in representation and thus in its learning capabilities. A preliminary approach to circumventing these limitations by devising a simple constructive training method that adds neurons during training while still preserving the powerful fully recurrent structure is given. Through simulations it is shown that such a method can learn many types of regular grammars which the recurrent cascade correlation method is unable to learn
  • Keywords
    grammars; learning (artificial intelligence); recurrent neural nets; constructive training method; fully recurrent structure; minimal neural network structure; recurrent cascade correlation; recurrent neural networks; regular grammars; Convergence; Educational institutions; Learning automata; Military computing; Neural networks; Neurons; Predictive models; Recurrent neural networks; Signal processing; Upper bound;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Neural Networks, 1993., IEEE International Conference on
  • Conference_Location
    San Francisco, CA
  • Print_ISBN
    0-7803-0999-5
  • Type

    conf

  • DOI
    10.1109/ICNN.1993.298727
  • Filename
    298727