• DocumentCode
    1166480
  • Title

    Learning exponential state-growth languages by hill climbing

  • Author

    Tabor, Whitney

  • Author_Institution
    Dept. of Psychol., Connecticut Univ., Storrs, CT, USA
  • Volume
    14
  • Issue
    2
  • fYear
    2003
  • fDate
    3/1/2003 12:00:00 AM
  • Firstpage
    444
  • Lastpage
    446
  • Abstract
    Training recurrent neural networks on infinite state languages has been successful with languages in which the minimal number of machine states grows linearly with the length of the sentence, but has faired poorly with exponential state-growth languages. The new architecture learns several exponential state-growth languages in near perfect by hill climbing.
  • Keywords
    fractals; learning (artificial intelligence); probability; recurrent neural nets; exponential state-growth languages; fractal learning; hill climbing; infinite state languages; probability distribution; recurrent neural networks; recursive computations; Computer architecture; Computer networks; Counting circuits; Fractals; Learning automata; Neural networks; Probability distribution; Recurrent neural networks; Turing machines; Vocabulary;
  • fLanguage
    English
  • Journal_Title
    Neural Networks, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    1045-9227
  • Type

    jour

  • DOI
    10.1109/TNN.2003.809421
  • Filename
    1189642