• DocumentCode
    2437263
  • Title

    The effects of varying memory vector size in a network that learns to learn

  • Author

    Brown, Gordon D A ; Hyland, Peter ; Hulme, Charles

  • Author_Institution
    Dept. of Psychol., Wales Univ., Bangor, UK
  • Volume
    4
  • fYear
    1994
  • fDate
    27 Jun-2 Jul 1994
  • Firstpage
    2291
  • Abstract
    Simulation results show that DARNET, a network model that learns using a gradient-descent procedure to perform single-trial learning, can make efficient use of whatever number of memory trace vector elements it is provided with. However, the effective maximum capacity of the system is determined by the architecture of the model to be the number of items that can be stored in a composite memory trace vector of fixed dimensionality, the optimal size depending on the number of elements in each to-be-associated input vector. DARNET has previously been shown to provide a good account of some relevant psychological data and the present work adds to the authors\´ understanding of the constraints governing its memory capacity and ability to "learn-to-learn"
  • Keywords
    learning (artificial intelligence); neural nets; DARNET; gradient-descent procedure; maximum capacity; memory capacity; memory vector size; single-trial learning; Convolution; Encoding; Humans; Intelligent networks; Interference; Learning systems; Performance evaluation; Psychology; Testing;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Neural Networks, 1994. IEEE World Congress on Computational Intelligence., 1994 IEEE International Conference on
  • Conference_Location
    Orlando, FL
  • Print_ISBN
    0-7803-1901-X
  • Type

    conf

  • DOI
    10.1109/ICNN.1994.374576
  • Filename
    374576