DocumentCode
768087
Title
On-line learning with minimal degradation in feedforward networks
Author
De Angulo, Vincente Ruiz ; Torras, Carme
Author_Institution
Joint Res. Centre, Comm. of the Eur. Communities, Ispra, Italy
Volume
6
Issue
3
fYear
1995
fDate
5/1/1995 12:00:00 AM
Firstpage
657
Lastpage
668
Abstract
Dealing with nonstationary processes requires quick adaptation while at the same time avoiding catastrophic forgetting. A neural learning technique that satisfies these requirements, without sacrificing the benefits of distributed representations, is presented. It relies on a formalization of the problem as the minimization of the error over the previously learned input-output patterns, subject to the constraint of perfect encoding of the new pattern. Then this constrained optimization problem is transformed into an unconstrained one with hidden-unit activations as variables. This new formulation leads to an algorithm for solving the problem, which we call learning with minimal degradation (LMD). Some experimental comparisons of the performance of LMD with backpropagation are provided which, besides showing the advantages of using LMD, reveal the dependence of forgetting on the learning rate in backpropagation. We also explain why overtraining affects forgetting and fault tolerance, which are seen as related problems
Keywords
backpropagation; encoding; feedforward neural nets; optimisation; real-time systems; backpropagation; constrained optimization; encoding; feedforward neural networks; forgetting factor; hidden-unit activations; minimal degradation learning; online learning; Adaptive systems; Backpropagation algorithms; Constraint optimization; Degradation; Encoding; Fault tolerance; Informatics; Intelligent networks; Interference; Systems engineering and theory;
fLanguage
English
Journal_Title
Neural Networks, IEEE Transactions on
Publisher
ieee
ISSN
1045-9227
Type
jour
DOI
10.1109/72.377971
Filename
377971
Link To Document