DocumentCode
288324
Title
Improving generalization performance by information minimization
Author
Kamimura, Ryotaro ; Takagi, Toshiyuki ; Nakanishi, Shohachiro
Author_Institution
Inf. Sci. Lab., Tokai Univ., Kanagawa, Japan
Volume
1
fYear
1994
fDate
27 Jun-2 Jul 1994
Firstpage
143
Abstract
In this paper, we attempt to show that the information stored in networks must be as small as possible for the improvement of the generalization performance under the condition that the networks can produce targets with appropriate accuracy. The information is defined by the difference between maximum entropy or uncertainty and observed entropy. Borrowing a definition of fuzzy entropy, the uncertainty function is defined for the internal representation and represented by the equation: -υi log υi-(1-υi ) log (1-υi), where υi is a hidden unit activity. After having formulated an update rule for the minimization of the information, we applied the method to a problem of language acquisition: the inference of the past tense forms of regular verbs. Experimental results confirmed that by our method, the information was significantly decreased and the generalization performance was greatly improved
Keywords
generalisation (artificial intelligence); grammars; inference mechanisms; information theory; knowledge acquisition; maximum entropy methods; minimisation; natural languages; neural nets; uncertainty handling; fuzzy entropy; generalization performance improvement; hidden unit activity; inference; information minimization; language acquisition; maximum entropy; neural nets; observed entropy; past tense; uncertainty function; verbs; Entropy; Equations; Information science; Laboratories; Minimization methods; Uncertainty;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks, 1994. IEEE World Congress on Computational Intelligence., 1994 IEEE International Conference on
Conference_Location
Orlando, FL
Print_ISBN
0-7803-1901-X
Type
conf
DOI
10.1109/ICNN.1994.374153
Filename
374153
Link To Document