DocumentCode :
2770177
Title :
Supervised Information Maximization by Weighted Distance
Author :
Kamimura, Ryotaro
Author_Institution :
Tokai Univ., Hiratsuka
fYear :
0
fDate :
0-0 0
Firstpage :
1790
Lastpage :
1796
Abstract :
In this paper, we propose a method to extend information-theoretic competitive learning to supervised competitive learning. We have shown that information maximization correspond to competition in neurons. However, this information maximization cannot be used to specify which neurons should be winners. Thus, it is impossible to incorporate teacher information in information maximization. For dealing with this teacher information, we use weighted distance between input patterns and connection weights. Even if distance between input patterns and connection weights is not so small, the distance are made smaller by the parameter considering teacher information. By this weighted distance, we can naturally incorporate teacher information and extend unsupervised competitive learning to supervised information-theoretic competitive learning.
Keywords :
unsupervised learning; information-theoretic competitive learning; supervised competitive learning; supervised information maximization; unsupervised competitive learning; weighted distance; Entropy; Euclidean distance; Information science; Information technology; Information theory; Laboratories; Mutual information; Neural networks; Neurons; Supervised learning; Gaussian function; competition; entropy maximization; guide; mutual information maximization; weighted distance;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 2006. IJCNN '06. International Joint Conference on
Conference_Location :
Vancouver, BC
Print_ISBN :
0-7803-9490-9
Type :
conf
DOI :
10.1109/IJCNN.2006.246896
Filename :
1716326
Link To Document :
بازگشت