Title :
D-entropy minimization: integration of mutual information maximization and minimization
Author :
Kamimura, Ryotaro
Author_Institution :
Inf. Sci. Lab., Tokai Univ., Kanagawa, Japan
Abstract :
In this paper, we propose a D-entropy minimisation method, aiming to unify information maximization and minimization methods. The D-entropy is defined by difference between Renyi´s entropy and Shannon´s entropy. The D-entropy minimization corresponds to both mutual information maximization and minimization. Thus, the method can be used to interpret explicitly internal representations and to improve generalization. The D-entropy minimization was applied to two problems: six alphabet character recognition and the inference of well-formedness of artificial strings. Experimental results confirmed that by minimizing the D-entropy a small number of principal hidden units can be detected and generalization performance can be improved
Keywords :
character recognition; generalisation (artificial intelligence); information theory; learning (artificial intelligence); minimisation; minimum entropy methods; neural nets; probability; D-entropy minimization; Renyi entropy; Shannon entropy; character recognition; generalization; information maximization; information minimization; neural learning; neural nets; probability; Artificial neural networks; Character recognition; Entropy; Information science; Laboratories; Minimization methods; Mutual information; Neural networks; Noise reduction; Uncertainty;
Conference_Titel :
Neural Networks,1997., International Conference on
Conference_Location :
Houston, TX
Print_ISBN :
0-7803-4122-8
DOI :
10.1109/ICNN.1997.616174