Title :
Fault tolerant learning using Kullback-Leibler divergence
Author :
Sum, John ; Leung, Chi-Sing ; Hsu, Lipin
Author_Institution :
Nat. Chung Hsing Univ., Taichung
fDate :
Oct. 30 2007-Nov. 2 2007
Abstract :
In this paper, an objective function for training a fault tolerant neural network is derived based on the idea of Kullback-Leibler (KL) divergence. The new objective function is then applied to a radial basis function (RBF) network that is with multiplicative weight noise. Simulation results have demonstrated that the RBF network trained in accordance with the new objective function is of better fault tolerance ability, in compared with the one trained by explicit regularization. As KL divergence has relation to Bayesian learning, a discussion on the proposed objective function and the other Bayesian type objective functions is discussed.
Keywords :
Bayes methods; fault tolerant computing; radial basis function networks; Bayesian type objective function; Kullback-Leibler divergence; fault tolerant neural network; multiplicative weight noise; radial basis function network; Additive white noise; Bayesian methods; Biomedical engineering; Electronic commerce; Fault tolerance; Multilayer perceptrons; Neural network hardware; Neural networks; Radial basis function networks; Signal to noise ratio;
Conference_Titel :
TENCON 2007 - 2007 IEEE Region 10 Conference
Conference_Location :
Taipei
Print_ISBN :
978-1-4244-1272-3
Electronic_ISBN :
978-1-4244-1272-3
DOI :
10.1109/TENCON.2007.4429073