DocumentCode
1797447
Title
An improved RBM based on Bayesian Regularization
Author
Guangyuan Pan ; Junfei Qiao ; Wei Chai ; Dimopoulos, Nikitas
Author_Institution
Inst. of Intell. Syst., Beijing Univ. of Technol., Beijing, China
fYear
2014
fDate
6-11 July 2014
Firstpage
2935
Lastpage
2939
Abstract
Restricted Boltzmann Machine is a fundamental method in deep learning networks. Training and generalization is an ill-defined problem in that many different networks may achieve the training goal; however each will respond differently to an unknown input. Traditional approaches include stopping the training early and/or restricting the size of the network These approaches ameliorate the problem of over-fitting where the network learns the patterns presented but is unable to generalize. Bayesian regularization addresses these issues by requiring the weights of the network to attain a minimum magnitude. This ensures that non-contributing weights are reduced significantly and the resulting network represents the essence of the inter-relations of the training. Bayesian Regularization simply introduces an additional term to the objective function. This term comprises the sum of the squares of the weights. The optimization process therefore not only achieves the objective of the original cost (i.e. the minimization of an error metric) but it also ensures that this objective is achieved with minimum-magnitude weights. We have introduced Bayesian Regularization in the training of Restricted Boltzmann Machines and have applied this method in experiments of hand-written numbers classification. Our experiments showed that by adding Bayesian regularization in the training of RBMs, we were able to improve the generalization capabilities of the trained network by reducing its recognition errors by more than 1.6%.
Keywords
Boltzmann machines; belief networks; generalisation (artificial intelligence); learning (artificial intelligence); Bayesian regularization; deep learning networks; hand-written number classification; improved RBM; minimum-magnitude weights; noncontributing weights; optimization process; restricted Boltzmann machine; Approximation methods; Bayes methods; Educational institutions; Feature extraction; Neurons; Testing; Training; Restricted Boltzmann Machine; classification; over fitting; regularization;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks (IJCNN), 2014 International Joint Conference on
Conference_Location
Beijing
Print_ISBN
978-1-4799-6627-1
Type
conf
DOI
10.1109/IJCNN.2014.6889458
Filename
6889458
Link To Document