• DocumentCode
    1797447
  • Title

    An improved RBM based on Bayesian Regularization

  • Author

    Guangyuan Pan ; Junfei Qiao ; Wei Chai ; Dimopoulos, Nikitas

  • Author_Institution
    Inst. of Intell. Syst., Beijing Univ. of Technol., Beijing, China
  • fYear
    2014
  • fDate
    6-11 July 2014
  • Firstpage
    2935
  • Lastpage
    2939
  • Abstract
    Restricted Boltzmann Machine is a fundamental method in deep learning networks. Training and generalization is an ill-defined problem in that many different networks may achieve the training goal; however each will respond differently to an unknown input. Traditional approaches include stopping the training early and/or restricting the size of the network These approaches ameliorate the problem of over-fitting where the network learns the patterns presented but is unable to generalize. Bayesian regularization addresses these issues by requiring the weights of the network to attain a minimum magnitude. This ensures that non-contributing weights are reduced significantly and the resulting network represents the essence of the inter-relations of the training. Bayesian Regularization simply introduces an additional term to the objective function. This term comprises the sum of the squares of the weights. The optimization process therefore not only achieves the objective of the original cost (i.e. the minimization of an error metric) but it also ensures that this objective is achieved with minimum-magnitude weights. We have introduced Bayesian Regularization in the training of Restricted Boltzmann Machines and have applied this method in experiments of hand-written numbers classification. Our experiments showed that by adding Bayesian regularization in the training of RBMs, we were able to improve the generalization capabilities of the trained network by reducing its recognition errors by more than 1.6%.
  • Keywords
    Boltzmann machines; belief networks; generalisation (artificial intelligence); learning (artificial intelligence); Bayesian regularization; deep learning networks; hand-written number classification; improved RBM; minimum-magnitude weights; noncontributing weights; optimization process; restricted Boltzmann machine; Approximation methods; Bayes methods; Educational institutions; Feature extraction; Neurons; Testing; Training; Restricted Boltzmann Machine; classification; over fitting; regularization;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Neural Networks (IJCNN), 2014 International Joint Conference on
  • Conference_Location
    Beijing
  • Print_ISBN
    978-1-4799-6627-1
  • Type

    conf

  • DOI
    10.1109/IJCNN.2014.6889458
  • Filename
    6889458