• DocumentCode
    671544
  • Title

    A study of transformation-invariances of deep belief networks

  • Author

    Zheng Shou ; Yuhao Zhang ; Cai, H.J.

  • Author_Institution
    Int. Sch. of Software, Wuhan Univ., Wuhan, China
  • fYear
    2013
  • fDate
    4-9 Aug. 2013
  • Firstpage
    1
  • Lastpage
    8
  • Abstract
    In order to learn transformation-invariant features, several effective deep architectures like hierarchical feature learning and variant Deep Belief Networks (DBN) have been proposed. Considering the complexity of those variants, people are interested in whether DBN itself has transformation-invariances. First of all, we use original DBN to test original data. Almost same error rates will be achieved, if we change weights in the bottom interlayer according to transformations occurred in testing data. It implies that weights in the bottom interlayer can store the knowledge to handle transformations such as rotation, shifting, and scaling. Along with the continuous learning ability and good storage of DBN, we present our Weight-Transformed Training Algorithm (WTTA) without augmenting other layers, units or filters to original DBN. Based upon original training method, WTTA is aiming at transforming weights and is still unsupervised. For MNIST handwritten digits recognizing experiments, we adopted 784-100-100-100 DBN to compare the differences of recognizing ability in weights-transformed ranges. Most error rates generated by WTTA were below 25% while most rates generated by original training algorithm exceeded 25%. Then we also did an experiment on part of MIT-CBCL face database, with varying illumination, and the best testing accuracy can be achieved is 87.5%. Besides, similar results can be achieved by datasets covering all kinds of transformations, but WTTA only needs original training data and transform weights after each training loop. Consequently, we can mine inherent transformation-invariances of DBN by WTTA, and DBN itself can recognize transformed data at satisfying error rates without inserting other components.
  • Keywords
    belief networks; feature extraction; handwritten character recognition; lighting; neural nets; object recognition; probability; unsupervised learning; DBN; MIT-CBCL face database; MNIST handwritten digits recognizing experiments; WTTA; deep belief networks; hierarchical feature learning; illumination variation; multilayer probabilistic generative model; robust feature extraction models; rotation transformation; scaling transformation; shifting transformation; training loop; transformation-invariances; transformation-invariant features; unsupervised learning algorithm; weight-transformed training algorithm; Error analysis; Feature extraction; Neurons; Pattern recognition; Training; Transforms;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Neural Networks (IJCNN), The 2013 International Joint Conference on
  • Conference_Location
    Dallas, TX
  • ISSN
    2161-4393
  • Print_ISBN
    978-1-4673-6128-6
  • Type

    conf

  • DOI
    10.1109/IJCNN.2013.6706884
  • Filename
    6706884