• DocumentCode
    2771963
  • Title

    Deep, super-narrow neural network is a universal classifier

  • Author

    Szymanski, Lech ; McCane, Brendan

  • Author_Institution
    Dept. of Comput. Sci., Univ. of Otago, Dunedin, New Zealand
  • fYear
    2012
  • fDate
    10-15 June 2012
  • Firstpage
    1
  • Lastpage
    8
  • Abstract
    Deep architecture models are known to be conducive to good generalisation for certain types of classification tasks. Existing unsupervised and semi-supervised training methods do not explain why and when deep internal representations will be effective. We investigate the fundamental principles of representation in deep architectures by devising a method for binary classification in multi-layer feed forward networks with limited breadth. We show that, given enough layers, a super-narrow neural network, with two neurons per layer, is capable of shattering any separable binary dataset. We also show that datasets that exhibit certain type of symmetries are better suited for deep representation and may require only few hidden layers to produce desired classification.
  • Keywords
    feedforward neural nets; pattern classification; unsupervised learning; binary classification; classification task; deep architecture model; deep architectures representation; deep belief nets; multilayer feedforward neural network; semisupervised training method; separable binary dataset; super narrow neural network; unsupervised training method; Biological neural networks; Computational modeling; Computer architecture; Neurons; Spirals; Training; Vectors;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Neural Networks (IJCNN), The 2012 International Joint Conference on
  • Conference_Location
    Brisbane, QLD
  • ISSN
    2161-4393
  • Print_ISBN
    978-1-4673-1488-6
  • Electronic_ISBN
    2161-4393
  • Type

    conf

  • DOI
    10.1109/IJCNN.2012.6252513
  • Filename
    6252513