• DocumentCode
    1930516
  • Title

    Combining evolving neural network classifiers using bagging

  • Author

    Sohn, Sunghwan ; Dagli, Cihan H.

  • Author_Institution
    Dept. of Eng. Manage., Missouri Univ., Rolla, MO, USA
  • Volume
    4
  • fYear
    2003
  • fDate
    20-24 July 2003
  • Firstpage
    3218
  • Abstract
    The performance of the neural network classifier significantly depends on its architecture and generalization. It is usual to find the proper architecture by trial and error. This is time consuming and may not always find the optimal network. For this reason, we apply genetic algorithms to the automatic generation of neural networks. Many researchers have provided that combining multiple classifiers improves generalization. One of the most effective combining methods is bagging. In bagging, training sets are selected by resampling from the original training set and classifiers trained with these sets are combined by voting. We implement the bagging technique into the training of evolving neural network classifiers to improve generalization.
  • Keywords
    generalisation (artificial intelligence); genetic algorithms; learning (artificial intelligence); neural net architecture; pattern classification; bagging; evolving neural network classifiers; generalization; genetic algorithms; neural net training; Algorithm design and analysis; Bagging; Computer architecture; Genetic algorithms; Neural networks; Research and development management; Robustness; Speech recognition; Systems engineering and theory; Voting;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Neural Networks, 2003. Proceedings of the International Joint Conference on
  • ISSN
    1098-7576
  • Print_ISBN
    0-7803-7898-9
  • Type

    conf

  • DOI
    10.1109/IJCNN.2003.1224088
  • Filename
    1224088