• DocumentCode
    445969
  • Title

    A simple hierarchical approximation RBF neural network

  • Author

    Doerschuk, Peggy Israel ; Pawaskar, Sainath Shrikant

  • Author_Institution
    Dept. of Comput. Sci., Lamar Univ., Beaumont, TX, USA
  • Volume
    3
  • fYear
    2005
  • fDate
    31 July-4 Aug. 2005
  • Firstpage
    1389
  • Abstract
    The approximation algorithm introduced by Asim Roy et al. (1997) generates a hybrid neural network with RBF neurons and other types of hidden neurons for function approximation. The network is trained in stages, with RBF neurons at the early stages corresponding to general features in the space and those in later stages corresponding to more specific features. The other types of hidden neurons are added with a view to improving generalization and reducing the number of RBF neurons. The algorithm uses linear programming to design and train the hybrid network. We investigate simplifying the algorithm with a view to eliminating the need for the other types of hidden neurons and linear programming. The simple hierarchical approximation algorithm (´SHA´) achieves comparable results in terms of accuracy without the added complexity introduced by the other types of hidden neurons.
  • Keywords
    function approximation; learning (artificial intelligence); linear programming; radial basis function networks; function approximation; linear programming; radial basis function neural network; radial basis function neuron; simple hierarchical approximation; Algorithm design and analysis; Approximation algorithms; Computer science; Electronic mail; Function approximation; Hybrid power systems; Linear programming; Neural networks; Neurons; Testing;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Neural Networks, 2005. IJCNN '05. Proceedings. 2005 IEEE International Joint Conference on
  • Print_ISBN
    0-7803-9048-2
  • Type

    conf

  • DOI
    10.1109/IJCNN.2005.1556077
  • Filename
    1556077