Title :
Two-Stage Orthogonal Least Squares Methods for Neural Network Construction
Author :
Long Zhang ; Kang Li ; Er-Wei Bai ; Irwin, George W.
Author_Institution :
Sch. of Electron., Electr. Eng. & Comput. Sci., Queen´s Univ. Belfast, Belfast, UK
Abstract :
A number of neural networks can be formulated as the linear-in-the-parameters models. Training such networks can be transformed to a model selection problem where a compact model is selected from all the candidates using subset selection algorithms. Forward selection methods are popular fast subset selection approaches. However, they may only produce suboptimal models and can be trapped into a local minimum. More recently, a two-stage fast recursive algorithm (TSFRA) combining forward selection and backward model refinement has been proposed to improve the compactness and generalization performance of the model. This paper proposes unified two-stage orthogonal least squares methods instead of the fast recursive-based methods. In contrast to the TSFRA, this paper derives a new simplified relationship between the forward and the backward stages to avoid repetitive computations using the inherent orthogonal properties of the least squares methods. Furthermore, a new term exchanging scheme for backward model refinement is introduced to reduce computational demand. Finally, given the error reduction ratio criterion, effective and efficient forward and backward subset selection procedures are proposed. Extensive examples are presented to demonstrate the improved model compactness constructed by the proposed technique in comparison with some popular methods.
Keywords :
least squares approximations; neural nets; set theory; TSFRA; backward model refinement; backward subset selection procedures; error reduction ratio criterion; forward selection methods; forward subset selection procedures; linear-in-the-parameter models; model selection problem; neural network construction; orthogonal properties; recursive-based methods; suboptimal models; two-stage fast recursive algorithm; two-stage orthogonal least square methods; unified two-stage orthogonal least square methods; Computational modeling; Cost function; Least squares methods; Matching pursuit algorithms; Neural networks; Numerical models; Vectors; Backward model refinement; computational complexity; forward selection; linear-in-the-parameters model; orthogonal least square (OLS);
Journal_Title :
Neural Networks and Learning Systems, IEEE Transactions on
DOI :
10.1109/TNNLS.2014.2346399