DocumentCode
948002
Title
Reduced Pattern Training Based on Task Decomposition Using Pattern Distributor
Author
Guan, Sheng-Uei ; Bao, Chunyu ; Neo, TseNgee
Author_Institution
Brunel Univ., Uxbridge
Volume
18
Issue
6
fYear
2007
Firstpage
1738
Lastpage
1749
Abstract
Task decomposition with pattern distributor (PD) is a new task decomposition method for multilayered feedforward neural networks (NNs). Pattern distributor network is proposed that implements this new task decomposition method. We propose a theoretical model to analyze the performance of pattern distributor network. A method named reduced pattern training (RPT) is also introduced, aiming to improve the performance of pattern distribution. Our analysis and the experimental results show that RPT improves the performance of pattern distributor network significantly. The distributor module´s classification accuracy dominates the whole network´s performance. Two combination methods, namely, crosstalk-based combination and genetic-algorithm (GA)-based combination, are presented to find suitable grouping for the distributor module. Experimental results show that this new method can reduce training time and improve network generalization accuracy when compared to a conventional method such as constructive backpropagation or a task decomposition method such as output parallelism (OP).
Keywords
feedforward neural nets; genetic algorithms; learning (artificial intelligence); crosstalk-based combination; genetic-algorithm-based combination; multilayered feedforward neural networks; pattern distributor network; reduced pattern training; task decomposition method; Crosstalk-based combination; full pattern training (FPT); genetic-algorithm-based combination; pattern distributor; reduced pattern training (RPT); task decomposition;
fLanguage
English
Journal_Title
Neural Networks, IEEE Transactions on
Publisher
ieee
ISSN
1045-9227
Type
jour
DOI
10.1109/TNN.2007.899711
Filename
4359177
Link To Document