Title :
PANFIS: A Novel Incremental Learning Machine
Author :
Pratama, Mahardhika ; Anavatti, Sreenatha G. ; Angelov, Plamen P. ; Lughofer, Edwin
Author_Institution :
Sch. of Eng. & Inf. Technol., Univ. of New South Wales, Canberra, ACT, Australia
Abstract :
Most of the dynamics in real-world systems are compiled by shifts and drifts, which are uneasy to be overcome by omnipresent neuro-fuzzy systems. Nonetheless, learning in nonstationary environment entails a system owning high degree of flexibility capable of assembling its rule base autonomously according to the degree of nonlinearity contained in the system. In practice, the rule growing and pruning are carried out merely benefiting from a small snapshot of the complete training data to truncate the computational load and memory demand to the low level. An exposure of a novel algorithm, namely parsimonious network based on fuzzy inference system (PANFIS), is to this end presented herein. PANFIS can commence its learning process from scratch with an empty rule base. The fuzzy rules can be stitched up and expelled by virtue of statistical contributions of the fuzzy rules and injected datum afterward. Identical fuzzy sets may be alluded and blended to be one fuzzy set as a pursuit of a transparent rule base escalating human´s interpretability. The learning and modeling performances of the proposed PANFIS are numerically validated using several benchmark problems from real-world or synthetic datasets. The validation includes comparisons with state-of-the-art evolving neuro-fuzzy methods and showcases that our new method can compete and in some cases even outperform these approaches in terms of predictive fidelity and model complexity.
Keywords :
computational complexity; fuzzy neural nets; fuzzy reasoning; fuzzy set theory; knowledge based systems; learning (artificial intelligence); statistical analysis; PANFIS; autonomous rule base assembling; computational load; degree of nonlinearity; empty rule base; fuzzy rule stitching; fuzzy sets; human interpretability escalation; incremental machine learning performance; memory demand; model complexity; modeling performance; nonstationary environment; omnipresent neurofuzzy system; parsimonious network based on fuzzy inference system; predictive fidelity; scratch; statistical fuzzy rule; transparent rule base; Covariance matrices; Ellipsoids; Fuzzy logic; Fuzzy sets; Fuzzy systems; Training; Vectors; Evolving neuro-fuzzy systems (ENFSs); incremental learning; sample-wise training;
Journal_Title :
Neural Networks and Learning Systems, IEEE Transactions on
DOI :
10.1109/TNNLS.2013.2271933