DocumentCode
301296
Title
Fusing completing models to improve accuracy
Author
Elder, John F.
Author_Institution
Rice Univ., USA
Volume
1
fYear
1995
fDate
22-25 Oct 1995
Firstpage
149
Abstract
Inductive modeling or “machine learning” algorithms are able to discover structure in high-dimensional data in a nearly automated fashion. These adaptive statistical methods-from decision trees and polynomial networks, to projection pursuit models, additive networks, and cascade correlation neural networks-repeatedly search for, and add on, the model component judged best at that state. Because of the huge model space of possible components, the choice is typical greedy. In fact, it is usual for the analyst and algorithm to be greedy at three levels: when choosing: 1) a term within a model, 2) a model within a family (class of method), and 3) a family within a collection of techniques. It is better in each stage, we argue, to “take a longer view” to: 1) consider terms in larger sets, 2) merge competing models within a family, and 3) to fuse information from disparate models, making the combination more robust. Example benefits of fusion are demonstrated on a challenging classification dataset, where one must infer the species of a bat from its chirps
Keywords
Algorithm design and analysis; Chirp; Decision trees; Fuses; Machine learning algorithms; Neural networks; Polynomials; Protection; Robustness; Training data;
fLanguage
English
Publisher
ieee
Conference_Titel
Systems, Man and Cybernetics, 1995. Intelligent Systems for the 21st Century., IEEE International Conference on
Conference_Location
Vancouver, BC
Print_ISBN
0-7803-2559-1
Type
conf
DOI
10.1109/ICSMC.1995.537749
Filename
537749
Link To Document