Title :
Learning linear threshold functions
Author_Institution :
Div. of Math. Comput. Sci. & Stat., Texas Univ., San Antonio, TX, USA
Abstract :
This paper surveys recent theoretical results on algorithms for learning linear threshold functions (aka perceptrons, linear discriminants). When the examples are linearly separable, Littlestone has introduced algorithms that are much faster than the perceptron algorithm when there are many irrelevant attributes. When the examples are not linearly separable, Gallant´s Ratchet algorithm will converge to the optimal linear threshold function, albeit slowly, and Bylander´s perceptron-noise algorithm will converge quickly to optimal in the special case of classification noise. Empirical comparisons on some nonlinearly-separable real-world datasets are presented
Keywords :
computational complexity; learning (artificial intelligence); perceptrons; Bylander perceptron-noise algorithm; Gallant´s Ratchet algorithm; Littlestone algorithm; aka perceptrons; convergence; linear discriminants; linear threshold functions; linear threshold learning; machine learning; Algorithm design and analysis; Computer science; Convergence; Linear programming; Machine learning; Machine learning algorithms; Mathematics; Neural networks; Nonhomogeneous media; Statistics;
Conference_Titel :
Systems, Man, and Cybernetics, 1994. Humans, Information and Technology., 1994 IEEE International Conference on
Conference_Location :
San Antonio, TX
Print_ISBN :
0-7803-2129-4
DOI :
10.1109/ICSMC.1994.400002