DocumentCode
671599
Title
Maximal margin learning vector quantisation
Author
Trung Le ; Dat Tran ; Van Nguyen ; Wanli Ma
Author_Institution
Fac. of Inf. Technol., HCMc Univ. of Pedagogy, Ho Chi Minh City, Vietnam
fYear
2013
fDate
4-9 Aug. 2013
Firstpage
1
Lastpage
6
Abstract
Kernel Generalised Learning Vector Quantisation (KGLVQ) was proposed to extend Generalised Learning Vector Quantisation into the kernel feature space to deal with complex class boundaries and thus yielded promising performance for complex classification tasks in pattern recognition. However KGLVQ does not follow the maximal margin principle, which is crucial for kernel-based learning methods. In this paper we propose a maximal margin approach (MLVQ) to the KGLVQ algorithm. MLVQ inherits the merits of KGLVQ and also follows the maximal margin principle to improve the generalisation capability. Experiments performed on the well-known data sets available in UCI repository show promising classification results for the proposed method.
Keywords
learning (artificial intelligence); pattern classification; vector quantisation; KGLVQ algorithm; UCI repository; complex classification tasks; kernel feature space; kernel generalised learning vector quantisation; maximal margin learning vector quantisation; pattern recognition; Kernel; Linear programming; Prototypes; Support vector machines; Training; Vector quantization; Vectors;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks (IJCNN), The 2013 International Joint Conference on
Conference_Location
Dallas, TX
ISSN
2161-4393
Print_ISBN
978-1-4673-6128-6
Type
conf
DOI
10.1109/IJCNN.2013.6706940
Filename
6706940
Link To Document