Title :
A decomposition method for large-scale sparse coding in representation learning
Author :
Yifeng Li ; Caron, Richard J. ; Ngom, Alioune
Author_Institution :
Child & Family Res. Inst., Univ. of British Columbia, Vancouver, BC, Canada
Abstract :
In representation learning, sparse representation is a parsimonious principle that a sample can be approximated by a sparse superposition of dictionary atoms. Sparse coding is the core of this technique. Since the dictionary is often redundant, the dictionary size can be very large. Many optimization methods have been proposed in the literature for sparse coding. However, the efficiency of the optimization for a tremendous number of dictionary atoms is still a bottleneck. In this paper, we propose to use decomposition method for large-scale sparse coding models. Our experimental results show that our method is very efficient.
Keywords :
encoding; learning (artificial intelligence); quadratic programming; decomposition method; large-scale sparse coding models; parsimonious principle; quadratic programming; representation learning; sparse dictionary atoms superposition; sparse representation; Computational modeling; Dictionaries; Encoding; Equations; Mathematical model; Optimization; Vectors;
Conference_Titel :
Neural Networks (IJCNN), 2014 International Joint Conference on
Conference_Location :
Beijing
Print_ISBN :
978-1-4799-6627-1
DOI :
10.1109/IJCNN.2014.6889394