DocumentCode
110191
Title
Efficient Dual Approach to Distance Metric Learning
Author
Chunhua Shen ; Junae Kim ; Fayao Liu ; Lei Wang ; van den Hengel, A.
Author_Institution
Australian Center for Visual Technol., Univ. of Adelaide, Adelaide, SA, Australia
Volume
25
Issue
2
fYear
2014
fDate
Feb. 2014
Firstpage
394
Lastpage
406
Abstract
Distance metric learning is of fundamental interest in machine learning because the employed distance metric can significantly affect the performance of many learning methods. Quadratic Mahalanobis metric learning is a popular approach to the problem, but typically requires solving a semidefinite programming (SDP) problem, which is computationally expensive. The worst case complexity of solving an SDP problem involving a matrix variable of size D×D with O (D) linear constraints is about O(D6.5) using interior-point methods, where D is the dimension of the input data. Thus, the interior-point methods only practically solve problems exhibiting less than a few thousand variables. Because the number of variables is D (D+1)/2, this implies a limit upon the size of problem that can practically be solved around a few hundred dimensions. The complexity of the popular quadratic Mahalanobis metric learning approach thus limits the size of problem to which metric learning can be applied. Here, we propose a significantly more efficient and scalable approach to the metric learning problem based on the Lagrange dual formulation of the problem. The proposed formulation is much simpler to implement, and therefore allows much larger Mahalanobis metric learning problems to be solved. The time complexity of the proposed method is roughly O (D3), which is significantly lower than that of the SDP approach. Experiments on a variety of data sets demonstrate that the proposed method achieves an accuracy comparable with the state of the art, but is applicable to significantly larger problems. We also show that the proposed method can be applied to solve more general Frobenius norm regularized SDP problems approximately.
Keywords
learning (artificial intelligence); mathematical programming; matrix algebra; Lagrange dual formulation; distance metric learning; general Frobenius norm regularized SDP problems; interior-point methods; linear constraints; machine learning; matrix variable; quadratic Mahalanobis metric learning approach; semidefinite programming problem; Complexity theory; Learning systems; Linear programming; Measurement; Optimization; Symmetric matrices; Training; Convex optimization; Lagrange duality; Mahalanobis distance; metric learning; semidefinite programming (SDP);
fLanguage
English
Journal_Title
Neural Networks and Learning Systems, IEEE Transactions on
Publisher
ieee
ISSN
2162-237X
Type
jour
DOI
10.1109/TNNLS.2013.2275170
Filename
6588959
Link To Document