• DocumentCode
    110191
  • Title

    Efficient Dual Approach to Distance Metric Learning

  • Author

    Chunhua Shen ; Junae Kim ; Fayao Liu ; Lei Wang ; van den Hengel, A.

  • Author_Institution
    Australian Center for Visual Technol., Univ. of Adelaide, Adelaide, SA, Australia
  • Volume
    25
  • Issue
    2
  • fYear
    2014
  • fDate
    Feb. 2014
  • Firstpage
    394
  • Lastpage
    406
  • Abstract
    Distance metric learning is of fundamental interest in machine learning because the employed distance metric can significantly affect the performance of many learning methods. Quadratic Mahalanobis metric learning is a popular approach to the problem, but typically requires solving a semidefinite programming (SDP) problem, which is computationally expensive. The worst case complexity of solving an SDP problem involving a matrix variable of size D×D with O (D) linear constraints is about O(D6.5) using interior-point methods, where D is the dimension of the input data. Thus, the interior-point methods only practically solve problems exhibiting less than a few thousand variables. Because the number of variables is D (D+1)/2, this implies a limit upon the size of problem that can practically be solved around a few hundred dimensions. The complexity of the popular quadratic Mahalanobis metric learning approach thus limits the size of problem to which metric learning can be applied. Here, we propose a significantly more efficient and scalable approach to the metric learning problem based on the Lagrange dual formulation of the problem. The proposed formulation is much simpler to implement, and therefore allows much larger Mahalanobis metric learning problems to be solved. The time complexity of the proposed method is roughly O (D3), which is significantly lower than that of the SDP approach. Experiments on a variety of data sets demonstrate that the proposed method achieves an accuracy comparable with the state of the art, but is applicable to significantly larger problems. We also show that the proposed method can be applied to solve more general Frobenius norm regularized SDP problems approximately.
  • Keywords
    learning (artificial intelligence); mathematical programming; matrix algebra; Lagrange dual formulation; distance metric learning; general Frobenius norm regularized SDP problems; interior-point methods; linear constraints; machine learning; matrix variable; quadratic Mahalanobis metric learning approach; semidefinite programming problem; Complexity theory; Learning systems; Linear programming; Measurement; Optimization; Symmetric matrices; Training; Convex optimization; Lagrange duality; Mahalanobis distance; metric learning; semidefinite programming (SDP);
  • fLanguage
    English
  • Journal_Title
    Neural Networks and Learning Systems, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    2162-237X
  • Type

    jour

  • DOI
    10.1109/TNNLS.2013.2275170
  • Filename
    6588959