Title :
Nonlinear Metric Learning with Kernel Density Estimation
Author :
Yujie He ; Yi Mao ; Wenlin Chen ; Yixin Chen
Author_Institution :
Dept. of Comput. Sci. & Eng., Washington Univ., St. Louis, MO, USA
Abstract :
Metric learning, the task of learning a good distance metric, is a key problem in machine learning with ample applications. This paper introduces a novel framework for nonlinear metric learning, called kernel density metric learning (KDML), which is easy to use and provides nonlinear, probability-based distance measures. KDML constructs a direct nonlinear mapping from the original input space into a feature space based on kernel density estimation. The nonlinear mapping in KDML embodies established distance measures between probability density functions, and leads to accurate classification on datasets for which existing linear metric learning methods would fail. It addresses the severe challenge to distance-based classifiers when features are from heterogeneous domains and, as a result, the Euclidean or Mahalanobis distance between original feature vectors is not meaningful. We also propose two ways to determine the kernel bandwidths, including an adaptive local scaling approach and an integrated optimization algorithm that learns the Mahalanobis matrix and kernel bandwidths together. KDML is a general framework that can be combined with any existing metric learning algorithm. As concrete examples, we combine KDML with two leading metric learning algorithms, large margin nearest neighbors (LMNN) and neighborhood component analysis (NCA). KDML can naturally handle not only numerical features, but also categorical ones, which is rarely found in previous metric learning algorithms. Extensive experimental results on various datasets show that KDML significantly improves existing metric learning algorithms in terms of classification accuracy.
Keywords :
learning (artificial intelligence); matrix algebra; optimisation; pattern classification; probability; Euclidean distance; KDML; LMNN; Mahalanobis distance; Mahalanobis matrix; NCA; adaptive local scaling approach; dataset classification; direct nonlinear mapping; distance metric; distance-based classifiers; feature space; integrated optimization algorithm; kernel bandwidths; kernel density estimation; kernel density metric learning; large margin nearest neighbors; linear metric learning methods; machine learning; neighborhood component analysis; nonlinear metric learning; nonlinear-based distance measure; probability density functions; probability-based distance measure; Algorithm design and analysis; Density measurement; Euclidean distance; Kernel; Learning systems; Vectors; Classification; classification; kernel density estimation; large margin nearest neighbors; metric learning; neighborhood components analysis;
Journal_Title :
Knowledge and Data Engineering, IEEE Transactions on
DOI :
10.1109/TKDE.2014.2384522