Fast solvers and efficient implementations for distance metric learning

  • Weinberger K
  • Saul L
  • 169


    Mendeley users who have this article in their library.
  • 173


    Citations of this article.


In this paper we study how to improve nearest neighbor classification by learning a Mahalanobis distance metric. We build on a recently proposed framework for distance metric learning known as large margin nearest neighbor (LMNN) classification. Our paper makes three contributions. First, we describe a highly efficient solver for the particular instance of semidefinite programming that arises in LMNN classification; our solver can handle problems with billions of large margin constraints in a few hours. Second, we show how to reduce both training and testing times using metric ball trees; the speedups from ball trees are further magnified by learning low dimensional representations of the input space. Third, we show how to learn different Mahalanobis distance metrics in different parts of the input space. For large data sets, the use of locally adaptive distance metrics leads to even lower error rates.

Get free article suggestions today

Mendeley saves you time finding and organizing research

Sign up here
Already have an account ?Sign in

Find this document


  • Kilian Q. Weinberger

  • Lawrence K. Saul

Cite this document

Choose a citation style from the tabs below

Save time finding and organizing research with Mendeley

Sign up for free