Please use this identifier to cite or link to this item: https://hdl.handle.net/2440/72071
Type: Conference paper
Title: Efficiently learning a distance metric for large margin nearest neighbor classification
Author: Park, K.
Shen, C.
Hao, Z.
Kim, J.
Citation: Proceedings of the Twenty-Fifth AAAI Conference on Artificial Intelligence and the Twenty-Third Innovative Applications of Artificial Intelligence Conference, 7-11 August 2011, San Francisco, California, USA: pp.453-458
Publisher: AAAI Press
Publisher Place: online
Issue Date: 2011
ISBN: 9781577355083
Conference Name: AAAI Conference on Artificial Intelligence (25th : 2011 : San Francisco, U.S.A.)
Statement of
Responsibility: 
K. Park, C. Shen, Z. Hao, J. Kim
Abstract: We concern the problem of learning a Mahalanobis distance metric for improving nearest neighbor classification. Our work is built upon the large margin nearest neighbor (LMNN) classification framework. Due to the semidefiniteness constraint in the optimization problem of LMNN, it is not scalable in terms of the dimensionality of the input data. The original LMNN solver partially alleviates this problem by adopting alternating projection methods instead of standard interior-point methods. Still, at each iteration, the computation complexity is at least O(D3) (D is the dimension of input data). In this work, we propose a column generation based algorithm to solve the LMNN optimization problem much more efficiently. Our algorithm is much more scalable in that at each iteration, it does not need full eigen-decomposition. Instead, we only need to find the leading eigenvalue and its corresponding eigenvector, which is of O(D2) complexity. Experiments show the efficiency and efficacy of our algorithms.
Rights: Copyright status unknown
Appears in Collections:Aurora harvest 5
Computer Science publications

Files in This Item:
File Description SizeFormat 
RA_hdl_72071.pdf
  Restricted Access
Restricted Access533.17 kBAdobe PDFView/Open


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.