Generative Local Metric Learning for Nearest Neighbor Classificationopen access
- Authors
- Noh, Yung-Kyun; Zhang, Byoung-Tak; Lee, Daniel D.
- Issue Date
- Jan-2018
- Publisher
- IEEE COMPUTER SOC
- Keywords
- Metric learning; nearest neighbor classification; f-divergence; generative-discriminative hybridization
- Citation
- IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, v.40, no.1, pp.106 - 118
- Indexed
- SCIE
SCOPUS
- Journal Title
- IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE
- Volume
- 40
- Number
- 1
- Start Page
- 106
- End Page
- 118
- URI
- https://scholarworks.bwise.kr/hanyang/handle/2021.sw.hanyang/150664
- DOI
- 10.1109/TPAMI.2017.2666151
- ISSN
- 0162-8828
- Abstract
- We consider the problem of learning a local metric in order to enhance the performance of nearest neighbor classification. Conventional metric learning methods attempt to separate data distributions in a purely discriminative manner; here we show how to take advantage of information from parametric generative models. We focus on the bias in the information-theoretic error arising from finite sampling effects, and find an appropriate local metric that maximally reduces the bias based upon knowledge from generative models. As a byproduct, the asymptotic theoretical analysis in this work relates metric learning to dimensionality reduction from a novel perspective, which was not understood from previous discriminative approaches. Empirical experiments show that this learned local metric enhances the discriminative nearest neighbor performance on various datasets using simple class conditional generative models such as a Gaussian.
- Files in This Item
-
Go to Link
- Appears in
Collections - 서울 공과대학 > 서울 컴퓨터소프트웨어학부 > 1. Journal Articles
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.