Working hard to know your neighbor's margins: Local descriptor learning loss

A. Mishchuk, D. Mishkin, F. Radenovic, J. Matas. Working hard to know your neighbor's margins: Local descriptor learning loss. pages online, 12, 2017.

Autoren
  • Anastasiya Mishchuk
  • Dmytro Mishkin
  • Filip Radenovic
  • Jiri Matas
BuchAdvances in Neural Information Processing Systems 30 - Proc. NIPS 2017
TypIn Konferenzband
Monat12
Jahr2017
Seitenonline
Abstract

We introduce a loss for metric learning, which is inspired by the Lowe’s matching criterion for SIFT. We show that the proposed loss, that maximizes the distance between the closest positive and closest negative example in the batch, is better than complex regularization methods; it works well for both shallow and deep convolution network architectures. Applying the novel loss to the L2Net CNN architecture results in a compact descriptor named HardNet. It has the same dimensionality as SIFT (128) and shows state-of-art performance in wide baseline stereo, patch verification and instance retrieval benchmarks.