Comparing apples to apples in the evaluation of binary coding methods

05/05/2014
by   Mohammad Rastegari, et al.
0

We discuss methodological issues related to the evaluation of unsupervised binary code construction methods for nearest neighbor search. These issues have been widely ignored in literature. These coding methods attempt to preserve either Euclidean distance or angular (cosine) distance in the binary embedding space. We explain why when comparing a method whose goal is preserving cosine similarity to one designed for preserving Euclidean distance, the original features should be normalized by mapping them to the unit hypersphere before learning the binary mapping functions. To compare a method whose goal is to preserves Euclidean distance to one that preserves cosine similarity, the original feature data must be mapped to a higher dimension by including a bias term in binary mapping functions. These conditions ensure the fair comparison between different binary code methods for the task of nearest neighbor search. Our experiments show under these conditions the very simple methods (e.g. LSH and ITQ) often outperform recent state-of-the-art methods (e.g. MDSH and OK-means).

READ FULL TEXT
research
10/06/2019

Exact and/or Fast Nearest Neighbors

Prior methods for retrieval of nearest neighbors in high dimensions are ...
research
08/05/2020

Fast top-K Cosine Similarity Search through XOR-Friendly Binary Quantization on GPUs

We explore the use of GPU for accelerating large scale nearest neighbor ...
research
02/04/2021

The Analysis from Nonlinear Distance Metric to Kernel-based Drug Prescription Prediction System

Distance metrics and their nonlinear variant play a crucial role in mach...
research
12/01/2015

Implicit Sparse Code Hashing

We address the problem of converting large-scale high-dimensional image ...
research
01/14/2019

Search Efficient Binary Network Embedding

Traditional network embedding primarily focuses on learning a dense vect...
research
06/07/2021

The Distance Transform and its Computation

Distance transformation is an image processing technique used for many d...
research
06/09/2022

Neural Bregman Divergences for Distance Learning

Many metric learning tasks, such as triplet learning, nearest neighbor r...

Please sign up or login with your details

Forgot password? Click here to reset