Compression Network with Transformer for Approximate Nearest Neighbor Search

07/30/2021
by   Haokui Zhang, et al.
0

We propose a generic feature compression method for Approximate Nearest Neighbor Search (ANNS) problems, which speeds up existing ANNS methods in a plug-and-play manner. Specifically, we propose a new network structure called Compression Network with Transformer (CNT) to compress the feature into a low dimensional space, and an inhomogeneous neighborhood relationship preserving (INRP) loss that aims to maintain high search accuracy. In CNT, we use multiple compression projections to cast the feature into many low dimensional spaces, and then use transformer to globally optimize these projections such that the features are well compressed following the guidance from our loss function. The loss function is designed to assign high weights on point pairs that are close in original feature space, and keep their distances in projected space. Keeping these distances helps maintain the eventual top-k retrieval accuracy, and down weighting others creates room for feature compression. In experiments, we run our compression method on public datasets, and use the compressed features in graph based, product quantization and scalar quantization based ANNS solutions. Experimental results show that our compression method can significantly improve the efficiency of these methods while preserves or even improves search accuracy, suggesting its broad potential impact on real world applications.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/15/2016

Scalable Image Retrieval by Sparse Product Quantization

Fast Approximate Nearest Neighbor (ANN) search technique for high-dimens...
research
08/25/2017

Subspace Approximation for Approximate Nearest Neighbor Search in NLP

Most natural language processing tasks can be formulated as the approxim...
research
12/22/2015

Transformed Residual Quantization for Approximate Nearest Neighbor Search

The success of product quantization (PQ) for fast nearest neighbor searc...
research
12/17/2021

Nearest neighbor search with compact codes: A decoder perspective

Modern approaches for fast retrieval of similar vectors on billion-scale...
research
07/25/2022

C3-SL: Circular Convolution-Based Batch-Wise Compression for Communication-Efficient Split Learning

Most existing studies improve the efficiency of Split learning (SL) by c...
research
06/16/2022

Unsupervised Space Partitioning for Nearest Neighbor Search

Approximate Nearest Neighbor Search (ANNS) in high dimensional spaces is...
research
11/29/2021

Feature-Gate Coupling for Dynamic Network Pruning

Gating modules have been widely explored in dynamic network pruning to r...

Please sign up or login with your details

Forgot password? Click here to reset