Knowledge Base Index Compression via Dimensionality and Precision Reduction

04/06/2022
by   Vilém Zouhar, et al.
0

Recently neural network based approaches to knowledge-intensive NLP tasks, such as question answering, started to rely heavily on the combination of neural retrievers and readers. Retrieval is typically performed over a large textual knowledge base (KB) which requires significant memory and compute resources, especially when scaled up. On HotpotQA we systematically investigate reducing the size of the KB index by means of dimensionality (sparse random projections, PCA, autoencoders) and numerical precision reduction. Our results show that PCA is an easy solution that requires very little data and is only slightly worse than autoencoders, which are less stable. All methods are sensitive to pre- and post-processing and data should always be centered and normalized both before and after dimension reduction. Finally, we show that it is possible to combine PCA with using 1bit per dimension. Overall we achieve (1) 100× compression with 75 with 92

READ FULL TEXT

page 4

page 8

research
12/30/2020

A Memory Efficient Baseline for Open Domain Question Answering

Recently, retrieval systems based on dense representations have led to i...
research
01/24/2022

Artefact Retrieval: Overview of NLP Models with Knowledge Base Access

Many NLP models gain performance by having access to a knowledge base. A...
research
05/01/2020

How to reduce dimension with PCA and random projections?

In our "big data" age, the size and complexity of data is steadily incre...
research
05/25/2021

A Survey on Complex Knowledge Base Question Answering: Methods, Challenges and Solutions

Knowledge base question answering (KBQA) aims to answer a question over ...
research
05/07/2021

Empirical Evaluation of Pre-trained Transformers for Human-Level NLP: The Role of Sample Size and Dimensionality

In human-level NLP tasks, such as predicting mental health, personality,...
research
07/16/2022

Learnable Mixed-precision and Dimension Reduction Co-design for Low-storage Activation

Recently, deep convolutional neural networks (CNNs) have achieved many e...
research
03/30/2020

Pruned Wasserstein Index Generation Model and wigpy Package

Recent proposal of Wasserstein Index Generation model (WIG) has shown a ...

Please sign up or login with your details

Forgot password? Click here to reset