DeepMapping: The Case for Learned Data Mapping for Compression and Efficient Query Processing

07/12/2023
by   Lixi Zhou, et al.
0

Storing tabular data in a way that balances storage and query efficiencies is a long standing research question in the database community. While there are several lossless compression techniques in the literature, in this work we argue and show that a novel Deep Learned Data Mapping (or DeepMapping) abstraction, which relies on the impressive memorization capabilities of deep neural networks, can provide better storage cost, better latency, and better run-time memory footprint, all at the same time. Our proposed DeepMapping abstraction transforms a data set into multiple key-value mappings and constructs a multi-tasking neural network model that outputs the corresponding values for a given input key. In order to deal with the memorization errors, DeepMapping couples the learned neural network with a light-weight auxiliary data structure capable of correcting errors. The auxiliary structure further enables DeepMapping to efficiently deal with insertions, deletions, and updates, without having to re-train the mapping. Since the shape of the network has a significant impact on the overall size of the DeepMapping structure, we further propose a multi-task hybrid architecture search strategy to identify DeepMapping architectures that strike a desirable balance among memorization capacity, size, and efficiency. Extensive experiments with synthetic and benchmark datasets, including TPC-H and TPC-DS, demonstrated that the proposed DeepMapping approach can significantly reduce the latency of the key-based queries, while simultaneously improving both offline and run-time storage requirements against several cutting-edge competitors.

READ FULL TEXT

page 4

page 5

page 6

page 7

page 8

page 9

page 10

page 13

research
11/13/2017

Weightless: Lossy Weight Encoding For Deep Neural Network Compression

The large memory requirements of deep neural networks limit their deploy...
research
10/06/2020

Rank and run-time aware compression of NLP Applications

Sequence model based NLP applications can be large. Yet, many applicatio...
research
03/08/2022

YONO: Modeling Multiple Heterogeneous Neural Networks on Microcontrollers

With the advancement of Deep Neural Networks (DNN) and large amounts of ...
research
06/12/2019

Run-Time Efficient RNN Compression for Inference on Edge Devices

Recurrent neural networks can be large and compute-intensive, yet many a...
research
12/25/2018

JALAD: Joint Accuracy- and Latency-Aware Deep Structure Decoupling for Edge-Cloud Execution

Recent years have witnessed a rapid growth of deep-network based service...
research
02/15/2021

Neural Network Compression for Noisy Storage Devices

Compression and efficient storage of neural network (NN) parameters is c...
research
05/24/2018

Multi-Task Zipping via Layer-wise Neuron Sharing

Future mobile devices are anticipated to perceive, understand and react ...

Please sign up or login with your details

Forgot password? Click here to reset