Progressive Compressed Records: Taking a Byte out of Deep Learning Data

11/01/2019
by   Michael Kuchnik, et al.
6

Deep learning training accesses vast amounts of data at high velocity, posing challenges for datasets retrieved over commodity networks and storage devices. We introduce a way to dynamically reduce the overhead of fetching and transporting training data with a method we term Progressive Compressed Records (PCRs). PCRs deviate from previous formats by using progressive compression to convert a single dataset into multiple datasets of increasing fidelity—all without adding to the total dataset size. Empirically, we implement PCRs and evaluate them on a wide range of datasets: ImageNet, HAM10000, Stanford Cars, and CelebA-HQ. Our results show that different tasks can tolerate different levels of compression. PCRs use an on-disk layout that enables applications to efficiently and dynamically access appropriate levels of compression at runtime. In turn, we demonstrate that PCRs can seamlessly enable a 2x speedup in training time on average over baseline formats.

READ FULL TEXT

page 3

page 15

research
09/26/2017

Learning to Inpaint for Image Compression

We study the design of deep architectures for lossy image compression. W...
research
05/03/2023

ProgDTD: Progressive Learned Image Compression with Double-Tail-Drop Training

Progressive compression allows images to start loading as low-resolution...
research
05/16/2022

Lost in Compression: the Impact of Lossy Image Compression on Variable Size Object Detection within Infrared Imagery

Lossy image compression strategies allow for more efficient storage and ...
research
12/12/2021

DPICT: Deep Progressive Image Compression Using Trit-Planes

We propose the deep progressive image compression using trit-planes (DPI...
research
11/28/2017

Homomorphic Parameter Compression for Distributed Deep Learning Training

Distributed training of deep neural networks has received significant re...
research
08/07/2023

A General Framework for Progressive Data Compression and Retrieval

In scientific simulations, observations, and experiments, the cost of tr...
research
04/19/2021

SAPE: Spatially-Adaptive Progressive Encoding for Neural Optimization

Multilayer-perceptrons (MLP) are known to struggle with learning functio...

Please sign up or login with your details

Forgot password? Click here to reset