HEMP: High-order Entropy Minimization for neural network comPression

07/12/2021
by   Enzo Tartaglione, et al.
0

We formulate the entropy of a quantized artificial neural network as a differentiable function that can be plugged as a regularization term into the cost function minimized by gradient descent. Our formulation scales efficiently beyond the first order and is agnostic of the quantization scheme. The network can then be trained to minimize the entropy of the quantized parameters, so that they can be optimally compressed via entropy coding. We experiment with our entropy formulation at quantizing and compressing well-known network architectures over multiple datasets. Our approach compares favorably over similar methods, enjoying the benefits of higher order entropy estimate, showing flexibility towards non-uniform quantization (we use Lloyd-max quantization), scalability towards any entropy order to be minimized and efficiency in terms of compression. We show that HEMP is able to work in synergy with other approaches aiming at pruning or quantizing the model itself, delivering significant benefits in terms of storage size compressibility without harming the model's performance.

READ FULL TEXT
research
12/05/2016

Towards the Limit of Network Quantization

Network quantization is one of network compression techniques to reduce ...
research
12/18/2018

Entropy-Constrained Training of Deep Neural Networks

We propose a general framework for neural network compression that is mo...
research
11/25/2022

Homology-constrained vector quantization entropy regularizer

This paper describes an entropy regularization term for vector quantizat...
research
01/20/2023

Optimized learned entropy coding parameters for practical neural-based image and video compression

Neural-based image and video codecs are significantly more power-efficie...
research
07/22/2022

Quantized Sparse Weight Decomposition for Neural Network Compression

In this paper, we introduce a novel method of neural network weight comp...
research
06/23/2023

QNNRepair: Quantized Neural Network Repair

We present QNNRepair, the first method in the literature for repairing q...
research
06/17/2020

Universally Quantized Neural Compression

A popular approach to learning encoders for lossy compression is to use ...

Please sign up or login with your details

Forgot password? Click here to reset