Quantization of Generative Adversarial Networks for Efficient Inference: a Methodological Study

08/31/2021
by   Pavel Andreev, et al.
0

Generative adversarial networks (GANs) have an enormous potential impact on digital content creation, e.g., photo-realistic digital avatars, semantic content editing, and quality enhancement of speech and images. However, the performance of modern GANs comes together with massive amounts of computations performed during the inference and high energy consumption. That complicates, or even makes impossible, their deployment on edge devices. The problem can be reduced with quantization – a neural network compression technique that facilitates hardware-friendly inference by replacing floating-point computations with low-bit integer ones. While quantization is well established for discriminative models, the performance of modern quantization techniques in application to GANs remains unclear. GANs generate content of a more complex structure than discriminative models, and thus quantization of GANs is significantly more challenging. To tackle this problem, we perform an extensive experimental study of state-of-art quantization techniques on three diverse GAN architectures, namely StyleGAN, Self-Attention GAN, and CycleGAN. As a result, we discovered practical recipes that allowed us to successfully quantize these models for inference with 4/8-bit weights and 8-bit activations while preserving the quality of the original full-precision models.

READ FULL TEXT

page 7

page 12

research
01/24/2019

QGAN: Quantized Generative Adversarial Networks

The intensive computation and memory requirements of generative adversar...
research
05/21/2023

Integer or Floating Point? New Outlooks for Low-Bit Quantization on Large Language Models

Efficient deployment of large language models (LLMs) necessitates low-bi...
research
04/18/2023

DeepGEMM: Accelerated Ultra Low-Precision Inference on CPU Architectures using Lookup Tables

A lot of recent progress has been made in ultra low-bit quantization, pr...
research
04/07/2020

Deep Attentive Generative Adversarial Network for Photo-Realistic Image De-Quantization

Most of current display devices are with eight or higher bit-depth. Howe...
research
06/15/2021

A White Paper on Neural Network Quantization

While neural networks have advanced the frontiers in many applications, ...
research
01/14/2021

On the quantization of recurrent neural networks

Integer quantization of neural networks can be defined as the approximat...
research
03/22/2021

Evaluating Post-Training Compression in GANs using Locality-Sensitive Hashing

The analysis of the compression effects in generative adversarial networ...

Please sign up or login with your details

Forgot password? Click here to reset