Guaranteed Quantization Error Computation for Neural Network Model Compression

04/26/2023
by   Wesley Cooke, et al.
0

Neural network model compression techniques can address the computation issue of deep neural networks on embedded devices in industrial systems. The guaranteed output error computation problem for neural network compression with quantization is addressed in this paper. A merged neural network is built from a feedforward neural network and its quantized version to produce the exact output difference between two neural networks. Then, optimization-based methods and reachability analysis methods are applied to the merged neural network to compute the guaranteed quantization error. Finally, a numerical example is proposed to validate the applicability and effectiveness of the proposed approach.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/02/2022

Approximate Bisimulation Relations for Neural Networks and Application to Assured Neural Network Compression

In this paper, we propose a concept of approximate bisimulation relation...
research
07/09/2022

CEG4N: Counter-Example Guided Neural Network Quantization Refinement

Neural networks are essential components of learning-based software syst...
research
07/01/2019

Weight Normalization based Quantization for Deep Neural Network Compression

With the development of deep neural networks, the size of network models...
research
01/17/2023

Safety Verification of Neural Network Control Systems Using Guaranteed Neural Network Model Reduction

This paper aims to enhance the computational efficiency of safety verifi...
research
12/18/2019

TOCO: A Framework for Compressing Neural Network Models Based on Tolerance Analysis

Neural network compression methods have enabled deploying large models o...
research
05/15/2019

DeepCABAC: Context-adaptive binary arithmetic coding for deep neural network compression

We present DeepCABAC, a novel context-adaptive binary arithmetic coder f...
research
02/20/2020

Neural Network Compression Framework for fast model inference

In this work we present a new framework for neural networks compression ...

Please sign up or login with your details

Forgot password? Click here to reset