Deep Residual Learning in the JPEG Transform Domain

12/31/2018
by   Max Ehrlich, et al.
0

We introduce a general method of performing Residual Network inference and learning in the JPEG transform domain that allows the network to consume compressed images as input. Our formulation leverages the linearity of the JPEG transform to redefine convolution and batch normalization with a tune-able numerical approximation for ReLu. The result is mathematically equivalent to the spatial domain network up to the ReLu approximation accuracy. A formulation for image classification and a model conversion algorithm for spatial domain networks are given as examples of the method. We show that the sparsity of the JPEG format allows for faster processing of images with little to no penalty in the network accuracy.

READ FULL TEXT
research
10/13/2021

Transform and Bitstream Domain Image Classification

Classification of images within the compressed domain offers significant...
research
11/14/2016

Identity Matters in Deep Learning

An emerging design principle in deep learning is that each layer of a de...
research
05/14/2015

Neural Network with Unbounded Activation Functions is Universal Approximator

This paper presents an investigation of the approximation property of ne...
research
05/28/2016

Weighted Residuals for Very Deep Networks

Deep residual networks have recently shown appealing performance on many...
research
07/16/2019

Single-bit-per-weight deep convolutional neural networks without batch-normalization layers for embedded systems

Batch-normalization (BN) layers are thought to be an integrally importan...
research
11/15/2022

DCT Perceptron Layer: A Transform Domain Approach for Convolution Layer

In this paper, we propose a novel Discrete Cosine Transform (DCT)-based ...
research
10/19/2017

Block DCT filtering using vector processing

Filtering is an important issue in signals and images processing. Many i...

Please sign up or login with your details

Forgot password? Click here to reset