An Empirical Method to Quantify the Peripheral Performance Degradation in Deep Networks

12/04/2020
by   Calden Wloka, et al.
2

When applying a convolutional kernel to an image, if the output is to remain the same size as the input then some form of padding is required around the image boundary, meaning that for each layer of convolution in a convolutional neural network (CNN), a strip of pixels equal to the half-width of the kernel size is produced with a non-veridical representation. Although most CNN kernels are small to reduce the parameter load of a network, this non-veridical area compounds with each convolutional layer. The tendency toward deeper and deeper networks combined with stride-based down-sampling means that the propagation of this region can end up covering a non-negligable portion of the image. Although this issue with convolutions has been well acknowledged over the years, the impact of this degraded peripheral representation on modern network behavior has not been fully quantified. What are the limits of translation invariance? Does image padding successfully mitigate the issue, or is performance affected as an object moves between the image border and center? Using Mask R-CNN as an experimental model, we design a dataset and methodology to quantify the spatial dependency of network performance. Our dataset is constructed by inserting objects into high resolution backgrounds, thereby allowing us to crop sub-images which place target objects at specific locations relative to the image border. By probing the behaviour of Mask R-CNN across a selection of target locations, we see clear patterns of performance degredation near the image boundary, and in particular in the image corners. Quantifying both the extent and magnitude of this spatial anisotropy in network performance is important for the deployment of deep networks into unconstrained and realistic environments in which the location of objects or regions of interest are not guaranteed to be well localized within a given image.

READ FULL TEXT

page 2

page 5

page 6

page 8

page 9

page 11

page 12

research
04/13/2021

Tracking translation invariance in CNNs

Although Convolutional Neural Networks (CNNs) are widely used, their tra...
research
05/30/2017

A Kernel Redundancy Removing Policy for Convolutional Neural Network

Deep Convolutional Neural Networks (CNN) have won a significant place in...
research
12/17/2017

Spatial As Deep: Spatial CNN for Traffic Scene Understanding

Convolutional neural networks (CNNs) are usually built by stacking convo...
research
04/24/2015

Object Level Deep Feature Pooling for Compact Image Representation

Convolutional Neural Network (CNN) features have been successfully emplo...
research
02/06/2022

Hyper-Convolutions via Implicit Kernels for Medical Imaging

The convolutional neural network (CNN) is one of the most commonly used ...
research
06/29/2021

Attaining entropy production and dissipation maps from Brownian movies via neural networks

Quantifying entropy production (EP) is essential to understand stochasti...

Please sign up or login with your details

Forgot password? Click here to reset