Deep Convolutional Networks as shallow Gaussian Processes

08/16/2018
by   Adrià Garriga-Alonso, et al.
0

We show that the output of a (residual) convolutional neural network (CNN) with an appropriate prior over the weights and biases is a Gaussian process (GP) in the limit of infinitely many convolutional filters, extending similar results for dense networks. For a CNN, the equivalent kernel can be computed exactly and, unlike "deep kernels", has very few parameters: only the hyperparameters of the original CNN. Further, we show that this kernel has two properties that allow it to be computed efficiently; the cost of evaluating the kernel for a pair of images is similar to a single forward pass through the original CNN with only one filter per layer. The kernel equivalent to a 32-layer ResNet obtains 0.84 GPs with a comparable number of parameters.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/08/2020

Using PSPNet and UNet to analyze the internal parameter relationship and visualization of the convolutional neural network

Convolutional neural network(CNN) has achieved great success in many fie...
research
11/03/2019

Enhanced Convolutional Neural Tangent Kernels

Recent research shows that for training with ℓ_2 loss, convolutional neu...
research
06/24/2017

Irregular Convolutional Neural Networks

Convolutional kernels are basic and vital components of deep Convolution...
research
10/25/2018

A Gaussian Process perspective on Convolutional Neural Networks

In this paper we cast the well-known convolutional neural network in a G...
research
01/12/2021

Convolutional Neural Network Simplification with Progressive Retraining

Kernel pruning methods have been proposed to speed up, simplify, and imp...
research
07/30/2018

Transformationally Identical and Invariant Convolutional Neural Networks by Combining Symmetric Operations or Input Vectors

Transformationally invariant processors constructed by transformed input...
research
11/23/2021

Critical initialization of wide and deep neural networks through partial Jacobians: general theory and applications to LayerNorm

Deep neural networks are notorious for defying theoretical treatment. Ho...

Please sign up or login with your details

Forgot password? Click here to reset