Neural Networks, Ridge Splines, and TV Regularization in the Radon Domain

06/10/2020
by   Rahul Parhi, et al.
0

We develop a variational framework to understand the properties of the functions learned by neural networks fit to data. We propose and study of a family of continuous-domain linear inverse problems with total variation-like regularization in the Radon domain subject to data fitting constraints. We derive a representer theorem showing that finite-width, single-hidden layer neural networks are solutions to these inverse problems. We draw on many techniques from variational spline theory and so we propose the notion of a ridge spline, which corresponds to fitting data with a single-hidden layer neural network. The representer theorem is reminiscent of the classical Reproducing Kernel Hilbert space representer theorem, but the neural network problem is set in a non-Hilbertian Banach space. Although the learning problems are posed in the continuous-domain, similar to kernel methods, the problems can be recast as finite-dimensional neural network training problems. These neural network training problems have regularizers which are related to the well-known weight decay and path-norm regularizers. Thus, our result gives insight into functional characteristics of trained neural networks and also into the design neural network regularizers. We also show that these regularizers promote neural network solutions with desirable generalization properties.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/07/2021

What Kinds of Functions do Deep Neural Networks Learn? Insights from Variational Spline Theory

We develop a variational framework to understand the properties of funct...
research
09/20/2021

Understanding neural networks with reproducing kernel Banach spaces

Characterizing the function spaces corresponding to neural networks can ...
research
10/05/2019

Minimum "Norm" Neural Networks are Splines

We develop a general framework based on splines to understand the interp...
research
12/27/2017

Neural network augmented inverse problems for PDEs

In this paper we show how to augment classical methods for inverse probl...
research
06/10/2022

Intrinsic dimensionality and generalization properties of the ℛ-norm inductive bias

We study the structural and statistical properties of ℛ-norm minimizing ...
research
05/25/2023

Vector-Valued Variation Spaces and Width Bounds for DNNs: Insights on Weight Decay Regularization

Deep neural networks (DNNs) trained to minimize a loss term plus the sum...
research
05/21/2019

Total variation multiscale estimators for linear inverse problems

Even though the statistical theory of linear inverse problems is a well-...

Please sign up or login with your details

Forgot password? Click here to reset