Training Neural Networks as Learning Data-adaptive Kernels: Provable Representation and Approximation Benefits

01/21/2019
by   Xialiang Dou, et al.
0

Consider the problem: given data pair (x, y) drawn from a population with f_*(x) = E[y | x = x], specify a neural network and run gradient flow on the weights over time until reaching any stationarity. How does f_t, the function computed by the neural network at time t, relate to f_*, in terms of approximation and representation? What are the provable benefits of the adaptive representation by neural networks compared to the pre-specified fixed basis representation in the classical nonparametric literature? We answer the above questions via a dynamic reproducing kernel Hilbert space (RKHS) approach indexed by the training process of neural networks. We show that when reaching any local stationarity, gradient flow learns an adaptive RKHS representation, and performs the global least squares projection onto the adaptive RKHS, simultaneously. In addition, we prove that as the RKHS is data-adaptive and task-specific, the residual for f_* lies in a subspace that is smaller than the orthogonal complement of the RKHS, formalizing the representation and approximation benefits of neural networks.

READ FULL TEXT
research
05/09/2019

A Novel Adaptive Kernel for the RBF Neural Networks

In this paper, we propose a novel adaptive kernel for the radial basis f...
research
02/01/2023

Gradient Descent in Neural Networks as Sequential Learning in RKBS

The study of Neural Tangent Kernels (NTKs) has provided much needed insi...
research
06/07/2019

Recurrent Kernel Networks

Substring kernels are classical tools for representing biological sequen...
research
08/31/2023

Training Neural Networks Using Reproducing Kernel Space Interpolation and Model Reduction

We introduce and study the theory of training neural networks using inte...
research
04/30/2022

NeuralEF: Deconstructing Kernels by Deep Neural Networks

Learning the principal eigenfunctions of an integral operator defined by...
research
05/19/2018

Integral representation of the global minimizer

We have obtained an integral representation of the shallow neural networ...
research
07/10/2023

Self Expanding Neural Networks

The results of training a neural network are heavily dependent on the ar...

Please sign up or login with your details

Forgot password? Click here to reset