VC dimension of partially quantized neural networks in the overparametrized regime

10/06/2021
by   Yutong Wang, et al.
0

Vapnik-Chervonenkis (VC) theory has so far been unable to explain the small generalization error of overparametrized neural networks. Indeed, existing applications of VC theory to large networks obtain upper bounds on VC dimension that are proportional to the number of weights, and for a large class of networks, these upper bound are known to be tight. In this work, we focus on a class of partially quantized networks that we refer to as hyperplane arrangement neural networks (HANNs). Using a sample compression analysis, we show that HANNs can have VC dimension significantly smaller than the number of weights, while being highly expressive. In particular, empirical risk minimization over HANNs in the overparametrized regime achieves the minimax rate for classification with Lipschitz posterior class probability. We further demonstrate the expressivity of HANNs empirically. On a panel of 121 UCI datasets, overparametrized HANNs match the performance of state-of-the-art full-precision models.

READ FULL TEXT
research
02/10/2018

On the Universal Approximability of Quantized ReLU Neural Networks

Compression is a key step to deploy large neural networks on resource-co...
research
11/02/2018

Minimax Estimation of Neural Net Distance

An important class of distance metrics proposed for training generative ...
research
10/02/2020

The Efficacy of L_1 Regularization in Two-Layer Neural Networks

A crucial problem in neural networks is to select the most appropriate n...
research
01/26/2023

WL meet VC

Recently, many works studied the expressive power of graph neural networ...
research
06/06/2021

Towards an Understanding of Benign Overfitting in Neural Networks

Modern machine learning models often employ a huge number of parameters ...
research
12/12/2012

Dimension Correction for Hierarchical Latent Class Models

Model complexity is an important factor to consider when selecting among...
research
06/04/2022

Combinatorial optimization for low bit-width neural networks

Low-bit width neural networks have been extensively explored for deploym...

Please sign up or login with your details

Forgot password? Click here to reset