How Many Samples are Needed to Learn a Convolutional Neural Network?

05/21/2018
by   Simon S. Du, et al.
0

A widespread folklore for explaining the success of convolutional neural network (CNN) is that CNN is a more compact representation than the fully connected neural network (FNN) and thus requires fewer samples for learning. We initiate the study of rigorously characterizing the sample complexity of learning convolutional neural networks. We show that for learning an m-dimensional convolutional filter with linear activation acting on a d-dimensional input, the sample complexity of achieving population prediction error of ϵ is O (m/ϵ^2), whereas its FNN counterpart needs at least Ω(d/ϵ^2) samples. Since m ≪ d, this result demonstrates the advantage of using CNN. We further consider the sample complexity of learning a one-hidden-layer CNN with linear activation where both the m-dimensional convolutional filter and the r-dimensional output weights are unknown. For this model, we show the sample complexity is O((m+r)/ϵ^2) when the ratio between the stride size and the filter size is a constant. For both models, we also present lower bounds showing our sample complexities are tight up to logarithmic factors. Our main tools for deriving these results are localized empirical process and a new lemma characterizing the convolutional structure. We believe these tools may inspire further developments in understanding CNN.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/12/2019

Tight Sample Complexity of Learning One-hidden-layer Convolutional Neural Networks

We study the sample complexity of learning one-hidden-layer convolutiona...
research
09/14/2016

Understanding Convolutional Neural Networks with A Mathematical Model

This work attempts to address two fundamental questions about the struct...
research
07/13/2017

A Brief Study of In-Domain Transfer and Learning from Fewer Samples using A Few Simple Priors

Domain knowledge can often be encoded in the structure of a network, suc...
research
05/08/2023

High-Dimensional Smoothed Entropy Estimation via Dimensionality Reduction

We study the problem of overcoming exponential sample complexity in diff...
research
05/09/2018

Controlling the privacy loss with the input feature maps of the layers in convolutional neural networks

We propose the method to sanitize the privacy of the IFM(Input Feature M...
research
02/01/2017

Design, Analysis and Application of A Volumetric Convolutional Neural Network

The design, analysis and application of a volumetric convolutional neura...
research
06/15/2021

Predicting Unreliable Predictions by Shattering a Neural Network

Piecewise linear neural networks can be split into subfunctions, each wi...

Please sign up or login with your details

Forgot password? Click here to reset