1 Introduction
In recent years, deep neural networks have improved the state of the art performance substantially in computer vision
[Krizhevsky et al., 2012; He et al., 2016; Salman and Liu, 2019], machine translation [Sutskever et al., 2014], speech recognition [Graves et al., 2013], healthcare [Miotto et al., 2017; Salman et al., 2019] and game playing [Silver et al., 2017]among other applications. However, the underlying mechanisms that enable them to perform well are still not well understood. Even though they typically have more parameters than the training samples and exhibit very large capacities, they generalize well on real datasets trained via stochastic gradient descent or its variants. In an insightful paper,
Zhang et al. [2016] have identified a number of intriguing phenomena of such networks. In particular, they demonstrate that overparametrized neural networks can achieve accuracy trained on datasets with the original labels and generalize well. At the same time, the exact same neural network architectures can also achieveaccuracy on the datasets with random labels, and therefore “memorize” the training samples. Clearly, this is not consistent with statistical learning theory
[Vapnik, 1998]and biasvariance tradeoff
[Geman et al., 1992], where models should match the (unknown) capacity of the underlying processes in order to generalize well. Understanding and explaining this typical behavior of deep neural networks has attracted a lot of attention recently with the hope of revealing the underlying mechanisms of how deep neural networks generalize.Fundamentally, while training, deep neural networks iteratively minimize a loss function defined as the sum of the loss on the training samples. The parameters in the trained network depend on the initial parameter values, the optimization process, and training data. As 100% accuracy on the training samples can be achieved even with random labels, finding good solutions for such networks that minimize the loss is therefore not a key issue. While regularization techniques can affect the parameters of trained networks,
Zhang et al. [2016] have demonstrated that their effects are typically small, suggesting that they are not a key component. Therefore, the generalization performance of a trained overparametrized network should depend on the training data and network architecture. In this paper, we focus on deep ReLU networks. We show that such networks generalize consistently and reliably by interpolating among the training points. Using generalization intervals defined as the range of the data that have the same classification along a direction, we discover that pairwise generalization intervals on datasets with real and random labels are almost identical for high dimensional inputs (e.g., MNIST and CIFAR10 samples). Furthermore, we show that pairwise interpolations approximate the underlying manifold in the data well, enabling the networks to generalize well. We show that the properties are remarkably consistent among networks with different architectures and on different datasets. The properties enable us to characterize the generalization performance of neural networks based on their behaviors on the training sets only, which we call intrinsic generalization. This notion of generalization is very different from the typical definition of the performance gap on the training set and test set. While intrinsic generalization of a network on a training set can be quantified through generalization intervals, the gapbased generalization performance can not be studied without having a validation set or test set. Furthermore, the gapbased definition is extrinsic as it can vary when a different validation set is used. In other words, for the first time, we demonstrate the underlying mechanisms that enable overparametrized networks to generalize well when all the training samples are classified correctly. The systematic results demonstrate the effectiveness of the proposed method and therefore validate the proposed solution.The rest of the paper is organized as follows. In the next section, we review recent works that are closely related to our study. After that, we present the theoretical foundation of the generalization mechanism via interpolation for deep ReLU networks. We introduce a novel notation called generalization interval (GI) to quantify the generalization of such networks. Then, we illustrate our proposed ideas on intrinsic generalization behavior of deep ReLU networks with systematic experiments on representative datasets such as MNIST and CIFAR10 along with a twodimensional synthetic dataset. Finally, we discuss correlations between generalization intervals on training sets and validation accuracy and whether there exists a mechanism in deep neural networks that supports deep memorization. We conclude the paper with a brief summary and plan for future work.
2 Related Work
The ability to perform well on new data is one of the hallmarks of any useful machine learning system, including deep neural networks. Traditionally, the statistical learning theory
[Vapnik, 1998] ties a model’s ability to generalize well to whether its capacity matches the underlying (unknown) complexity of the problem. In contrast, in recent years, deep neural networks have empirically improved the performance of many tasks even with more parameters than the number of training samples. Understanding the generalization ability and mechanism of deep neural networks have become central to get their full practical benefit in real applications [Jakubovitz et al., 2018]. While important, it is difficult to study the generalization performance of a deep neural network as it is defined as the performance gap on a training set and test one.Zhang et al. [2016] have demonstrated concretely that the same deep neural networks can generalize and at the same time can “memorize” training set with random labels, which makes it clear that the capacity of deep neural networks [Arpit et al., 2017] can not explain their ability to generalize. Consequentially, a central research question is to discover the properties of deep networks, the data, or the interplay that allow the network to generalize well and at the same time “memorize”.
Mechanically, training deep neural networks is based on minimizing a loss iteratively using gradient descent from an initial, often randomized, solution. As deep neural networks implement continuous functions, similar inputs will produce similar outputs. If we use the literal definition of memorization, i.e., the learning process associates a particular input with a particular label [Collins et al., 2018], deep neural networks are not capable of memorization only. Therefore, deep neural networks must “generalize” in that after training on a training set, they will also produce answers for other inputs. For example, for classification, a trained (deep neural network) model divides the input space into decision regions, and the inputs in the same region produce the same output. In other words, all the inputs are associated with the same output. As a result, for deep neural networks, the question becomes how they generalize, rather than whether they generalize.
It appears that the discussions about memorization and generalization of deep neural networks are often due to different definitions of memorization.For example, in [Arpit et al., 2017; Chatterjee, 2018], memorization is simply defined to be a point when a network reaches accuracy on the training set. Note that for the network to generalize, it should perform well on the training set, i.e., it should memorize at the same time. This could not help elucidate the underlying mechanism that enables a deep neural network to generalize.
In this paper, to quantify the intrinsic ability of a deep neural network to generalize, we introduce generalization intervals to measure how far a trained deep network can generalize. Through systematic experiments, we discover that pairwise generalization intervals are remarkably similar on datasets with real and random labels, confirming that generalization of deep neural networks is independent of the labeling of data. Furthermore, we show that on real datasets, pairwise interpolations provide a way to approximate the underlying manifolds that allow such models to generalize well.
3 Theoretical Foundation for Deep ReLU Networks
3.1 Notations and ReLU Networks
In this paper, we use the following commonly used notations. We focus on feedforward deep neural networks having Rectified Linear Unit (ReLU) activations for classification which is defined by
. The neural network functions can be defined as , where and are the input and output dimension respectively. We consideris an input vector and
is a target vector. We define a set of n training samples as . The functions are parameterized by , which is a vector that includes all the parameters (i.e., weights and biases). Furthermore, we would use the function interchangeably to denote the preactivation output of the last layer (i.e., output from the penultimate layer).By restricting the activation function to be ReLU, a neural network provides a piecewise approximation of complex decision regions
[Montúfar et al., 2014] by minimizing the empirical risk,(1) 
where denotes some loss function such as loss, hinge loss, cross entropy loss etc. The gradient based methods update the weights by
(2) 
where is known as the step size. To simplify our classification model, we have considered the cross entropy loss. Let denotes a particular layer,
the last layer of the network, k the index of a neuron and c the number of classes. Given that the last layer is a softmax layer, the activations for the neurons in the last layer would be defined by
(3) 
where denotes weighted summation of activations from prior layer. The cross entropy loss is defined as follows.
(4) 
Now, from the definition of the loss, we can derive the weight updates for the last layer as the following.
(5) 
where denotes the ReLU activations of previous layer. Note that is 1 for the target class (which can be a real label or a random label assigned to the sample) and 0 for all the other classes. Equation 5 shows the importance of labels for gradient descent. The weights are getting closer to the representation of the neuron in the assigned class in the previous layer while pushing away from that of the neurons in the other classes, representing competitive learning [Rumelhart and Zipser, 1988].
3.2 Generalization Intervals
In order to study generalization behavior quantitatively, we introduce the Generalization Interval (GI), which is defined as the generalization interval of a trained deep neural network at input point along a particular direction as the following.
(6)  
Intuitively, the above equation defines the generalization interval at a given point as the range of the inputs that have stable classification along a particular direction. When the samples are from a meaningful data distribution, i.e., when the samples are from an underlying manifold, the generalization intervals along the tangent directions of the manifold should be large. By connecting the samples with the same true labels in a pairwise manner, one can approximate the tangent directions smoothly of the manifold. In contrast, the directions with other labels approximate the normal directions generally, resulting in poor generalizations (i.e., small generalization intervals).
Since there are many such pairwise paths, a question is whether multiple lines interfere with each other. In a high dimensional space (such as MNIST or CIFAR10; see the Experiments Section), the probability of two lines intersecting with the other lines or getting very close to other lines is low. However, when the dimension gets smaller, the chance of having two lines interfering with each other will be higher. In addition, the interference affects the performance of a trained neural network only if the two points are from different classes. On datasets with random labels, one would expect higher chances of interference if closer points with different labels are much more common. However, we argue that deep neural networks still do not simply memorize training points even when the labels are randomized. Please see the Experiments Section for numerical results.
3.3 Generalization Along Tangent and Normal Directions
(to be viewed in color) When the models are trained with real labels, outputs from the penultimate layer along the direction defined by class 9 samples. (left) first principal component. (middle) perpendicular (mean class 9 to mean class 1) to the first principal component. (right) principal component having a small eigenvalue (i.e., approximately a normal direction).
As ReLU networks partition the input space into linear regions [Montúfar et al., 2014], we consider two samples from same class denoted as and . The linear network can be defined as
(7) 
where is the input and are the weights of the layers 1,…,L respectively. While the weights remain constant in one linear region, the weights in general depend on , as indicated in Equation 7.
At a particular point, we have a specific linear network. Therefore, we have the following linear model at .
(8) 
If we consider a particular direction from point to another point x, then we have the network as
(9) 
The change of the output of the network in that particular direction is
(10)  
If two training samples (i.e., and ) are classified correctly, then the difference between and should be small. If the two points in the same linear region, the first part of Equation 10 (i.e., ) will be zero; if the linear regions are similar, the part should be small. In general, we have established a bound of the change of weight matrices of ReLU networks; see Appendix for detailed mathematical analysis. Currently we are working on improving the bound. More generally, whether the change of the output (i.e., ) would be small or not along the direction () depends largely on the change of weight shown on the second part of the equation (i.e., ). The part corresponds to tangent direction loosely when both and are on the same manifold.
This observation is closely related to the one by Simard et al. [1998], where learning should be more efficient along the tangent directions, compared to that of normal vectors on the underlying manifold. Their proposed tangent propagation algorithm makes sure that learning involves using the information of the derivatives of the classification function that is supplied by the tangent vectors. The tangent vectors are defined explicitly in their method, while in our case, we argue that the samples from the same class approximate the tangent vectors of the underlying manifold using pairwise interpolations defined by the data. For example, direction approximates a tangent vector on the underlying manifold, while approximates a normal direction.
The analysis shows that deep ReLU networks generalize via interpolation. In the random data point case, we have checked the generalization intervals that show that they generalize similarly like the real data points. If there is an underlying manifold, samples from the same class form a compact and possibly directional set, a small number of samples may be sufficient to approximate the manifold, enabling a trained network to generalize well from a small training set. When there is no manifold as in the random label case, it is difficult to characterize the set that a network generalizes to, but it still interpolates between samples with the same label.
4 Experiments
4.1 Datasets
Our experiments are based on the widely used representative MNIST [Lecun et al., 1998] and CIFAR10 datasets [Krizhevsky, 2009]. The CIFAR10 dataset consists of 3232 color images in 10 classes. The images are split into 50,000 training samples and 10,000 test/validation samples. On the other hand, The MNIST dataset consists of 70,000 handwritten digit images of size 2828 for each image; 60,000 training samples and 10,000 test/validation samples. However, to get some computational advantage, we have considered 1000 training samples (100 from each of the 10 classes). We have trained our deep neural network models with real labels and randomized labels after randomizing the class labels. We have made sure our network architectures and optimization procedures to get the training error to be very close to 0. We have mainly used the models that have three hidden dense layers with 128 ReLU units in each of the layers. However, to illustrate our proposed methods working on different architectures too, we have also conducted experiments on models with 1, 2, and 4 hidden layers having 512, 256, and 128 ReLU units respectively in each of the layers. The weights are initialized with random uniform, and the networks are optimized using SGD with learning rate of 0.01. The validation accuracy of the model trained on MNIST dataset is found around when trained on real labels (i.e., randomized) and when trained on random labels (i.e., randomized). To illustrate the impact of the dimension of input space on our proposed methods, we have also considered a twodimensional synthetic dataset based on the paper by Rozza et al. [2014]. This dataset is known as the “two moons” dataset having an equal number of points in two classes (red and blue). Those points are split such that we get 300 training samples and 100 test/validation samples.
4.2 Pairwise Interpolations That Approximate the Underlying Manifold
Figure 1 shows the output of the network trained on meaningful data (i.e., with real labels); along the path defined by samples from the same class, the neural network output for the correct class does not change much (i.e., stable) compared to other directions and the network classifies all the inputs along the path correctly. This behavior of the networks shows that they interpolate between the samples.
To demonstrate the neural networks’ ability to approximate the underlying manifold, we show the network outputs along the principal components for a class, which capture the tangent directions of the local manifolds. Similarly, minor components approximate normal directions. In a highdimensional space, there are many normal directions to a local lowdimensional manifold. As examples, Figure 2(left) shows the output along the first principal component of a particular class, while Figure 2(middle) depicts the result along a normal direction by making the direction from class 9 (mean) to class 1 (mean) perpendicular to the first principal component. In addition, Figure 2(right) shows the output for another component that has very small eigenvalue. Along the principal directions, the output of the correct class is more or less constant, which is expected as it approximates the tangent directions. Similarly, Figure 3 illustrates the behavior of a network trained on random labels. While the neural network output of the correct class changes more compared to that in Figure 1(left), all the samples along the path are still classified “correctly”, i.e., the same as the random label assigned to the two samples. However, as expected, the mean of random class “9” is not very different from that of any other random class, and there is no meaningful local manifold to differentiate different classes. Figure 3(right) verifies that there is indeed no manifold to approximate. However, the models still interpolate pairwise linearly as the other plots show.
4.3 Generalization Intervals for Lines and Triangles on MNIST and CIFAR10
To illustrate our idea of intrinsic generalization via interpolation, we have conducted a series of experiments on MNIST and CIFAR10 datasets. For the pairwise lines, we observe almost similar generalization intervals for both real and random labels. In addition, to illustrate the role of the dimension of input space in interpolation, we explore generalization intervals on the MNIST dataset with reduced dimensions. To compute the generalization intervals, we consider the directions from particular training points to other points in the same class for both real and random labels. We have limited the maximal range as 3 (from 1 to 2) for pairwise lines centered at one of the two points. Figure 4(left) shows the distributions of generalization intervals with unchanged dimension of MNIST dataset (); the generalization intervals do not vary significantly between real and random labels. As one expects, different paths may interfere with each other when they are close. While the interference is small in the original dimension, the effects are shown as the small bump around 1.3. When the dimension is reduced, the interference should be more prominent. To demonstrate such effects, Figure 4(middle) and 4(right) illustrate the behavior when the samples are downsampled. Both of the figures clearly show that, with lower dimensions, generalization intervals get lower with random labels. Note that there is no significant change in generalization intervals with real labels. This further confirms our claim of generalization via interpolation of the underlying manifold of the data. We have also repeated the experiments on CIFAR10 dataset. As shown in Figure 7, they behave similarly with real and random labels as on the MNIST dataset.
To further validate the effect of dimension of data, we observe the pairwise generalization intervals on “two moons” dataset. Figure 8 illustrates the density of the generalization intervals for one of the two classes before and after randomizing the labels. This shows that the density changes dramatically after randomizing the labels of twodimensional data. According to current definitions, apparently, it seems the network is memorizing data with random labels as the generalization intervals are much lower compared to data with real labels. In the twodimensional case, almost all the lines have high probability of being interfered with each other.
To study the generalization intervals of higherorder structures, Figure 5 shows the density of the generalization intervals along the directions defined by the pairwise lines connecting the midpoints of a triangle (i.e., the midlines; range as 5) formed by the samples of MNIST from class 5. We observe that the generalization intervals do not change much with those directions with real labels. However, we observe a significant change in the case of random labels. This confirms that there is no underlying manifold for random labels and midline directions behave like random directions.
We also find similar behavior with networks having different number of layers. Figure 6 shows the pairwise generalization interval distributions for three different architectures. Even though the number of layers is different, the distributions of the pairwise generalization intervals remain roughly the same. As the number of layers increases, the nonlinearity increases, and the bump around 1.3 becomes more visible. Therefore, the experiments suggest that generalization is largely decided by how the networks generalize between the points. In addition, the experiments confirm that the results are valid with different datasets and architectures.
To further validate the relationship between generalization intervals and generalization performance of a neural network, Figure 9 shows the correlation of generalization intervals with training and validation performance for MNIST dataset during the training process. One can see clearly that the generalization intervals correlate very well with the model’s generalization performance on the validation set.
5 Discussion
In this paper, by analyzing the mechanism of training deep neural networks, we demonstrate quantitatively that deep neural networks generalize from training samples via interpolation, even when the labels are randomized. The remarkable similarity of distributions of pairwise generalization intervals on datasets with real and random labels answer the question that deep neural networks generalize and do not simply memorize (i.e., associate a label to a particular input only). Furthermore, pairwise interpolations provide a good approximation of the underlying manifolds of real datasets. The experiments lead to several further research questions. For example, how regularization techniques and neural network architectures affect details of specific solutions even though the pairwise generalization intervals are similar. In particular, the differences can and will affect the generalization performance on particular validation sets as variations are expected. Additionally, as shown by Zhang et al. [2016], the performance differences due to regularization techniques are often small. However, small differences could also be significant because models are often designed and trained for a small amount of improvements. It is important to analyze whether such improvements are extrinsic simply due to optimization or the choice of test sets or other factors. Similarly, as data are important in determining the generalization performance of deep neural networks, neural network architectures should be important. Clearly, different architectures can behave differently locally and therefore affect the details of generalization along pairwise paths. In this paper, we have used principal components to capture linear and local manifolds. Manifolds of real datasets are often complex and exhibit globally nonlinear structures. As the interpolations are datadriven, we expect that the interpolations can also approximate nonlinear manifolds well. How to quantify the relationships between nonlinearity of manifolds and deep neural network generalization needs to be further investigated.
The correlation of generalization intervals with training and validation accuracy demonstrates an intrinsic way to quantify the generalization performance of deep neural networks without relying on particular choices of validation sets. The performance measure can be useful for neural architecture search and for designing better neural network components and architectures. The linear interpolations, while effective for ReLU networks without any additional nonlinearity, need to handle nonlinearity due to other nonlinear components such as max pooling and maxout. Such nonlinear components allow inputs not to be aligned but have the same outputs. For example, we have observed that the correlation between the generalization intervals and validation accuracy for the CIFAR10 dataset is lower than that on the MNIST dataset. This is being investigated further.
6 Conclusion
In this paper, we demonstrate for the first time that deep ReLU neural networks generalize through interpolations. While the pairwise generalization intervals on real and random datasets are remarkably similar on high dimensional datasets such as MNIST and CIFAR10, the pairwise interpolations also approximate the underlying manifolds well when they exist, enabling the models to generalize well. While we have characterized systematically for networks with ReLU as the sole source of nonlinearity, how to compute the generalization intervals efficiently for networks with additional nonlinearity such as max pooling and maxout still need to be investigated. Even though regularization techniques and neural network architectures may not have a significant impact on generalization, they can and do have impacts. Analyzing and modeling their impacts is also being studied.
References
 A closer look at memorization in deep networks. In Proceedings of the 34th International Conference on Machine Learning, ICML, Cited by: §2, §2.
 Learning and memorization. In Proceedings of the 35th International Conference on Machine Learning, pp. 755–763. Cited by: §2.
 Detecting memorization in relu networks. CoRR. External Links: 1810.03372 Cited by: §2.
 Neural networks and the bias/variance dilemma. Neural Computation. Cited by: §1.

Speech recognition with deep recurrent neural networks
. CoRR abs/1303.5778. External Links: 1303.5778 Cited by: §1. 
Deep residual learning for image recognition.
2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)
, pp. 770–778. Cited by: §1. 
Generalization error in deep learning
. CoRR. External Links: 1808.01174 Cited by: §2.  ImageNet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems 25, F. Pereira, C. J. C. Burges, L. Bottou, and K. Q. Weinberger (Eds.), pp. 1097–1105. Cited by: §1.
 Learning multiple layers of features from tiny images. Technical report . Cited by: §4.1.
 Gradientbased learning applied to document recognition. Proceedings of the IEEE 86 (11), pp. 2278–2324. External Links: Document, ISSN 00189219 Cited by: §4.1.
 Deep learning for healthcare: review, opportunities and challenges. Briefings in Bioinformatics, pp. 1236–1246. External Links: Document Cited by: §1.
 On the number of linear regions of deep neural networks. In Proceedings of the 27th International Conference on Neural Information Processing Systems  Volume 2, NIPS’14, pp. 2924–2932. Cited by: §3.1, §3.3.

A novel graphbased fisher kernel method for semisupervised learning
. Cited by: §4.1.  Connectionist models and their implications: readings from cognitive science. D. Waltz and J. A. Feldman (Eds.), pp. 205–242. External Links: ISBN 0893914568 Cited by: §3.1.
 Sparsity as the implicit gating mechanism for residual blocks. In 2019 International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, pp. 1–6. External Links: Document Cited by: §1.
 Consensusbased Interpretable Deep Neural Networks with Application to Mortality Prediction. arXiv eprints, pp. arXiv:1905.05849. External Links: 1905.05849 Cited by: §1.
 Mastering the game of go without human knowledge. Nature 550, pp. 354–359. Cited by: §1.
 Transformation invariance in pattern recognitiontangent distance and tangent propagation. In Neural Networks: Tricks of the Trade, This Book is an Outgrowth of a 1996 NIPS Workshop, Cited by: §3.3.
 Sequence to sequence learning with neural networks. In Advances in Neural Information Processing Systems 27, Z. Ghahramani, M. Welling, C. Cortes, N. D. Lawrence, and K. Q. Weinberger (Eds.), pp. 3104–3112. Cited by: §1.
 Statistical learning theory. adaptive and learning systems for signal process ing, communications, and control.. Wiley. Cited by: §1, §2.
 Understanding deep learning requires rethinking generalization. CoRR abs/1611.03530. Cited by: §1, §1, §2, §5.
Appendix A Generalization and Interpolation of Deep ReLU networks based on Matrix Norm
In this supplementary subsection, we apply basic operator theory to show that the rate of change of the output from the deep ReLU network is overall bounded with respect to the rate of change of the input. Moreover, this boundary is “reasonable” in most portion of the input space. Then, we will further interpret the results shown in Figure 1, 2 and 3 of the main paper.
Let’s start with deep linear networks. That is, we first remove the ReLU layers. A multilayer linear network can be defined as
where is the input and are the matrices representation of layers 1,…,L respectively. Then, suppose and so that the linear network can be represented as , which is a linear projection by matrix from to .
Before going on, we shall briefly introduce the notions of vector’s norm and matrix norm: Let be a real number. Then, we can define a measure of distance called  on the real space :
It only takes a routine derivation to show that is indeed a norm, which is not discussed in this paper. However, we do want to note that: When , we shall get the typical Euclidean distance. And when approaches positive infinity, the norm becomes the :
If we apply the same norm on both and , there induces a corresponding (or ) on the linear projection as:
where mean the supreme of the scales in this set.
And we note that in the special case of , the matrix norm will have the formula:
which is simply the maximum absolute row sum of the matrix.
Intuitively, the matrix norm put an upper bound to the variance in with respect to that in : With this property, we provide a strict mathematical analysis on linear network generalization.
Suppose we are working with the dataset MNIST, so that and , with each dimension in corresponding to the digitclass . Also, we assume that the linear network is welltrained: Say, for any sample , the dimension of the labeled digitclass in is larger than any other dimension by at least . In this case, suppose the dimension of the labeled digitclass is . Then, , and so that the output after softmax will be like . But the exact value of may be set differently based on the dimension of and the training requirements. We call the . We shall put in use shortly when interpreting experimental results.
Now, we add the ReLU layers back, so that
Suppose , and the ’th dimension of is less than zero. Then, the ReLU function will turn the ’th dimension of into zero, which is also equivalent to turning all the elements in row of into zeros. In fact, this means that for the fixed value , we shall have , where is the matrix same to except for some rows to be all zeros.
We note that different values of will lead to different , i.e., different rows in will become zeros. But due to the continuity of , will be fixed in local areas of input space . In fact, if , the input space can be theoretically separated into regions: with for , so that is fixed in each . Hence, we can see that the operation is piecewise linear throughout the input space . (Also note that we use to represent in our main paper. But in this supplementary subsection, we always use to represent the matrix after ReLU function.)
Taking this analysis into the entire deep ReLU network, we can have that: There exists a space separation with and for any , such that
where each matrix is the same to except for some rows to be all zeros, and all the matrices (and hence ) are fixed in each . Therefore, we can see that the deep ReLU network is throughout the input space .
Since is obtained by erasing rows in into zeros, it is obvious that for any . However, although holds true, is not guaranteed. Hence, we cannot claim that . Therefore, we cannot claim that .
But anyway, the deep ReLU network is indeed bounded by , due to . In fact, we claim that for most , is suppose to be in the same magnitude of : Suppose , i.e. . The variance should be randomly positive or negative. Then, if the row in is not erased to zeros, we shall have . And so that the difference between the absolute summation of ’th row in and that in will be:
According to this form, the absolute summation in one row of shall hardly show significant change comparing to the same row in . This is because most amount of variance may be cancelled out due to the randomness in . Hence, shall hold true for most when . When , we may set , so that . Then the same result shall be obtained by mathematical induction.
We use the concept “most” for many times in the above analysis. In fact, the more nodes in the hidden layers, the closer and should be. This is because the random differences caused by ReLU may be balanced in the high dimension of row and column . Given that our deep ReLU network has 3 hidden layers with 128 nodes in each layer, we assume that the probability of any input to locate in a region with is 99. We call this statement the .
Now, we are ready to interpret the experimental results in the previous section. Once again, we only use MNIST as our example. The situations with CIFAR10 and other datasets shall be the same.
We first deal with the performance of deep ReLU network trained by actual labels. We assume that the training samples are “dense”, with samples from the same class clustered together. Actually, dense and clustered samples are typical for many datasets. And these clusters usually form specific structures such as manifolds or hyperplanes. But to be specific, we assume that there is a domain in the sample space satisfying:
Every sample in belongs to the class .
For any , there exists a sample from class such that Given the MNIST dataset, this means that each pixel in is within the range of of the corresponding pixel in .
Then, for any input , suppose is its closest sample. We assume that and , where and are the small regions with fixed , corresponding to the welltrained deep ReLU network . Then, suppose the line segment (starting at and ending at ) passes through the boundaries between small regions . Then, suppose the points are on each boundary respectively.
According to the definition of , we have that , and there exists positive numbers with such that (we define and ). In addition, since the deep ReLU network is continuous, inputs on the boundary between two regions can take either from each region. That is, for .
Then, we have (for simplicity, we use to note )
According to our norm distribution assumption, 99 of will have . We assume that , which is a small range. So, the value should not be huge enough to have more than 2 or 3 with . As a result, we can say that in most cases, we have
This means that each dimension in is within the range of of the corresponding dimension in . So for any dimension in , we have that:
That is, the dimension of class in will not be surpassed by any other dimension, which means that shall be classified into . Finally, due to the ambiguity of in and the norm distribution assumption, we can conclude that the deep ReLU network is correctly generalized to most portion of the domain .
This conclusion can explain the results shown by Figure 1 and 2 well: As we said, most dense cluster domains preserve specific structures such as manifolds or hyperplanes in the input space. Hence, when moving along two samples , from the same class, or along the first principle component direction, it is very likely that is within the manifold . Hence, the classification won’t change due to the generalization of the deep ReLU network in .
However, when and come from different classes, or when moves perpendicular to the first principle component along normal directions, the most common situation is that will move from domain to domain , so that the classification will change with steady when x passes through the boundary of the two domains. But when moving from along a random direction, it is very likely that
will go beyond the hyperplane and arrive in areas where samples are sparse. Hence, the deep ReLU network are not well generalized in these areas and dimensions in
shall oscillate with chaos.But when we randomly assign the labels, the manifolds with dense clustered samples will be broken down into pieces according to samples assigned with different labels. So, there is hardly any dense domain for the deep ReLU network to produce a good generalization. As a result, moving along any directions are less stable than the case with actual labels, as shown in Figure 3. Specifically, moving along the first principle component direction has no difference than moving along any random direction. This is because the random labels break down each manifold , so that no principle component direction exists anymore.
However, even with random labels, if both and are from the same class, moving along is relatively stable. This is because the deep ReLU network is still piecewise linear and piecewise bounded by along . So if and are almost the same, then all the dimensions in cannot oscillate “at will” during the moving along , since at last each dimension in need to go back to the same. Hence, even with actual labels, moving along with the same label on and are relatively stable.
In summary, we show that a deep ReLU network does piecewise linear interpolation rather than memorization in the input space. And its interpolation is bounded mostly. Hence, if samples from the same classes form dense clusters in the input space, the deep ReLU network shall generalize well in the domains covered by the dense clusters.
Comments
There are no comments yet.