deepy
A highly extensible deep learning framework
view repo
There is plenty of theoretical and empirical evidence that depth of neural networks is a crucial ingredient for their success. However, network training becomes more difficult with increasing depth and training of very deep networks remains an open problem. In this extended abstract, we introduce a new architecture designed to ease gradient-based training of very deep networks. We refer to networks with this architecture as highway networks, since they allow unimpeded information flow across several layers on "information highways". The architecture is characterized by the use of gating units which learn to regulate the flow of information through a network. Highway networks with hundreds of layers can be trained directly using stochastic gradient descent and with a variety of activation functions, opening up the possibility of studying extremely deep and efficient architectures.
READ FULL TEXT VIEW PDF
Theoretical and empirical evidence indicates that the depth of neural
ne...
read it
The vanishing gradient problem was a major obstacle for the success of d...
read it
Deep neural network architectures have recently produced excellent resul...
read it
Training very deep networks is an important open problem in machine lear...
read it
Recurrent Neural Networks (RNNs) play a major role in the field of seque...
read it
Understanding the power of depth in feed-forward neural networks is an
o...
read it
The past year saw the introduction of new architectures such as Highway
...
read it
A highly extensible deep learning framework
An implementation of Highway Networks in Caffe
Low-rank Highway Networks
This is the implementation of Highway version of Deep convolutional network, based on https://arxiv.org/pdf/1505.00387.pdf.
Torch highway network implementation + training/evaluation on MNIST from https://arxiv.org/abs/1505.00387.
Many recent empirical breakthroughs in supervised machine learning have been achieved through the application of deep neural networks. Network depth (referring to the number of successive computation layers) has played perhaps the most important role in these successes. For instance, the top-5 image classification accuracy on the 1000-class ImageNet dataset has increased from
84% (Krizhevsky et al., 2012) to 95% (Szegedy et al., 2014; Simonyan & Zisserman, 2014) through the use of ensembles of deeper architectures and smaller receptive fields (Ciresan et al., 2011a, b, 2012) in just a few years.On the theoretical side, it is well known that deep networks can represent certain function classes exponentially more efficiently than shallow ones (e.g. the work of Håstad (1987); Håstad & Goldmann (1991) and recently of Montufar et al. (2014)). As argued by Bengio et al. (2013), the use of deep networks can offer both computational and statistical efficiency for complex tasks.
However, training deeper networks is not as straightforward as simply adding layers. Optimization of deep networks has proven to be considerably more difficult, leading to research on initialization schemes (Glorot & Bengio, 2010; Saxe et al., 2013; He et al., 2015), techniques of training networks in multiple stages (Simonyan & Zisserman, 2014; Romero et al., 2014)
or with temporary companion loss functions attached to some of the layers
(Szegedy et al., 2014; Lee et al., 2015).In this extended abstract, we present a novel architecture that enables the optimization of networks with virtually arbitrary depth. This is accomplished through the use of a learned gating mechanism for regulating information flow which is inspired by Long Short Term Memory recurrent neural networks
(Hochreiter & Schmidhuber, 1995). Due to this gating mechanism, a neural network can have paths along which information can flow across several layers without attenuation. We call such paths information highways, and such networks highway networks.In preliminary experiments, we found that highway networks as deep as 900 layers can be optimized using simple Stochastic Gradient Descent (SGD) with momentum. For up to 100 layers we compare their training behavior to that of traditional networks with normalized initialization (Glorot & Bengio, 2010; He et al., 2015). We show that optimization of highway networks is virtually independent of depth, while for traditional networks it suffers significantly as the number of layers increases. We also show that architectures comparable to those recently presented by Romero et al. (2014) can be directly trained to obtain similar test set accuracy on the CIFAR-10 dataset without the need for a pre-trained teacher network.
We use boldface letters for vectors and matrices, and italicized capital letters to denote transformation functions.
and denote vectors of zeros and ones respectively, anddenotes an identity matrix. The function
is defined as .A plain feedforward neural network typically consists of layers where the layer (
) applies a non-linear transform
(parameterized by ) on its input to produce its output . Thus, is the input to the network and is the network’s output. Omitting the layer index and biases for clarity,(1) |
is usually an affine transform followed by a non-linear activation function, but in general it may take other forms.
For a highway network, we additionally define two non-linear transforms and such that
(2) |
We refer to as the transform gate and as the carry gate, since they express how much of the output is produced by transforming the input and carrying it, respectively. For simplicity, in this paper we set , giving
(3) |
The dimensionality of and must be the same for Equation (missing) to be valid. Note that this re-parametrization of the layer transformation is much more flexible than Equation (missing). In particular, observe that
(4) |
Similarly, for the Jacobian of the layer transform,
(5) |
Thus, depending on the output of the transform gates, a highway layer can smoothly vary its behavior between that of a plain layer and that of a layer which simply passes its inputs through. Just as a plain layer consists of multiple computing units such that the unit computes , a highway network consists of multiple blocks such that the block computes a block state and transform gate output . Finally, it produces the block output , which is connected to the next layer.
As mentioned earlier, Equation (missing) requires that the dimensionality of and be the same. In cases when it is desirable to change the size of the representation, one can replace with
obtained by suitably sub-sampling or zero-padding
. Another alternative is to use a plain layer (without highways) to change dimensionality and then continue with stacking highway layers. This is the alternative we use in this study.Convolutional highway layers are constructed similar to fully connected layers. Weight-sharing and local receptive fields are utilized for both and transforms. We use zero-padding to ensure that the block state and transform gate feature maps are the same size as the input.
For plain deep networks, training with SGD stalls at the beginning unless a specific weight initialization scheme is used such that the variance of the signals during forward and backward propagation is preserved initially
(Glorot & Bengio, 2010; He et al., 2015). This initialization depends on the exact functional form of .For highway layers, we use the transform gate defined as , where is the weight matrix and
the bias vector for the transform gates. This suggests a simple initialization scheme which is independent of the nature of
: can be initialized with a negative value (e.g. -1, -3 etc.) such that the network is initially biased towards carry behavior. This scheme is strongly inspired by the proposal of Gers et al. (1999) to initially bias the gates in a Long Short-Term Memory recurrent network to help bridge long-term temporal dependencies early in learning. Note that , so the conditions in Equation (missing) can never be exactly true.In our experiments, we found that a negative bias initialization was sufficient for learning to proceed in very deep networks for various zero-mean initial distributions of and different activation functions used by . This is significant property since in general it may not be possible to find effective initialization schemes for many choices of .
Very deep plain networks become difficult to optimize even if using the variance-preserving initialization scheme form (He et al., 2015). To show that highway networks do not suffer from depth in the same way we train run a series of experiments on the MNIST digit classification dataset. We measure the cross entropy error on the training set, to investigate optimization, without conflating them with generalization issues.
We train both plain networks and highway networks with the same architecture and varying depth. The first layer is always a regular fully-connected layer followed by 9, 19, 49, or 99 fully-connected plain or highway layers and a single softmax output layer. The number of units in each layer is kept constant and it is 50 for highways and 71 for plain networks. That way the number of parameters is roughly the same for both. To make the comparison fair we run a random search of 40 runs for both plain and highway networks to find good settings for the hyperparameters. We optimized the initial learning rate, momentum, learning rate decay rate, activation function for (either ReLU or tanh) and, for highway networks, the value for the transform gate bias (between -1 and -10). All other weights were initialized following the scheme introduced by (He et al., 2015).
The convergence plots for the best performing networks for each depth can be seen in Figure 1
. While for 10 layers plain network show very good performance, their performance significantly degrades as depth increases. Highway networks on the other hand do not seem to suffer from an increase in depth at all. The final result of the 100 layer highway network is about 1 order of magnitude better than the 10 layer one, and is on par with the 10 layer plain network. In fact, we started training a similar 900 layer highway network on CIFAR-100 which is only at 80 epochs as of now, but so far has shown no signs of optimization difficulties. It is also worth pointing out that the highway networks always converge significantly faster than the plain ones.
Network | Number of Layers | Number of Parameters | Accuracy |
---|---|---|---|
Fitnet Results reported by Romero et al. (2014) | |||
Teacher | 5 | 9M | 90.18% |
Fitnet 1 | 11 | 250K | 89.01% |
Fitnet 2 | 11 | 862K | 91.06% |
Fitnet 3 | 13 | 1.6M | 91.10% |
Fitnet 4 | 19 | 2.5M | 91.61% |
Highway networks | |||
Highway 1 (Fitnet 1) | 11 | 236K | 89.18% |
Highway 2 (Fitnet 4) | 19 | 2.3M | 92.24% |
Highway 3* | 19 | 1.4M | 90.68% |
Highway 4* | 32 | 1.25M | 90.34% |
using maxout networks are also shown. Fitnets were trained using a two step training procedure using soft targets from the trained Teacher network, which was trained using backpropagation. We trained all highway networks directly using backpropagation. * indicates networks which were trained only on a set of 40K out of 50K examples in the training set.
Deep highway networks are easy to optimize, but are they also beneficial for supervised learning where we are interested in generalization performance on a test set? To address this question, we compared highway networks to the thin and deep architectures termed
Fitnets proposed recently by Romero et al. (2014) on the CIFAR-10 dataset augmented with random translations. Results are summarized in Table 1.Romero et al. (2014) reported that training using plain backpropogation was only possible for maxout networks with depth up to 5 layers when number of parameters was limited to 250K and number of multiplications to 30M. Training of deeper networks was only possible through the use of a two-stage training procedure and addition of soft targets produced from a pre-trained shallow teacher network (hint-based training). Similarly it was only possible to train 19-layer networks with a budget of 2.5M parameters using hint-based training.
We found that it was easy to train highway networks with number of parameters and operations comparable to fitnets directly using backpropagation. As shown in Table 1, Highway 1 and Highway 4, which are based on the architecture of Fitnet 1 and Fitnet 4 respectively obtain similar or higher accuracy on the test set. We were also able to train thinner and deeper networks: a 19-layer highway network with 1.4M parameters and a 32-layer highway network with 1.25M parameter both perform similar to the teacher network of Romero et al. (2014).
In Figure 2 we show some inspections on the inner workings of the best^{1}^{1}1obtained via random search over hyperparameters to minimize the best training set error achieved using each configuration 50 hidden layer fully-connected highway networks trained on MNIST (top row) and CIFAR-100 (bottom row). The first three columns show, for each transform gate, the bias, the mean activity over 10K random samples, and the activity for a single random sample respectively. The block outputs for the same single sample are displayed in the last column.
The transform gate biases of the two networks were initialized to -2 and -4 respectively. It is interesting to note that contrary to our expectations most biases actually decreased further during training. For the CIFAR-100 network the biases increase with depth forming a gradient. Curiously this gradient is inversely correlated with the average activity of the transform gates as seen in the second column. This indicates that the strong negative biases at low depths are not used to shut down the gates, but to make them more selective. This behavior is also suggested by the fact that the transform gate activity for a single example (column 3) is very sparse. This effect is more pronounced for the CIFAR-100 network, but can also be observed to a lesser extent in the MNIST network.
The last column of Figure 2 displays the block outputs and clearly visualizes the concept of “information highways”. Most of the outputs stay constant over many layers forming a pattern of stripes. Most of the change in outputs happens in the early layers ( for MNIST and for CIFAR-100). We hypothesize that this difference is due to the higher complexity of the CIFAR-100 dataset.
In summary it is clear that highway networks actually utilize the gating mechanism to pass information almost unchanged through many layers. This mechanism serves not just as a means for easier training, but is also heavily used to route information in a trained network. We observe very selective activity of the transform gates, varying strongly in reaction to the current input patterns.
Learning to route information through neural networks has helped to scale up their application to challenging problems by improving credit assignment and making training easier (Srivastava et al., 2015). Even so, training very deep networks has remained difficult, especially without considerably increasing total network size.
Highway networks are novel neural network architectures which enable the training of extremely deep networks using simple SGD. While the traditional plain neural architectures become increasingly difficult to train with increasing network depth (even with variance-preserving initialization), our experiments show that optimization of highway networks is not hampered even as network depth increases to a hundred layers.
The ability to train extremely deep networks opens up the possibility of studying the impact of depth on complex problems without restrictions. Various activation functions which may be more suitable for particular problems but for which robust initialization schemes are unavailable can be used in deep highway networks. Future work will also attempt to improve the understanding of learning in highway networks.
This research was supported by the by EU project “NASCENCE” (FP7-ICT-317662). We gratefully acknowledge the support of NVIDIA Corporation with the donation of the Tesla K40 GPUs used for this research.
IEEE Conference on Computer Vision and Pattern Recognition
, 2012.Flexible, high performance convolutional neural networks for image classification.
In IJCAI, 2011b. URL http://www.aaai.org/ocs/index.php/IJCAI/IJCAI11/paper/download/3098/3425%0020http://dl.acm.org/citation.cfm?id=2283603.International Conference on Artificial Intelligence and Statistics
, pp. 249–256, 2010. URL http://machinelearning.wustl.edu/mlpapers/paper_files/AISTATS2010_GlorotB10.pdf.
Comments
There are no comments yet.