Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning

02/23/2016
by   Christian Szegedy, et al.
0

Very deep convolutional networks have been central to the largest advances in image recognition performance in recent years. One example is the Inception architecture that has been shown to achieve very good performance at relatively low computational cost. Recently, the introduction of residual connections in conjunction with a more traditional architecture has yielded state-of-the-art performance in the 2015 ILSVRC challenge; its performance was similar to the latest generation Inception-v3 network. This raises the question of whether there are any benefit in combining the Inception architecture with residual connections. Here we give clear empirical evidence that training with residual connections accelerates the training of Inception networks significantly. There is also some evidence of residual Inception networks outperforming similarly expensive Inception networks without residual connections by a thin margin. We also present several new streamlined architectures for both residual and non-residual Inception networks. These variations improve the single-frame recognition performance on the ILSVRC 2012 classification task significantly. We further demonstrate how proper activation scaling stabilizes the training of very wide residual Inception networks. With an ensemble of three residual and one Inception-v4, we achieve 3.08 percent top-5 error on the test set of the ImageNet classification (CLS) challenge

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/28/2017

Improved Inception-Residual Convolutional Neural Network for Object Recognition

Machine learning and computer vision have driven many of the greatest ad...
research
02/04/2017

Wide-Residual-Inception Networks for Real-time Object Detection

Since convolutional neural network(CNN)models emerged,several tasks in c...
research
09/29/2020

Deep discriminant analysis for task-dependent compact network search

Most of today's popular deep architectures are hand-engineered for gener...
research
08/02/2016

CUHK & ETHZ & SIAT Submission to ActivityNet Challenge 2016

This paper presents the method that underlies our submission to the untr...
research
05/22/2017

Facial Affect Estimation in the Wild Using Deep Residual and Convolutional Networks

Automated affective computing in the wild is a challenging task in the f...
research
04/12/2020

Inception LSTM

In this paper, we proposed a novel deep-learning method called Inception...
research
11/17/2016

PolyNet: A Pursuit of Structural Diversity in Very Deep Networks

A number of studies have shown that increasing the depth or width of con...

Please sign up or login with your details

Forgot password? Click here to reset