Wider or Deeper: Revisiting the ResNet Model for Visual Recognition

11/30/2016
by   Zifeng Wu, et al.
0

The trend towards increasingly deep neural networks has been driven by a general observation that increasing depth increases the performance of a network. Recently, however, evidence has been amassing that simply increasing depth may not be the best way to increase performance, particularly given other limitations. Investigations into deep residual networks have also suggested that they may not in fact be operating as a single deep network, but rather as an ensemble of many relatively shallow networks. We examine these issues, and in doing so arrive at a new interpretation of the unravelled view of deep residual networks which explains some of the behaviours that have been observed experimentally. As a result, we are able to derive a new, shallower, architecture of residual networks which significantly outperforms much deeper models such as ResNet-200 on the ImageNet classification dataset. We also show that this performance is transferable to other problem domains by developing a semantic segmentation approach which outperforms the state-of-the-art by a remarkable margin on datasets including PASCAL VOC, PASCAL Context, and Cityscapes. The architecture that we propose thus outperforms its comparators, including very deep ResNets, and yet is more efficient in memory use and sometimes also in training time. The code and models are available at https://github.com/itijyou/ademxapp

READ FULL TEXT

page 12

page 13

page 14

page 15

page 16

page 17

research
05/23/2016

Wide Residual Networks

Deep residual networks were shown to be able to scale up to thousands of...
research
09/19/2016

Multi-Residual Networks: Improving the Speed and Accuracy of Residual Networks

In this article, we take one step toward understanding the learning beha...
research
03/07/2017

Sharing Residual Units Through Collective Tensor Factorization in Deep Neural Networks

Residual units are wildly used for alleviating optimization difficulties...
research
10/09/2022

Stimulative Training of Residual Networks: A Social Psychology Perspective of Loafing

Residual networks have shown great success and become indispensable in t...
research
11/16/2016

Aggregated Residual Transformations for Deep Neural Networks

We present a simple, highly modularized network architecture for image c...
research
11/01/2021

RMNet: Equivalently Removing Residual Connection from Networks

Although residual connection enables training very deep neural networks,...
research
07/06/2017

Dual Path Networks

In this work, we present a simple, highly efficient and modularized Dual...

Please sign up or login with your details

Forgot password? Click here to reset