SIRe-Networks: Skip Connections over Interlaced Multi-Task Learning and Residual Connections for Structure Preserving Object Classification

10/06/2021
by   Danilo Avola, et al.
0

Improving existing neural network architectures can involve several design choices such as manipulating the loss functions, employing a diverse learning strategy, exploiting gradient evolution at training time, optimizing the network hyper-parameters, or increasing the architecture depth. The latter approach is a straightforward solution, since it directly enhances the representation capabilities of a network; however, the increased depth generally incurs in the well-known vanishing gradient problem. In this paper, borrowing from different methods addressing this issue, we introduce an interlaced multi-task learning strategy, defined SIRe, to reduce the vanishing gradient in relation to the object classification task. The presented methodology directly improves a convolutional neural network (CNN) by enforcing the input image structure preservation through interlaced auto-encoders, and further refines the base network architecture by means of skip and residual connections. To validate the presented methodology, a simple CNN and various implementations of famous networks are extended via the SIRe strategy and extensively tested on the CIFAR100 dataset; where the SIRe-extended architectures achieve significantly increased performances across all models, thus confirming the presented approach effectiveness.

READ FULL TEXT

page 3

page 5

research
03/23/2017

Single Image Super-resolution via a Lightweight Residual Convolutional Neural Network

Recent years have witnessed great success of convolutional neural networ...
research
08/31/2018

Multi-Cell Multi-Task Convolutional Neural Networks for Diabetic Retinopathy Grading Kang

Diabetic Retinopathy (DR) is a non-negligible eye disease among patients...
research
03/27/2019

Training Quantized Network with Auxiliary Gradient Module

In this paper, we seek to tackle two challenges in training low-precisio...
research
10/27/2020

A Multi-task Two-stream Spatiotemporal Convolutional Neural Network for Convective Storm Nowcasting

The goal of convective storm nowcasting is local prediction of severe an...
research
04/26/2018

IamNN: Iterative and Adaptive Mobile Neural Network for Efficient Image Classification

Deep residual networks (ResNets) made a recent breakthrough in deep lear...
research
10/08/2021

New Insights into Graph Convolutional Networks using Neural Tangent Kernels

Graph Convolutional Networks (GCNs) have emerged as powerful tools for l...
research
03/30/2020

Dataless Model Selection with the Deep Frame Potential

Choosing a deep neural network architecture is a fundamental problem in ...

Please sign up or login with your details

Forgot password? Click here to reset