Input Fast-Forwarding for Better Deep Learning

05/23/2017
by   Ahmed Ibrahim, et al.
0

This paper introduces a new architectural framework, known as input fast-forwarding, that can enhance the performance of deep networks. The main idea is to incorporate a parallel path that sends representations of input values forward to deeper network layers. This scheme is substantially different from "deep supervision" in which the loss layer is re-introduced to earlier layers. The parallel path provided by fast-forwarding enhances the training process in two ways. First, it enables the individual layers to combine higher-level information (from the standard processing path) with lower-level information (from the fast-forward path). Second, this new architecture reduces the problem of vanishing gradients substantially because the fast-forwarding path provides a shorter route for gradient backpropagation. In order to evaluate the utility of the proposed technique, a Fast-Forward Network (FFNet), with 20 convolutional layers along with parallel fast-forward paths, has been created and tested. The paper presents empirical results that demonstrate improved learning capacity of FFNet due to fast-forwarding, as compared to GoogLeNet (with deep supervision) and CaffeNet, which are 4x and 18x larger in size, respectively. All of the source code and deep learning models described in this paper will be made available to the entire research community

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/08/2017

Forward Thinking: Building and Training Neural Networks One Layer at a Time

We present a general framework for training deep neural networks without...
research
05/29/2023

Intelligent gradient amplification for deep neural networks

Deep learning models offer superior performance compared to other machin...
research
05/02/2017

Redundancy in active paths of deep networks: a random active path model

Deep learning has become a powerful and popular tool for a variety of ma...
research
09/15/2023

Make Deep Networks Shallow Again

Deep neural networks have a good success record and are thus viewed as t...
research
11/09/2022

Designing Network Design Strategies Through Gradient Path Analysis

Designing a high-efficiency and high-quality expressive network architec...
research
11/06/2016

The Shallow End: Empowering Shallower Deep-Convolutional Networks through Auxiliary Outputs

The depth is one of the key factors behind the great success of convolut...
research
10/12/2021

SoftNeuro: Fast Deep Inference using Multi-platform Optimization

Faster inference of deep learning models is highly demanded on edge devi...

Please sign up or login with your details

Forgot password? Click here to reset