Making Sense of Hidden Layer Information in Deep Networks by Learning Hierarchical Targets

05/03/2015
by   Abhinav Tushar, et al.
0

This paper proposes an architecture for deep neural networks with hidden layer branches that learn targets of lower hierarchy than final layer targets. The branches provide a channel for enforcing useful information in hidden layer which helps in attaining better accuracy, both for the final layer and hidden layers. The shared layers modify their weights using the gradients of all cost functions higher than the branching layer. This model provides a flexible inference system with many levels of targets which is modular and can be used efficiently in situations requiring different levels of results according to complexity. This paper applies the idea to a text classification task on 20 Newsgroups data set with two level of hierarchical targets and a comparison is made with training without the use of hidden layer branches.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/07/2022

Approximation error of single hidden layer neural networks with fixed weights

This paper provides an explicit formula for the approximation error of s...
research
08/05/2023

Neural Collapse in the Intermediate Hidden Layers of Classification Neural Networks

Neural Collapse (NC) gives a precise description of the representations ...
research
06/26/2021

Interflow: Aggregating Multi-layer Feature Mappings with Attention Mechanism

Traditionally, CNN models possess hierarchical structures and utilize th...
research
11/14/2017

Exploiting Layerwise Convexity of Rectifier Networks with Sign Constrained Weights

By introducing sign constraints on the weights, this paper proposes sign...
research
10/03/2019

Regularizing Neural Networks via Stochastic Branch Layers

We introduce a novel stochastic regularization technique for deep neural...
research
04/30/2020

Binary autoencoder with random binary weights

Here is presented an analysis of an autoencoder with binary activations ...
research
07/08/2018

Separability is not the best goal for machine learning

Neural networks use their hidden layers to transform input data into lin...

Please sign up or login with your details

Forgot password? Click here to reset