Interflow: Aggregating Multi-layer Feature Mappings with Attention Mechanism

06/26/2021
by   Zhicheng Cai, et al.
0

Traditionally, CNN models possess hierarchical structures and utilize the feature mapping of the last layer to obtain the prediction output. However, it can be difficulty to settle the optimal network depth and make the middle layers learn distinguished features. This paper proposes the Interflow algorithm specially for traditional CNN models. Interflow divides CNNs into several stages according to the depth and makes predictions by the feature mappings in each stage. Subsequently, we input these prediction branches into a well-designed attention module, which learns the weights of these prediction branches, aggregates them and obtains the final output. Interflow weights and fuses the features learned in both shallower and deeper layers, making the feature information at each stage processed reasonably and effectively, enabling the middle layers to learn more distinguished features, and enhancing the model representation ability. In addition, Interflow can alleviate gradient vanishing problem, lower the difficulty of network depth selection, and lighten possible over-fitting problem by introducing attention mechanism. Besides, it can avoid network degradation as a byproduct. Compared with the original model, the CNN model with Interflow achieves higher test accuracy on multiple benchmark datasets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/24/2023

Multi-stage feature decorrelation constraints for improving CNN classification performance

For the convolutional neural network (CNN) used for pattern classificati...
research
05/03/2015

Making Sense of Hidden Layer Information in Deep Networks by Learning Hierarchical Targets

This paper proposes an architecture for deep neural networks with hidden...
research
09/15/2017

Query-based Attention CNN for Text Similarity Map

In this paper, we introduce Query-based Attention CNN(QACNN) for Text Si...
research
05/07/2019

Intentional Attention Mask Transformation for Robust CNN Classification

Convolutional Neural Networks have achieved impressive results in variou...
research
03/24/2020

Dynamic Hierarchical Mimicking Towards Consistent Optimization Objectives

While the depth of modern Convolutional Neural Networks (CNNs) surpasses...
research
11/23/2016

Adaptive Feature Abstraction for Translating Video to Text

Previous models for video captioning often use the output from a specifi...
research
06/17/2021

Deep Learning Through the Lens of Example Difficulty

Existing work on understanding deep learning often employs measures that...

Please sign up or login with your details

Forgot password? Click here to reset