Self-Supervised Implicit Attention: Guided Attention by The Model Itself

06/15/2022
by   Jinyi Wu, et al.
0

We propose Self-Supervised Implicit Attention (SSIA), a new approach that adaptively guides deep neural network models to gain attention by exploiting the properties of the models themselves. SSIA is a novel attention mechanism that does not require any extra parameters, computation, or memory access costs during inference, which is in contrast to existing attention mechanism. In short, by considering attention weights as higher-level semantic information, we reconsidered the implementation of existing attention mechanisms and further propose generating supervisory signals from higher network layers to guide lower network layers for parameter updates. We achieved this by building a self-supervised learning task using the hierarchical features of the network itself, which only works at the training stage. To verify the effectiveness of SSIA, we performed a particular implementation (called an SSIA block) in convolutional neural network models and validated it on several image classification datasets. The experimental results show that an SSIA block can significantly improve the model performance, even outperforms many popular attention methods that require additional parameters and computation costs, such as Squeeze-and-Excitation and Convolutional Block Attention Module. Our implementation will be available on GitHub.

READ FULL TEXT

page 9

page 12

research
10/02/2021

Implicit and Explicit Attention for Zero-Shot Learning

Most of the existing Zero-Shot Learning (ZSL) methods focus on learning ...
research
05/25/2023

Perturbation-based Self-supervised Attention for Attention Bias in Text Classification

In text classification, the traditional attention mechanisms usually foc...
research
02/21/2021

A Hierarchical Conditional Random Field-based Attention Mechanism Approach for Gastric Histopathology Image Classification

In the Gastric Histopathology Image Classification (GHIC) tasks, which i...
research
06/21/2023

Constant Memory Attention Block

Modern foundation model architectures rely on attention mechanisms to ef...
research
11/14/2020

Graph-Based Neural Network Models with Multiple Self-Supervised Auxiliary Tasks

Self-supervised learning is currently gaining a lot of attention, as it ...
research
11/10/2021

Learning to ignore: rethinking attention in CNNs

Recently, there has been an increasing interest in applying attention me...
research
06/20/2021

Improving Ultrasound Tongue Image Reconstruction from Lip Images Using Self-supervised Learning and Attention Mechanism

Speech production is a dynamic procedure, which involved multi human org...

Please sign up or login with your details

Forgot password? Click here to reset