Near-Optimal Glimpse Sequences for Improved Hard Attention Neural Network Training

06/13/2019
by   William Harvey, et al.
0

We introduce the use of Bayesian optimal experimental design techniques for generating glimpse sequences to use in semi-supervised training of hard attention networks. Hard attention holds the promise of greater energy efficiency and superior inference performance. Employing such networks for image classification usually involves choosing a sequence of glimpse locations from a stochastic policy. As the outputs of observations are typically non-differentiable with respect to their glimpse locations, unsupervised gradient learning of such a policy requires REINFORCE-style updates. Also, the only reward signal is the final classification accuracy. For these reasons hard attention networks, despite their promise, have not achieved the wide adoption that soft attention networks have and, in many practical settings, are difficult to train. We find that our method for semi-supervised training makes it easier and faster to train hard attention networks and correspondingly could make them practical to consider in situations where they were not before.

READ FULL TEXT

page 4

page 6

page 14

page 15

page 16

page 17

research
04/08/2020

Empirical Perspectives on One-Shot Semi-supervised Learning

One of the greatest obstacles in the adoption of deep neural networks fo...
research
12/04/2018

Ladder Networks for Semi-Supervised Hyperspectral Image Classification

We used the Ladder Network [Rasmus et al. (2015)] to perform Hyperspectr...
research
09/22/2015

Learning Wake-Sleep Recurrent Attention Models

Despite their success, convolutional neural networks are computationally...
research
10/15/2020

Semi-supervised NMF Models for Topic Modeling in Learning Tasks

We propose several new models for semi-supervised nonnegative matrix fac...
research
10/04/2021

Adding Quaternion Representations to Attention Networks for Classification

This paper introduces a novel modification to axial-attention networks t...
research
05/24/2016

Hierarchical Memory Networks

Memory networks are neural networks with an explicit memory component th...
research
05/16/2017

Learning Hard Alignments with Variational Inference

There has recently been significant interest in hard attention models fo...

Please sign up or login with your details

Forgot password? Click here to reset