Top-down Flow Transformer Networks

12/06/2017
by   Zhiwei Jia, et al.
0

We study the deformation fields of feature maps across convolutional network layers under explicit top-down spatial transformations. We propose top-down flow transformer (TFT) by focusing on three transformations: translation, rotation, and scaling. We learn flow transformation generators that are able to account for the hidden layer deformations while maintaining the overall consistency across layers. The learned generators are shown to capture the underlying feature transformation processes that are independent of the particular training images. We observe favorable experimental results compared to the existing methods that tie transformations to fixed datasets. A comprehensive study on various datasets including MNIST, shapes, and natural images with both inner and inter datasets (trained on MNIST and validated in a number of datasets) evaluation demonstrates the advantages of our proposed TFT framework, which can be adopted in a variety of computer vision applications.

READ FULL TEXT

page 4

page 6

page 8

research
11/23/2015

What Happened to My Dog in That Network: Unraveling Top-down Generators in Convolutional Neural Networks

Top-down information plays a central role in human perception, but plays...
research
05/16/2018

Spatial Transformer Introspective Neural Network

Natural images contain many variations such as illumination differences,...
research
01/14/2020

The problems with using STNs to align CNN feature maps

Spatial transformer networks (STNs) were designed to enable CNNs to lear...
research
05/24/2017

Deep Rotation Equivariant Network

Recently, learning equivariant representations has attracted considerabl...
research
01/29/2018

Hierarchical Spatial Transformer Network

Computer vision researchers have been expecting that neural networks hav...
research
01/14/2022

Perspective Transformation Layer

Incorporating geometric transformations that reflect the relative positi...
research
09/21/2020

Learning Representation for Mixed Data Types with a Nonlinear Deep Encoder-Decoder Framework

Representation of data on mixed variables, numerical and categorical typ...

Please sign up or login with your details

Forgot password? Click here to reset