Entropy Transformer Networks: A Learning Approach via Tangent Bundle Data Manifold

07/24/2023
by   Pourya Shamsolmoali, et al.
0

This paper focuses on an accurate and fast interpolation approach for image transformation employed in the design of CNN architectures. Standard Spatial Transformer Networks (STNs) use bilinear or linear interpolation as their interpolation, with unrealistic assumptions about the underlying data distributions, which leads to poor performance under scale variations. Moreover, STNs do not preserve the norm of gradients in propagation due to their dependency on sparse neighboring pixels. To address this problem, a novel Entropy STN (ESTN) is proposed that interpolates on the data manifold distributions. In particular, random samples are generated for each pixel in association with the tangent space of the data manifold and construct a linear approximation of their intensity values with an entropy regularizer to compute the transformer parameters. A simple yet effective technique is also proposed to normalize the non-zero values of the convolution operation, to fine-tune the layers for gradients' norm-regularization during training. Experiments on challenging benchmarks show that the proposed ESTN can improve predictive accuracy over a range of computer vision tasks, including image reconstruction, and classification, while reducing the computational cost.

READ FULL TEXT
research
04/03/2023

VTAE: Variational Transformer Autoencoder with Manifolds Learning

Deep generative models have demonstrated successful applications in lear...
research
01/22/2019

Linearized Multi-Sampling for Differentiable Image Transformation

We propose a novel image sampling method for differentiable image transf...
research
06/01/2023

Dilated Convolution with Learnable Spacings: beyond bilinear interpolation

Dilated Convolution with Learnable Spacings (DCLS) is a recently propose...
research
09/18/2017

Minimal Effort Back Propagation for Convolutional Neural Networks

As traditional neural network consumes a significant amount of computing...
research
04/28/2021

Twins: Revisiting Spatial Attention Design in Vision Transformers

Very recently, a variety of vision transformer architectures for dense p...
research
09/07/2018

MixUp as Locally Linear Out-Of-Manifold Regularization

MixUp, a data augmentation approach through mixing random samples, has b...

Please sign up or login with your details

Forgot password? Click here to reset