Exploiting SIFT Descriptor for Rotation Invariant Convolutional Neural Network

03/30/2019
by   Abhay Kumar, et al.
0

This paper presents a novel approach to exploit the distinctive invariant features in convolutional neural network. The proposed CNN model uses Scale Invariant Feature Transform (SIFT) descriptor instead of the max-pooling layer. Max-pooling layer discards the pose, i.e., translational and rotational relationship between the low-level features, and hence unable to capture the spatial hierarchies between low and high level features. The SIFT descriptor layer captures the orientation and the spatial relationship of the features extracted by convolutional layer. The proposed SIFT Descriptor CNN therefore combines the feature extraction capabilities of CNN model and rotation invariance of SIFT descriptor. Experimental results on the MNIST and fashionMNIST datasets indicates reasonable improvements over conventional methods available in literature.

READ FULL TEXT
research
06/21/2020

Rotation Invariant Deep CBIR

Introduction of Convolutional Neural Networks has improved results on al...
research
06/17/2019

Modeling Music Modality with a Key-Class Invariant Pitch Chroma CNN

This paper presents a convolutional neural network (CNN) that uses input...
research
09/27/2018

Collective behavior recognition using compact descriptors

This paper presents a novel hierarchical approach for collective behavio...
research
04/24/2015

Object Level Deep Feature Pooling for Compact Image Representation

Convolutional Neural Network (CNN) features have been successfully emplo...
research
04/27/2019

Learning to Fuse Local Geometric Features for 3D Rigid Data Matching

This paper presents a simple yet very effective data-driven approach to ...
research
08/28/2013

A proposition of a robust system for historical document images indexation

Characterizing noisy or ancient documents is a challenging problem up to...
research
05/28/2011

Scale-Invariant Local Descriptor for Event Recognition in 1D Sensor Signals

In this paper, we introduce a shape-based, time-scale invariant feature ...

Please sign up or login with your details

Forgot password? Click here to reset