DeepAI AI Chat
Log In Sign Up

Backpropagation Training for Fisher Vectors within Neural Networks

by   Patrick Wieschollek, et al.

Fisher-Vectors (FV) encode higher-order statistics of a set of multiple local descriptors like SIFT features. They already show good performance in combination with shallow learning architectures on visual recognitions tasks. Current methods using FV as a feature descriptor in deep architectures assume that all original input features are static. We propose a framework to jointly learn the representation of original features, FV parameters and parameters of the classifier in the style of traditional neural networks. Our proof of concept implementation improves the performance of FV on the Pascal Voc 2007 challenge in a multi-GPU setting in comparison to a default SVM setting. We demonstrate that FV can be embedded into neural networks at arbitrary positions, allowing end-to-end training with back-propagation.


page 1

page 2

page 3

page 4


What is the right way to represent document images?

In this article we study the problem of document image representation ba...

End-to-end Learning of a Fisher Vector Encoding for Part Features in Fine-grained Recognition

Part-based approaches for fine-grained recognition do not show the expec...

Training Deep Networks with Structured Layers by Matrix Backpropagation

Deep neural network architectures have recently produced excellent resul...

High Order Structure Descriptors for Scene Images

Structure information is ubiquitous in natural scene images and it plays...

GPU-FV: Realtime Fisher Vector and Its Applications in Video Monitoring

Fisher vector has been widely used in many multimedia retrieval and visu...

End-to-end learning potentials for structured attribute prediction

We present a structured inference approach in deep neural networks for m...

Understanding the Fisher Vector: a multimodal part model

Fisher Vectors and related orderless visual statistics have demonstrated...