Is Second-order Information Helpful for Large-scale Visual Recognition?

03/23/2017
by   Peihua Li, et al.
0

By stacking layers of convolution and nonlinearity, convolutional networks (ConvNets) effectively learn from low-level to high-level features and discriminative representations. Since the end goal of large-scale recognition is to delineate complex boundaries of thousands of classes, adequate exploration of feature distributions is important for realizing full potentials of ConvNets. However, state-of-the-art works concentrate only on deeper or wider architecture design, while rarely exploring feature statistics higher than first-order. We take a step towards addressing this problem. Our method consists in covariance pooling, instead of the most commonly used first-order pooling, of high-level convolutional features. The main challenges involved are robust covariance estimation given a small sample of large-dimensional features and usage of the manifold structure of covariance matrices. To address these challenges, we present a Matrix Power Normalized Covariance (MPN-COV) method. We develop forward and backward propagation formulas regarding the nonlinear matrix functions such that MPN-COV can be trained end-to-end. In addition, we analyze both qualitatively and quantitatively its advantage over the well-known Log-Euclidean metric. On the ImageNet 2012 validation set, by combining MPN-COV we achieve over 4 respectively; integration of MPN-COV into 50-layer ResNet outperforms ResNet-101 and is comparable to ResNet-152. The source code will be available on the project page: http://www.peihuali.org/MPN-COV.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/29/2018

Global Second-order Pooling Convolutional Networks

Deep Convolutional Networks (ConvNets) are fundamental to, besides large...
research
11/29/2018

Global Second-order Pooling Neural Networks

Deep Convolutional Networks (ConvNets) are fundamental to, besides large...
research
04/15/2019

Deep CNNs Meet Global Covariance Pooling: Better Representation and Generalization

Compared with global average pooling in existing deep convolutional neur...
research
02/14/2022

Discriminability-enforcing loss to improve representation learning

During the training process, deep neural networks implicitly learn to re...
research
12/04/2017

Towards Faster Training of Global Covariance Pooling Networks by Iterative Matrix Square Root Normalization

Global covariance pooling in Convolutional neural neworks has achieved i...
research
06/24/2018

A Deeper Look at Power Normalizations

Power Normalizations (PN) are very useful non-linear operators in the co...
research
01/15/2022

Multi-level Second-order Few-shot Learning

We propose a Multi-level Second-order (MlSo) few-shot learning network f...

Please sign up or login with your details

Forgot password? Click here to reset