Towards Faster Training of Global Covariance Pooling Networks by Iterative Matrix Square Root Normalization

12/04/2017
by   Peihua Li, et al.
0

Global covariance pooling in Convolutional neural neworks has achieved impressive improvement over the classical first-order pooling. Recent works have shown matrix square root normalization plays a central role in achieving state-of-the-art performance. However, existing methods depending heavily on eigenvalue decomposition (EIG) or singular value decomposition (SVD), suffering from inefficient training due to limited support of EIG and SVD on GPU. Towards addressing this problem, we propose an iterative matrix square root normalization method for end-to-end training of covariance pooling networks. At the core of our method is a meta-layer designed with loop-embedded directed graph structure. The meta-layer consists of three consecutive nonlinear structured layers, which perform pre-normalization, coupled matrix iteration and post-compensation, respectively. Our method is much faster than EIG or SVD based ones, since it involves only matrix multiplications, suitable for parallel implementation on GPU. Moreover, the proposed network with ResNet architecture can converge in much less epochs, further accelerating network training. On large-scale ImageNet, we achieve competitive performance superior to existing covariance pooling networks. By finetuning our models pretrained on ImageNet, we establish state-of-the-art results on three challenging fine-grained benchmarks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/06/2021

Why Approximate Matrix Square Root Outperforms Accurate SVD in Global Covariance Pooling?

Global covariance pooling (GCP) aims at exploiting the second-order stat...
research
06/05/2019

Compact Approximation for Polynomial of Covariance Feature

Covariance pooling is a feature pooling method with good classification ...
research
04/15/2019

Deep CNNs Meet Global Covariance Pooling: Better Representation and Generalization

Compared with global average pooling in existing deep convolutional neur...
research
07/21/2017

Improved Bilinear Pooling with CNNs

Bilinear pooling of Convolutional Neural Network (CNN) features [22, 23]...
research
07/05/2022

Improving Covariance Conditioning of the SVD Meta-layer by Orthogonality

Inserting an SVD meta-layer into neural networks is prone to make the co...
research
02/20/2018

MoNet: Moments Embedding Network

Bilinear pooling has been recently proposed as a feature encoding layer,...
research
03/23/2017

Is Second-order Information Helpful for Large-scale Visual Recognition?

By stacking layers of convolution and nonlinearity, convolutional networ...

Please sign up or login with your details

Forgot password? Click here to reset