End-to-End Deep Kronecker-Product Matching for Person Re-identification

Person re-identification aims to robustly measure similarities between person images. The significant variation of person poses and viewing angles challenges for accurate person re-identification. The spatial layout and correspondences between query person images are vital information for tackling this problem but are ignored by most state-of-the-art methods. In this paper, we propose a novel Kronecker Product Matching module to match feature maps of different persons in an end-to-end trainable deep neural network. A novel feature soft warping scheme is designed for aligning the feature maps based on matching results, which is shown to be crucial for achieving superior accuracy. The multi-scale features based on hourglass-like networks and self-residual attention are also exploited to further boost the re-identification performance. The proposed approach outperforms state-of-the-art methods on the Market-1501, CUHK03, and DukeMTMC datasets, which demonstrates the effectiveness and generalization ability of our proposed approach.


Deep Group-shuffling Random Walk for Person Re-identification

Person re-identification aims at finding a person of interest in an imag...

Interpretable and Generalizable Deep Image Matching with Adaptive Convolutions

For image matching tasks, like face recognition and person re-identifica...

SphereReID: Deep Hypersphere Manifold Embedding for Person Re-Identification

Many current successful Person Re-Identification(ReID) methods train a m...

Adversarial Binary Coding for Efficient Person Re-identification

Person re-identification (ReID) aims at matching persons across differen...

Three-Stream Convolutional Networks for Video-based Person Re-Identification

This paper aims to develop a new architecture that can make full use of ...

Recognizing Partial Biometric Patterns

Biometric recognition on partial captured targets is challenging, where ...

Spectral Feature Transformation for Person Re-identification

With the surge of deep learning techniques, the field of person re-ident...

1 Introduction

Person re-identification (Re-ID) aims at finding a person of interest in an image gallery by comparing the query image of this person with all the other images in the gallery. It is an active research field in computer vision and has extensive applications in intelligent video surveillance, smart phone apps, and home robotics. Person Re-ID is closely related to yet harder than face verification, where human faces are usually well aligned, while person images show various poses and are captured from different viewing angles.

Recent years witnessed the success of deep learning methods for various computer vision problems. A large number of Convolutional Neural Network (CNN) based methods have been proposed for solving the problem of person Re-ID. Most state-of-the-art CNN based approaches aim at learning a highly non-linear mapping function that transforms person images into a common embedding space and ranking the gallery images according to their distances to the query image.

Figure 1: Illustration of our proposed framework with Kronecker Matching module and Soft Warping scheme for person re-identification. The KPM and soft warping are conducted between feature maps, which is end-to-end trainable in a deep network.

However, most existing CNN-based methods treat each person image as an individual sample without differentiating the semantic meanings of different pixels. State-of-the-art CNN networks either utilize global average pooling (e.g., ResNet [10]

) or direct vectorization (

e.g., VGG network [31]

) to convert the topmost feature maps into feature vectors. Directly comparing such feature vectors is ineffective because the same feature might encode different semantic concepts. The global average pooling simply averages features from all image spatial locations and abandons valuable spatial information. Features from different person regions might be compared. On the other hand, direct vectorization assumes that the person pose and camera viewpoints remain the same for all compared images so that features from the same image location could be compared, which is generally not true for person images from various surveillance cameras. In addition, existing CNN-based methods assume the same importance for all pixels in a person image. However, visual appearances of background regions are not informative and comparing them across different images leads to inaccurate similarity estimation.

To fully utilize spatial information of person images, we propose a deep neural network with Kronecker Product Matching (KPM) module to recover probabilistic correspondences between spatial regions across two images for more accurate person similarity estimation. Given the feature maps of two person images, the KPM module generates matching confidence maps to establish correspondences between them. Based on the matching confidence maps, a continuous warping scheme is adopted to deform the multi-scale feature maps of one image to match those of another image so that the feature maps of the two images could be better compared. A spatial attention mechanism is also adopted to automatically identify image regions of interest for re-identification. Extensive experiments and ablation studies on public person re-ID datasets show the effectiveness of our proposed method, which outperforms state-of-the-art approaches by large margins.

There were previous attempts on recovering person correspondences for person re-ID. Li et al[15]

proposed a patch matching layer that divides the pedestrian images into horizontal stripes and matches feature patches within each stripe between two images. However, the method assumes pedestrian images being vertically well aligned and cannot cope with larger deformations of persons’ spatial layouts. More importantly, this method directly feeds matching confidence maps into classifiers to determine the person similarities. Our experiments show that such confidence maps are not discriminative enough to obtain accurate similarity estimation. In contrast, our probabilistic feature warping module generates “softly” warped feature maps before estimating person similarities and results in significantly improved performance. To handle unaligned images, Zheng

et al[47] proposed a global matching algorithm, but the patches are divided on the image level and the features are hand-crafted. The person matching and person re-identification are performed in separate stages. In contrast, our proposed feature matching and warping modules are jointly optimized in a unified framework.

Our contribution can be summarized as threefold. First, we propose the Kronecker Product Matching module that is able to generate matching confidence maps between two images. Together with our proposed continuous warping scheme, the feature maps of person images could be stochastically deformed for end-to-end similarity learning via deep neural networks. Second, we exploit an hourglass-like network structure to generate multi-scale feature maps for person appearance encoding. The feature learning and warping are conducted at multiple scales for obtaining more robust person feature representations. Third, we investigate a series of important factors that could significantly impact the final performance, including loss functions, input aspect ratio, and specific network design, which provides guidance for the design of the proposed and future approaches.

Figure 2: Illustration of the proposed siamese-CNN network on a single scale with Kronecker Product Matching, soft feature warping, and self residual-attention modules for predicting whether the input image pair belong to the same person.

2 Related Work

Deep learning for person re-identification.  Person re-identification is a challenging problem which gains increasing attention in recent years [1, 6, 9, 36, 45, 29, 30]. State-of-the-arts person re-identification methods adopted deep learning techniques. Various loss functions, including pairwise verification loss, triplet loss, and classification loss, are utilized to train the deep learning models for re-ID. Ahmed et al. [1] designed a Cross-Input Neighbourhood Difference CNN model for person re-identification with a pair of cropped pedestrian images as input. Li et al. [15] proposed a patch matching layer that multiplies the activation of two images’ feature maps. Both of them employed a binary verification loss function for training. Xiao et al[40, 41]

trained CNN with a classification loss to learn the deep feature of person. Ding

et al[7] and Cheng et al[6] trained CNN with triplet samples and minimized feature distances between the same person while maximized the distances among different people. Most of these works only exploit the person identity information. Recently, pose information [43, 32] and pedestrian attribute information [21, 19] are also incorporated into person Re-ID systems, which enhance the performance of the trained models. A large number of metric learning methods for person re-identification have also been also proposed [24, 22, 13, 37, 42].

Spatial matching for person re-identification. There are some previous attempts on conducting spatial matching for person re-ID. Zheng et al[47] divided images into patches and utilized a global matching algorithm with hand-crafted features. Li et al[15] proposed a patch match layer for matching the features in horizontal stripes. Zheng et al[48] proposed a pedestrian alignment network for simultaneously aligning pedestrians within images and learning pedestrian descriptors. Bak et al[3] introduced a deformable patch-based model for accommodating pose changes and occlusions. Most of these approaches divide sample into patches at image level and cannot be trained in an end-to-end manner.

Multi-scale features for person re-identification Multi-scale features are beneficial for many computer vision tasks [23, 25, 51, 38, 39]. There are a few attempts of exploiting multi-scale features for person Re-ID. Liu et al[20] first introduced the multi-scale features into CNN training for person Re-ID, which directly downsamples image into different scales and fed them into shallow sub-networks. The outputs of these sub-networks are fused with features of the original scale for learning the similarity metric. Chen et al[5] proposed a network consisting of multiple branches for learning multi-scale features and one feature fusion branch. Different from [20], it incorporated person ID information for discriminative feature learning. These two works need multiple input images from multiple scales, which results in expensive computation. Qian et al[26] proposed a multi-scale stream layer, which was inspired by GoogLeNet [34] for learning features of different scales. But it needs multiple data streams for capturing multi-scale information, which results in redundant parameters. Our hourglass-like network could generate multi-scale features with fewer parameters, which is more efficiently.

3 Approach

Given a pair of pedestrian images, the similarity score between them is needed for determining whether these two images belong to the same person. Our proposed approach takes two pedestrian images as inputs and outputs their similarity score with a siamese convolutional neural network, which is shown in Figure 2. In this siamese-CNN, we propose Kronecker Product Matching module which will be introduced in Section 3.1 to generate matching confidence map and conduct soft warping between samples’ feature maps. Please be noted that Figure 2 only illustrates our proposed approach in one scale. In practice, we conduct our proposed operation under different scales and combine multi-scale features together for final prediction. The multi-scale architecture we utilize is an hourglass-like structure to be introduced in Section 3.2.

3.1 Kronecker Product Matching

The Kronecker Product Matching (KPM) module is designed to spatially match two sets of convolutional feature maps. Conventional approach would convert each set to a feature vector individually with global pooling or vectorization, then followed by a fully connected layer to compute their similarity. This ignores the spatial layout between the feature maps. KPM resolves this problem by establishing correspondence for each pixel across the two sets of feature maps. Based on the matching confidence map, probabilistic warping module deforms one of them to align it with the other, and computes their differences accordingly. We introduce KPM starting from the mathematical formulation of the problem.

3.1.1 Problem formulation

Given two input person images, the siamese-CNN outputs the two sets of convolutional feature maps, and , where is the number of pixels on each feature map, and are the -channel feature vector at each pixel. The goal of KPM and feature warping is to produce a vector that describes the difference between and , which can be further fed into a distance metric (e.g., distance where is a Malahanobis covariance matrix) or a neural network to compute the similarity.

For each feature vector , we would like to find the corresponding feature vector from that matches . Let

be an one-hot random variable out of

spatial locations, where if matches . It can be assumed to follow a multinomial distribution parameterized by with


The feature vector corresponding to could be stochastically selected as,


The difference vector at location can be computed by subtracting by the expected ,


On the other hand, different pixels could have different importance when comparing two sets of feature maps. For example, background regions in should make little contribution no matter if they have correspondences in or not. Thus we parameterize the importance of each by and compute the final difference vector between two aligned feature maps as

Figure 3: Illustration of Kronecker Product for computing dot-product similarities between features on the two feature maps.

3.1.2 Correspondence matching

KPM computes the correspondence probability

by calculating the inner product between every pair of and , and then normalizing across all the spatial locations in with the softmax function,



is the temperature hyperparameter. Lower temperature leads to lower entropy, which makes the distribution concentrate on a few high confidence locations.

CNN operations. Back to the CNN representations where feature maps have height and width , the feature vector at each location is indexed by row and column, e.g., and . The above inner products between 3-D feature maps of person images are actually a straightforward extension of Kronecker Product between two 2-D matrices, as demonstrated by blue rectangles in Figure 2 and Figure 3. The Kronecker Product results in an tensor with


Each submatrix can be viewed as the result of convolving an filter with the feature maps . The whole Kronecker Product can be efficiently computed by first permuting the dimension of to , and then using it as weights to convolve with .

Following Eq. (5), spatial softmax normalization is applied within each submatrix to obtain the matching confidence maps


After the softmax normalization we can compute the expectation in Eq. (3) by matrix multiplication between and the transposed (where the dimensions for are in front of those for ). This can be viewed as a continuous or soft warping that deforms the feature maps to pixelwisely match the feature maps . After the warping, the feature vectors across and are well aligned. Elementwise subtraction can be performed to obtain the difference vectors at all locations (as shown by the red cube in Figure 2), as


3.1.3 Residual self-attention for identifying discriminative regions

We propose to use a residual self-attention layer [35] (purple squares in Figure 2) to obtain the importance weights in Eq. (4) for . The feature maps are fed into two consecutive sets of

convolution, Batch Normalization (BN) 


, and ReLU layers. The output attention map

is then spatially normalized with the softmax function


with the temperature hyperparameter . We simply set , which means the image regions related to distinct regions should be strengthened by greater-than-1 weights.

At last, the above map of difference vectors in Eq. (8) is elementwisely weighted by and summed up to generate the final difference vector for classification,


An elementwise-square operation followed by a BN and a fully-connected layer is adopted to output the final confidence on whether the input pair belongs to a same person.

Figure 4: Illustration of the hourglass network for generating multi-scale feature maps with Kronecker Product Matching. (“GAP” denotes global average pooling and “ES” denotes elementwise square.)

3.1.4 Discussion with the existing approach

The patch matching layer in [15] considers spatial matching between convolutional feature maps for person Re-ID. However, they discard the original feature maps and utilize only the matching results to determine the similarity. We tried such a variant based on our framework, but the result shows that directly using matching confidence maps for similarity estimation is inferior to our proposed approach with not only feature matching but also feature warping operations. In addition, the correspondence matching in [15] is only conducted within each horizontal stripe, rather than the whole feature maps, which further restricts its discriminative ability on two vertically unaligned images.

3.2 Hourglass network for multi-scale KPM

Based on the KPM and residual self-attention modules, for fully exploiting the multi-scale information, we adopt a hourglass-like structure [23] to generate multi-scale feature maps. An overview of our hourglass network is illustrated in Figure 4. In our network, input images are resized to . The network generates feature maps from three different scales for feature matching, warping, and similarity estimation. The heights and widths of these feature maps are (scale-1), (scale-2) and (scale-3), respectively. Following the structure in [23], to gradually generate feature maps with larger sizes, we conduct convolution on lower-resolution feature maps to halve the channel dimension followed by bilinear upsampling to double the height and width dimension. The previous feature maps with the same spatial sizes are also processed by a convolution and elementwisely added to the output multi-feature maps. For feature maps of each scale, we conduct KPM, soft feature warping, elmentwise subtraction with residual self-attention, and the following operations in Section 3.1 to obtain the feature vector for classification. The feature vectors from the three scales are concatenated and fed in to a fully connected layer for estimating the final similarity score.

4 Experiments

To evaluate the overall performance and individual component of the proposed approach on person Re-ID, we conduct experiments on three popular public datasets.

4.1 Datasets and evaluation metric

CUHK03 [15] has 14,097 images of 1,467 identities captured by two cameras from the CUHK campus. It contains manually annotated images and automatically detected images. We use manually annotated ones in this paper.

Market-1501 [46] is a large-scale person Re-ID dataset which has 12,936 images for training and 19,732 images for testing. There are totally 1,501 identities in this dataset, captured from a real city market. All the images in this dataset are obtained by the DPM detector [8].

DukeMTMC [27] is also collected from a campus. It contains 1,812 identities from eight cameras. Among these 1,812 identities, 1,404 identities appear in more than two cameras. Here, we follow the setup in  [49] to divide DukeMTMC dataset into two parts: 16,522 images with 702 identities for training and 19,989 images with 702 identities for testing.

We adopt mean average precision (mAP) and top-1, top-5, top-10 cumulative matching characteristics (CMC) accuracies as evaluation metric. For different datasets, different CMC computation methods are adopted following each dataset’s own convention to calculate the final performance.

Factors Designs Market-1501 [46] CUHK03 [15] DukeMTMC [27]
mAP top-1 mAP top-1 mAP top-1
Loss functions Cross-entropy loss [41] 59.8 81.4 62.7 70.8 40.7 62.5
OIM loss [41] 60.9 82.1 72.5 77.5 47.7 68.1
Triplet loss [11] 67.9 85.1 80.7 84.3 54.6 73.1
Difference verification loss* 68.8 86.4 82.9 85.2 55.5 75.3
Input aspect ratio

1:1 (ImageNet pretrained)

66.6 86.1 79.8 85.4 54.9 57.1
1:1 (random initialization) 30.2 58.9 60.8 66.0 23.2 45.6
2:1 (random initialization) 32.4 61.3 61.5 67.6 24.9 46.6
2:1 (ImageNet pretrained)* 68.8 86.4 82.9 85.2 55.5 75.3
Feature difference
Absolute value 66. 85.0 81.7 84.1 54.6 73.8
Square before pooling 64.3 86.7 72.7 78.5 50.9 75.4
Square after pooling* 68.8 86.4 82.9 85.2 55.5 75.3
Table 1: Results by the base ResNet-50 network with different components on the Market-1501 [46], CUHK03 [15] and DukeMTMC [27] dataset. “*” denotes the component we use in our final model.

4.2 Implementation details

We implement our KPM hourglass network based on the ResNet-50 [10], which is pretrained on the ImageNet dataset. All the input images are resized to . The random horizontal flip and random erasing [50] are utilized for data augmentation. The positive-to-negative pair ratio for the each mini-batch is set to 1:3 for the Market-1501 dataset and 1:4 for the CUHK03 and DukeMTMC datasets. The temperature parameter in Eq. (7) is set to 0.05 and in Eq. (9

) is set to 1.0. We adopt SGD optimizer for model training and the initial learning rate is set to 0.01 which gradually drops to 0.001 after 50 epochs. It is then fixed for another 50 training epochs.

4.3 Base model design and analysis

For the Re-ID task, we observe that some basic network designs and training strategies have great impact on the final performance. In this section, we investigate loss functions, input image aspect ratio and difference feature normalization with a base ResNet-50 network. The analysis results could validate the design and hyperparameter setting of our proposed model in Figure 2.

Loss function. To investigate the effectiveness of the different loss functions on learning person features, we train a base ResNet-50 network with the cross-entropy (softmax) loss on predicting person identities [40], with the triplet loss on learning feature differences between samples [11], and with the proposed difference verification loss on verifying the similarities of input pairs. For training with our proposed difference verification loss, we utilize a siamese-structure, which is illustrated in Figure 5. A pair of images are fed into the siamese network. The feature difference between the two images are obtained by elementwise subtraction between the output feature maps of the last convolutional layer. The difference feature is then averaged across all locations and followed by a final fully-connected layer to compute the similarity confidence. Finally, the binary cross-entropy loss is adopted as the loss function.

For fair comparison, all the models are pretrained on ImageNet and we train the models with the cross-entropy loss [41], OIM loss [41] and triplet loss [11] following their original training scheme. The results by different loss functions are reported in Table 1. The meanAP and top-1 accuracy of these three loss functions are lower than that of the proposed difference verification loss function, which demonstrates the effectiveness of our difference verification loss function. In addition, the revised triplet loss in [11] needs a hard negative mining stage, which requires evaluating all training images with a partially trained model for hard negative sampling. We use the proposed siamese-ResNet-50 with difference verification loss in Figure 5 for the following experiments in this section.

Input aspect ratio. Conventional deep learning based methods adopt state-of-the-art deep networks (such as GoogLeNet and ResNet), for learning person features, which usually input images of squared shapes (e.g., ). However, since most person images have an height:width aspect ratio of roughly 2:1, we argue that resizing the images to square shapes would over-stretch the images in the horizontal dimension and hinder the feature learning capability. To validate our argument on the input image aspect ratio, we run experiments with different image aspect ratios corresponding to input image sizes of and . The proposed model is initialized with random parameters or ImageNet pretrained parameters. The results are shown in Table 1, which shows that the network trained with input aspect ratio 2:1 outperforms those trained with 1:1 aspect ratio. The 2:1 input aspect ratio results in gains of 2.2%, 0.3%, 1.7% on Market-1501, CUHK03, and DukeMTMC datasets with random initialization in terms of meanAP and achieve similar gains with ImageNet pretrained parameters.

Features difference normalization. For validating the importance of the elementwise square on the feature difference vector, we replace the elementwise square with absolute value operation in our base model in Figure 5, the meanAP drops by 1.9%, 1.2% , 0.9% and top-1 accuracies drop by 1.4%, 1.1%, 1.5% on three datasets. Then we also test removing the elementwise square from our base model and the model fails to converge. Furthermore, we also investigate the impact of where to conduct elementwise square. When we move the elementwise square before global average pooling, the meanAP drop by 3.7%, 7.8% , 4.6% on Market-1501, CUHK03, and DukeMTMC respectively. We also remove the subtraction and elementwise square operation, and simply concatenating the feature vectors of image pair for verification, but the training cannot converge.

Figure 5: Illustration of the base ResNet-50 network with the proposed pair difference classification
Methods Ref Market-1501 [46]
mAP top-1 top-5 top-10
DGD [40] CVPR’16 31.9 59.5   -   -
CADL [18] CVPR’17 47.1 73.8   -   -
P2S [52] CVPR’17 44.3 70.7   -   -
MSCAN [14] CVPR’17 53.1 76.3   -   -
SSM [2] CVPR’17 68.8 82.2   -   -
OIM Loss [41] CVPR’17 60.9 82.1   -   -
SpinNet [43] CVPR’17   - 76.9 91.5 94.6
JLML [16] IJCAI’17 65.5 85.1   -   -
VI+LSRO [49] ICCV’17 66.1 84.0   -   -
OL-MANS [51] ICCV’17   - 60.7   -   -
PDC [32] ICCV’17 63.4 84.1 92.7 94.9
PA [44] ICCV’17 63.4 81.0 92.0 94.7
SVDNet [33] ICCV’17 62.1 82.3 92.3 95.2
Ours 75.3 90.1 96.7 97.9
Table 2: mAP, top-1, top-5, and top-10 accuracies of compared methods on the Market-1501 dataset [46].
Methods Ref CUHK03 [15]
mAP top-1 top-5 top-10
DGD [40] CVPR’16   - 72.6 91.6 95.2
MSCAN [14] CVPR’17   - 74.2 94.3 97.5
SSM [2] CVPR’17   - 76.6 94.6 98.0
SpinNet [43] CVPR’17   - 88.5 97.8 98.6
OIM Loss [41] CVPR’17 72.5 77.5   -   -
Quadruplet [4] CVPR’17   - 75.5 95.2 99.2
JLML [16] IJCAI’17   - 83.2 98.0 99.4
OL-MANS [51] ICCV’17   - 61.7 88.4 95.2
PA [44] ICCV’17   - 85.4 97.6 99.4
SVDNet [33] ICCV’17 84.8 81.8 95.2 97.2
VI+LSRO [49] ICCV’17 87.4 84.6 97.6 98.9
PDC [32] ICCV’17   - 88.7 98.6 99.6
MuDeep [26] ICCV’17   - 76.3 96.0 98.4
Ours 89.2 91.1 98.3 99.1
Table 3: mAP, top-1, top-5 and top-10 accuracies by compared methods on the CUHK03 dataset [15]
Methods Ref DukeMTMC [27]
mAP top-1 top-5 top-10
BoW+KISSME [46] ICCV’15 12.2 25.1   -   -
LOMO+XQDA [17] CVPR’15 17.0 30.8   -   -
ACRN [28] CVPR’17 52.0 72.6 84.8 88.9
OIM Loss [41] CVPR’17 47.4 68.1   -   -
Basel+LSRO [49] ICCV’17 47.1 67.7   -   -
SVDNet [33] ICCV’17 56.8 76.7 86.4 89.9

63.2 80.3 89.5 91.9
Table 4: mAP, top-1, top-5, and top-10 accuracies by compared methods on the DukeMTMC dataset [27]

4.4 Comparison with existing person Re-ID methods

Results on Market-1501 dataset. Table 2 shows the experimental results of our approach and state-of-the-art methods on the Market-1501 dataset. Our approach results in the best performance on meanAP, top-1, top-5, and top-10 CMC accuracies, which demonstrates the effectiveness of the proposed method on this dataset. The SSM [2] method estimates the similarity between image pairs in the context of other pairs of instances. Compared with the SSM approach, our framework gains 6.5% and 7.9% in terms of meanAP and CMC top-1. Part Aligned (PA) [44] method jointly solves body detection and person Re-ID. Our method results in a meanAP of 75.3% and a CMC top-1 accuracy of 90.1%, which outperforms the PA by 63.4% and 81.0%. Similarly, SpindleNet (SpinNet) [43] and PDC [32] also integrate the human pose information in the person Re-ID pipeline which need human pose estimation pretrained model. Our approach does not need any pose estimation model yet still outperforms these methods.

Results on CUHK03 dataset. The Re-ID results on CUHK03 dataset is shown in Table 3. The meanAP and top-1 CMC accuracy of our framework are 89.2% and 91.1% which outperform those of state-of-the-art ones. For top-5 and top-10 CMC accuracies, PDC [32] yields slightly better performance than ours. However, PDC needs human pose estimation pretrained model, which is not utilized in our framework. MuDeep [26] learns discriminative features with different spatial scales and locations of person images. Our method improves the top-1 accuracy by 14.8% compared with MuDeep. Verif-Identif.+LSRO (VI+LSRO) [49] utilizes additional training data generated by GAN. Our method does not utilize any additional training data but still outperforms it.

Results on DukeMTMC dataset. In Table 4, we show the results of our framework and those by state-of-the-art ones. Our method outperforms all the compared frameworks. The gains on meanAP and CMC top-1 accuracy by our proposed framework are 6.4% and 3.6% compared with the state-of-the-art SVDNet [33]. ACRN [28] also integrates person attribute information into the training process. However, our method outperforms it by 6.3% and 7.5% on mAP and top-1 accuracy.

Figure 6: (Left) example attention maps by our approach of scale-1 and scale-2. Different distinct regions are focused to distinguish the pairs of person images. (Right) example matching confidence maps of feature vectors in red rectangles to those of another image.
Methods Market-1501 [46] CUHK03 [15] DukeMTMC [27]
mAP top-1 mAP top-1 mAP top-1
Baseline 68.8 86.4 82.9 85.2 55.5 75.3
Baseline+KPM 70.6 87.4 85.5 88.3 56.0 76.3
Baseline+KPM w/o warp 62.6 86.3 71.7 80.8 50.8 75.1
Baseline+KPM+RSA 72.0 88.7 86.2 89.3 58.1 77.0
Baseline+HG 69.9 87.7 84.1 87.0 60.4 79.0
Baseline+HG ML 69.1 86.8 - - - -
Baseline+KPM+RSA+HG 75.3 90.1 89.2 93.4 63.2 80.3
Table 5: Ablation studies on the Market-1501 [46], CUHK03 [15] and DukeMTMC [27] datasets. “+KPM”: with Kronecker Product Matching and soft warping modules. “+KPM w/o warp”: with only Kronecker Product Matching module. “+RSA”: with residual self-attention. “+HG”: with hourglass structure. “+HG ML” with hourglass structure and using multiple losses for different scales.

4.5 Ablation Study

Importance of different components. In this section, we investigate the effectiveness of each component in our proposed model by conducting a series of ablation studies on all three datasets. We treat the base ResNet-50 with pair difference classification in Section 4.3 as baseline model in this section. We first study the effectiveness of our Kronecker Product Matching and feature warping modules. We first train our baseline model with the KPM module, named Baseline+KPM. Its meanAP and CMC top-1 accuracy increase by 1.8%, 1.0% on the Market-1501 dataset, 2.6%, 3.1% on the CUHK03, and 0.5%, 1.0% on the DukeMTMC dataset compared with baseline model, which demonstrates that conducting soft warping with KPM helps align feature maps to enhance the performance. For validating the effectiveness of soft warping following KPM, we remove the soft warping operation and subtraction between warped feature maps which is illustrated in Eq. (8). Instead, we directly take the Kronecker Product Matching results from Eq. (7) and input them into a few of convolution and fully-connected layers to obtain the similarity confidence (denoted as Baseline+KPM w/o warp). Such an approach could be regarded as a extension of FPNN [15]. The mAP of this approach drops by 4.7%, 11.2% and 4.3% on the datasets and top-1 accuracies drop by 0.1%, 4.4% and 0.2%.

Then we investigate the importance of the proposed residual self-attention mechanism, which weights more on representative regions for verification. (denoted as Baseline+KPM+RSA). The meanAP has 1.4%, 0.7% and 2.1% increase on the Market-1501, CUHK03, and DukeMTMC datasets compared with Baseline+KPM model and the top-1 accuracies improve Baseline+KPM by 1.3%, 1.0% and 0.7% respectively. The improvements show the residual self-attention can help to implement the importance weights in Eq. (4). Next, we investigate the importance of the hourglass structure for obtaining multi-scale feature maps. Based on the baseline model, we integrate hourglass structure (denoted as Baseline+HG). Its mAP and top-1 accuracy increase by 1.1%, 1.3% on Market-1501, 1.2%, 1.8% on CUHK03, and 4.9%, 3.7% on DukeMTMC compared to the baseline model, which illustrate that integrating the multi-scale information into the framework is beneficial for distinguishing appearances of different people.

Furthermore, to investigate different hourglass structures’ impact. we also compare the results of two ways of combining information from multiple scales: 1) concatenating the feature vectors from different scales and to predict the similarity confidence (Baseline+HG); and 2) computing similarity confidences separately for each scale and averaging the confidences of multiple scales (denoted as Baseline+HG ML). Compared with Baseline+HG, the mAP and top-1 accuracy of Baseline+HG ML drop by 0.7% and 0.9% on Market-1501. Finally, the Baseline+KPM+RSA+HG denotes our proposed overall framework. Compared with our baseline model, its mAP and top-1 accuracy have 6.5% and 3.7% increase on Market-1501, 6.3% and 8.2% increase on CUHK03, 7.7% and 5.0% increase on DukeMTMC.

Temperature hyperparameters for KPM and RSA. In this section, we survey the temperature hyperparameters in Eq. (7) for Kronecker Product Matching (KPM) and in Eq. (9) for Residual Self-Attention (RSA). Decreasing and leads to lower entropy, thus the matching confidence map and the attention map would be more concentrated. In our approach, we prefer the matching maps to concentrate on a few high confidence locations while attention maps are dispersed to capture the whole foreground human region. Different temperatures’ impact on confidence maps and attention maps are shown in Figure 6. As illustrated in Table 6, such assumptions are validated by our experiments on the two temperature hyperparameters, where 0.05 is set for and is set to 1.0.

Temperature Evaluation Metric
for KPM for RSA mAP top-1
0.05 1.0 89.2 91.1
0.5 1.0 87.8 88.8
1.0 1.0 84.2 86.2
0.05 0.5 89.0 89.6
0.05 0.1 89.0 89.6
Table 6: mAP and top-1 accuracy by setting values for and for our proposed model on CUHK03 [15].

5 Conclusion

In this paper, we proposed an hourglass-like deep learning network Kronecker Product Matching, Soft Warping and Residual Self Attention for person re-identification. To match the correspondence feature between query image pair and further align the feature maps of query images, our proposed method incorporates Kronecker Product Matching and Soft Warping in the training process which is an end-to-end manner. Residual Self-Attention layer is also exploited for denoising the background noises. Furthermore, for capturing information from different scales we adopt hourglass structure in our model which could let model learn more robust and powerful feature. The proposed approach outperforms state-of-the-art methods on three large person re-identification datasets. Extensive component analysis of our framework demonstrates the effectiveness of our overall framework and individual components.


  • [1] E. Ahmed, M. Jones, and T. K. Marks. An improved deep learning architecture for person re-identification. In

    Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition

    , pages 3908–3916, 2015.
  • [2] S. Bai, X. Bai, and Q. Tian. Scalable person re-identification on supervised smoothed manifold. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), July 2017.
  • [3] S. Bak and P. Carr. Person re-identification using deformable patch metric learning. In Applications of Computer Vision (WACV), 2016 IEEE Winter Conference on, pages 1–9. IEEE, 2016.
  • [4] W. Chen, X. Chen, J. Zhang, and K. Huang. Beyond triplet loss: A deep quadruplet network for person re-identification. In CVPR, July 2017.
  • [5] Y. Chen, X. Zhu, and S. Gong. Person re-identification by deep learning multi-scale representations. In The IEEE International Conference on Computer Vision (ICCV), Oct 2017.
  • [6] D. Cheng, Y. Gong, S. Zhou, J. Wang, and N. Zheng. Person re-identification by multi-channel parts-based cnn with improved triplet loss function. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 1335–1344, 2016.
  • [7] S. Ding, L. Lin, G. Wang, and H. Chao. Deep feature learning with relative distance comparison for person re-identification. Pattern Recognition, 48(10):2993–3003, 2015.
  • [8] P. F. Felzenszwalb, R. B. Girshick, D. McAllester, and D. Ramanan. Object detection with discriminatively trained part-based models. IEEE transactions on pattern analysis and machine intelligence, 32(9):1627–1645, 2010.
  • [9] O. Hamdoun, F. Moutarde, B. Stanciulescu, and B. Steux. Person re-identification in multi-camera system by signature based on interest point descriptors collected on short video sequences. In Distributed Smart Cameras, 2008. ICDSC 2008. Second ACM/IEEE International Conference on, pages 1–6. IEEE, 2008.
  • [10] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 770–778, 2016.
  • [11] A. Hermans, L. Beyer, and B. Leibe. In defense of the triplet loss for person re-identification. arXiv preprint arXiv:1703.07737, 2017.
  • [12] S. Ioffe and C. Szegedy. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In

    International Conference on Machine Learning

    , pages 448–456, 2015.
  • [13] M. Koestinger, M. Hirzer, P. Wohlhart, P. M. Roth, and H. Bischof. Large scale metric learning from equivalence constraints. In Computer Vision and Pattern Recognition (CVPR), 2012 IEEE Conference on, pages 2288–2295. IEEE, 2012.
  • [14] D. Li, X. Chen, Z. Zhang, and K. Huang. Learning deep context-aware features over body and latent parts for person re-identification. In CVPR, July 2017.
  • [15] W. Li, R. Zhao, T. Xiao, and X. Wang. Deepreid: Deep filter pairing neural network for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 152–159, 2014.
  • [16] W. Li, X. Zhu, and S. Gong. Person re-identification by deep joint learning of multi-loss classification. In

    Proceedings of the 26th International Joint Conference on Artificial Intelligence

    , IJCAI’17, pages 2194–2200. AAAI Press, 2017.
  • [17] S. Liao, Y. Hu, X. Zhu, and S. Z. Li. Person re-identification by local maximal occurrence representation and metric learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 2197–2206, 2015.
  • [18] J. Lin, L. Ren, J. Lu, J. Feng, and J. Zhou. Consistent-aware deep learning for person re-identification in a camera network. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), July 2017.
  • [19] Y. Lin, L. Zheng, Z. Zheng, Y. Wu, and Y. Yang. Improving person re-identification by attribute and identity learning. arXiv preprint arXiv:1703.07220, 2017.
  • [20] J. Liu, Z.-J. Zha, Q. Tian, D. Liu, T. Yao, Q. Ling, and T. Mei. Multi-scale triplet cnn for person re-identification. In Proceedings of the 2016 ACM on Multimedia Conference, pages 192–196. ACM, 2016.
  • [21] Z. Liu, D. Wang, and H. Lu. Stepwise metric promotion for unsupervised video person re-identification. In The IEEE International Conference on Computer Vision (ICCV), Oct 2017.
  • [22] B. McFee and G. R. Lanckriet. Metric learning to rank. In Proceedings of the 27th International Conference on Machine Learning (ICML-10), pages 775–782, 2010.
  • [23] A. Newell, K. Yang, and J. Deng. Stacked hourglass networks for human pose estimation. In European Conference on Computer Vision, pages 483–499. Springer, 2016.
  • [24] S. Paisitkriangkrai, C. Shen, and A. van den Hengel. Learning to rank in person re-identification with metric ensembles. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 1846–1855, 2015.
  • [25] J. Pont-Tuset, P. Arbelaez, J. T. Barron, F. Marques, and J. Malik. Multiscale combinatorial grouping for image segmentation and object proposal generation. IEEE transactions on pattern analysis and machine intelligence, 39(1):128–140, 2017.
  • [26] X. Qian, Y. Fu, Y.-G. Jiang, T. Xiang, and X. Xue. Multi-scale deep learning architectures for person re-identification. arXiv preprint arXiv:1709.05165, 2017.
  • [27] E. Ristani, F. Solera, R. Zou, R. Cucchiara, and C. Tomasi. Performance measures and a data set for multi-target, multi-camera tracking. In European Conference on Computer Vision workshop on Benchmarking Multi-Target Tracking, 2016.
  • [28] A. Schumann and R. Stiefelhagen. Person re-identification by deep learning attribute-complementary information. In CVPRW, 2017 IEEE Conference on, pages 1435–1443. IEEE, 2017.
  • [29] Y. Shen, H. Li, T. Xiao, S. Yi, D. Chen, and X. Wang. Deep group-shuffling random walk for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 2265–2274, 2018.
  • [30] Y. Shen, T. Xiao, H. Li, S. Yi, and X. Wang. Learning deep neural networks for vehicle re-id with visual-spatio-temporal path proposals. In 2017 IEEE International Conference on Computer Vision (ICCV), pages 1918–1927. IEEE, 2017.
  • [31] K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition. CoRR, abs/1409.1556, 2014.
  • [32] C. Su, J. Li, S. Zhang, J. Xing, W. Gao, and Q. Tian. Pose-driven deep convolutional model for person re-identification. In The IEEE International Conference on Computer Vision (ICCV), Oct 2017.
  • [33] Y. Sun, L. Zheng, W. Deng, and S. Wang. Svdnet for pedestrian retrieval. In The IEEE International Conference on Computer Vision (ICCV), Oct 2017.
  • [34] C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and A. Rabinovich. Going deeper with convolutions. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 1–9, 2015.
  • [35] F. Wang, M. Jiang, C. Qian, S. Yang, C. Li, H. Zhang, X. Wang, and X. Tang. Residual attention network for image classification. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), July 2017.
  • [36] X. Wang, G. Doretto, T. Sebastian, J. Rittscher, and P. Tu. Shape and appearance context modeling. In Computer Vision, 2007. ICCV 2007. IEEE 11th International Conference on, pages 1–8. IEEE, 2007.
  • [37] K. Q. Weinberger and L. K. Saul. Distance metric learning for large margin nearest neighbor classification. Journal of Machine Learning Research, 10(Feb):207–244, 2009.
  • [38] F. Wu, S. Li, T. Zhao, and K. N. Ngan. Model-based face reconstruction using sift flow registration and spherical harmonics. In Pattern Recognition (ICPR), 2016 23rd International Conference on, pages 1774–1779. IEEE, 2016.
  • [39] F. Wu, S. Li, T. Zhao, and K. N. Ngan. 3d facial expression reconstruction using cascaded regression. arXiv preprint arXiv:1712.03491, 2017.
  • [40] T. Xiao, H. Li, W. Ouyang, and X. Wang. Learning deep feature representations with domain guided dropout for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 1249–1258, 2016.
  • [41] T. Xiao, S. Li, B. Wang, L. Lin, and X. Wang. Joint detection and identification feature learning for person search. In Proc. CVPR, 2017.
  • [42] D. Yi, Z. Lei, S. Liao, and S. Z. Li. Deep metric learning for person re-identification. In Pattern Recognition (ICPR), 2014 22nd International Conference on, pages 34–39. IEEE, 2014.
  • [43] H. Zhao, M. Tian, S. Sun, J. Shao, J. Yan, S. Yi, X. Wang, and X. Tang. Spindle net: Person re-identification with human body region guided feature decomposition and fusion. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 1077–1085, 2017.
  • [44] L. Zhao, X. Li, J. Wang, and Y. Zhuang. Deeply-learned part-aligned representations for person re-identification. arXiv preprint arXiv:1707.07256, 2017.
  • [45] R. Zhao, W. Oyang, and X. Wang. Person re-identification by saliency learning. IEEE transactions on pattern analysis and machine intelligence, 39(2):356–370, 2017.
  • [46] L. Zheng, L. Shen, L. Tian, S. Wang, J. Wang, and Q. Tian. Scalable person re-identification: A benchmark. In ICCV, pages 1116–1124, 2015.
  • [47] W.-S. Zheng, X. Li, T. Xiang, S. Liao, J. Lai, and S. Gong. Partial person re-identification. In The IEEE International Conference on Computer Vision (ICCV), December 2015.
  • [48] Z. Zheng, L. Zheng, and Y. Yang. Pedestrian alignment network for large-scale person re-identification. arXiv preprint arXiv:1707.00408, 2017.
  • [49] Z. Zheng, L. Zheng, and Y. Yang. Unlabeled samples generated by gan improve the person re-identification baseline in vitro. In Proceedings of the IEEE International Conference on Computer Vision, 2017.
  • [50] Z. Zhong, L. Zheng, G. Kang, S. Li, and Y. Yang. Random erasing data augmentation. arXiv preprint arXiv:1708.04896, 2017.
  • [51] J. Zhou, P. Yu, W. Tang, and Y. Wu. Efficient online local metric adaptation via negative samples for person re-identification. In The IEEE International Conference on Computer Vision (ICCV), Oct 2017.
  • [52] S. Zhou, J. Wang, J. Wang, Y. Gong, and N. Zheng. Point to set similarity based deep feature learning for person re-identification. In CVPR, July 2017.