Learning Diverse Features with Part-Level Resolution for Person Re-Identification

01/21/2020 ∙ by Ben Xie, et al. ∙ IEEE 0

Learning diverse features is key to the success of person re-identification. Various part-based methods have been extensively proposed for learning local representations, which, however, are still inferior to the best-performing methods for person re-identification. This paper proposes to construct a strong lightweight network architecture, termed PLR-OSNet, based on the idea of Part-Level feature Resolution over the Omni-Scale Network (OSNet) for achieving feature diversity. The proposed PLR-OSNet has two branches, one branch for global feature representation and the other branch for local feature representation. The local branch employs a uniform partition strategy for part-level feature resolution but produces only a single identity-prediction loss, which is in sharp contrast to the existing part-based methods. Empirical evidence demonstrates that the proposed PLR-OSNet achieves state-of-the-art performance on popular person Re-ID datasets, including Market1501, DukeMTMC-reID and CUHK03, despite its small model size.



There are no comments yet.


page 6

page 7

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

I Introduction

In recent years, person re-identification (Re-ID) has attracted increasing interest due to its fundamental role in emerging computer vision applications such as video surveillance, human identity validation, and authentication, and human-robot interaction

[1, 2, 3, 4, 5, 6]. The objective of person Re-ID is to match any query image with the images of the same person taken by the same or different cameras at different angles, time or location. Despite its recent progress, identifying the person of interest accurately and reliably is still very challenging due to huge variations in lighting, human pose, background, camera viewpoint, etc. With the ID-labeled training set, one of the main goals in the field of person Re-ID is to discover a low-dimensional but rich representation of any input image for person matching.

Fig. 1: The performance of different baselines on the DukeMTMC-reID and CUHK03-Detected datasets. We compare the proposed method with other baselines published in CVPR, ECCV and ICCV (2018/2019).

Person Re-ID was often formulated as a metric-learning problem (or a feature-embedding problem)  [7, 8, 9], where the distance between intra-class samples is required to be less than the distance between inter-class ones by at least a margin. Unfortunately, a direct implementation of this idea requires to group samples in a pairwise manner, which is known to be computationally intensive. Alternatively, a classification task is employed to find the feature-embedding solution due to its advantage on the implementation complexity. Currently, various state-of-the-art methods [10, 2, 6, 11, 12] for person Re-ID have evolved from a single metric-learning problem or a single discriminative classification problem to a multi-task problem, where both the discriminative loss and the triplet loss are employed [13]

. As each sample image is only labeled with the person ID, an end-to-end training approach usually has difficulty to learn diverse and rich features without elaborate design of the underlying neural network and further use of some regularization techniques.

In the past years, various part-based approaches [14, 15, 16] and dropout-based approaches [17] have been proposed in order to learn rich features from the ID-labeled dataset. Differing from conventional pose-based Re-ID approaches [7, 18, 19, 20], part-based approaches usually locate a number of body parts firstly, and force each part meeting an individual ID-prediction loss in getting discriminative part-level feature representations [21, 22, 23, 24]. Dropout-based approaches, however, intend to discover rich features from enlarging the dataset with various dropout-based data-augmentation methods, such as cutout [25] and random erasing [26]

, or from dropping the intermediate features from feature-extracting networks, such as Batch BropBlock


The performance of part-based methods relies heavily on the employed partition mechanism. Semantic partitions may offer stable cues to good alignment but are prone to noisy pose detections, as it requires that human body parts should be accurately identified and located. The uniform horizontal partition was widely employed in [15, 22], which, however, provides limited performance improvement.

This motivates the work in this paper, where we propose a novel two-branch lightweight architecture for discovering rich features in person Re-ID. In particular, we employ the idea of part-level feature resolution in developing a strong two-branch baseline for person Re-ID. Compared to the popular part-based method of PCB [15], our method differs mainly in two aspects. One is the use of global branch for facilitating the extraction of a global feature, and the other is the use of a single ID-prediction loss for part-level feature resolution. We briefly summarize the main contribution of this paper as follows:

  1. Based on the omni-scale network (OSNet) baseline [27]

    , we propose a lightweight two-branch network architecture (PLR-OSNet) for person Re-ID. Its global branch adopts a global-max-pooling layer while its local branch employs a part-level feature resolution scheme for producing only a single ID-prediction loss, which is in sharp contrast to existing part-based methods. The proposed architecture is shown to be effective for achieving feature diversity.

  2. Despite its small model size, the proposed PLR-OSNet is very efficient as depicted in Figure 1 for achieving the state-of-the-art results [28, 29, 6, 3, 2, 10, 30, 15, 31] on the three popular person Re-ID datasets, Marktet1501, DukeMTMC-reID and CUHK03. It achieves the rank-1 accuracy of 91.6% for DukeMTMC-reID and 83.5% for CUHK03-Labeled without using re-ranking. 111Source codes are available at https://github.com/AI-NERC-NUPT/PLR-OSNet

Ii Related Work

We review the relevant work about embedded feature learning, discriminative feature learning, part-based feature learning, and multi-scale feature learning. Besides of various problems for person Re-ID, we are particularly interested in achieving feature diversity throughout this section.

Ii-a Embedded Feature Learning

Person Re-ID can be formulated as a feature-embedding problem, which looks for a function mapping the high-dimensional pedestrian images into a low-dimensional feature space. This feature-embedding formulation requires the mapping function to ensure that given any anchor image, any positive image from the same person should has a lower distance to the anchor in the feature space compared to any negative image from a different person. This is known to be the objective of the triplet loss in training.

For an efficient feature-embedding learning, the batch hard triplet loss [32]

was proposed to mine the hardest positive and the hardest negative samples for each pedestrian image in a batch. However, it is sensitive to outlier samples and may discard useful information due to its hard selective approach. To deal with these problems, Ristani et al. proposed the batch-soft triplet loss

[32], which introduce a weighting factor for each pair distance. One hyper-parameter that exists in all of the triplet loss variations is the margin. To eliminate the manual parameter of the margin, the softplus function instead of

in the triplet loss function was introduced in

[12], which is known to be soft-margin triplet loss.

Ii-B Discriminative Feature Learning

A feature-embedding approach requires to group the pedestrian images into pairs for training, which is not efficient in general. Discriminative feature learning is more efficient by training a classification task, where each person is regarded as a single class. As each pedestrian image in a training dataset is only labeled by a single person ID, a fundamental problem in the field of person Re-ID is how to learn diverse features from the ID-labeled dataset.

To get diverse features from an end-to-end training approach, multi-branch network architectures have been widely employed [15, 2, 33], where a shared-net is often followed by multiple subnetwork branches. To achieve feature diversity, distinct mechanisms should be imposed among different branches, such as attention [6, 3], feature dropping [2, 34], and overlapped activation penalty [4] .

Fig. 2:

Single ID loss vs. multiple ID loss with part-level feature resolution. The input image goes through backbone network to obtain a 3-D tensor, which is vertically split into

= 4 part-level features, and then averages each part-level tensor into a vector. These

part-level vectors are used to drive independent ID losses in PCB, or simply concatenated for driving a single ID loss in training. The use of multiple ID loss in training may lead to false prediction of the person ID with some part-level features.

Ii-C Part-Based Feature Learning

Part-based feature learning with hand-crafted algorithms had been pursued for a long time for the purpose of person retrieval before the era of deep-learning. In [15], the Part-based Convolutional Baseline (PCB) network was proposed, which employs uniform partition on the conv-layer for learning part-level features. Essentially, it employed a 6-branch network by dividing the whole body into 6 horizontal stripes in the feature space and each part feature vector was used to produce an independent ID-prediction loss. The idea of PCB was very welcome and widely adopted for developing stronger methods in the recent years for person Re-ID [31][30][21].

The part-level feature learning has an intuitive advantage for extracting diverse features from the ID-labeled pedestrian images. However, the pristine division strategy usually suffers from misalignment [35] between corresponding parts due to large variations in poses, viewpoints and scales. In particular, the use of multiple ID-prediction loss (an independent ID loss for each part) may fail to capture the semantic part-level features since a pedestrian image may simply contain the semantically different parts at a uniformly-divided pedestrian part. This may partially explain the limited performance advantage of various PCB-based algorithms, compared to the state-of-the-art methods [34, 3].

Ii-D Multi-Scale Feature Learning

Recently, multi-scale feature learning, together with the multi-stream building block design [36, 37, 38]

, has been proved to be efficient for improving the performance of person Re-ID. By designing a residual block composed of multiple convolutional feature streams, each detecting features at a certain scale, the concept of omni-scale deep feature learning was further introduced in

[27] and a lightweight CNN architecture, termed OSNet, was cleverly constructed for learning omni-scale feature representations. Experiments demonstrated that OSNet performs very well for both tasks of classification and person Re-ID, despite its lightweight design.


Iii-a Part-Level Feature Resolution

Fig. 3: The overall network architecture of PLR-OSNet. During testing, the feature embedding concatenated from both global branch and local branch is used for the final matching distance computation.

For PCB with -part resolution, it produces part-level feature vectors by dividing the whole body into horizontal stripes in the feature space. As shown in Figure 2, the input image goes forward through the stacked convolutional layers from the backbone network to form a 3-D tensor . PCB employs a conventional average pooling layer to spatially down-sample into pieces of column vectors , followed by classifiers in order to produce ID-predication loss. Note that the classifier is implemented by a fully-connected (FC) layer and a softmax function. Hence, when the batch of input labeled samples are , PCB employs the multiple ID-prediction loss as


where , are the -th and -th column of the weight matrix (the -th classifier designated for ), respectively.

By forcing each part-level feature vector to meet an independent ID-prediction loss, one may obtain useful part-level features for discriminating different persons. However, many part-level feature vectors may simply fail to catch any discriminative information for different persons, as shown in Figure 2. Therefore, the use of PCB is practically limited for getting discriminative part-level information.

In order to learn discriminative features with part-level resolution, we propose to concatenate part-level feature vectors into a single column vector


which is further used to produce the ID-prediction loss


Here, , are the -th and -th column of the weight matrix (the single classifier for ), respectively. As the vector contains the full information about the input image, the use of a single ID-prediction loss could drive to learn sufficient discriminative information.

The proposed approach is somewhat similar to OSNet, where the tensor is followed by a global average pooling (GAP) for getting a global descriptor


Instead of using the GAP in OSNet, the proposed part-level resolution approach uses average pooling in each part to retrieve part-level feature vectors and the final descriptor (3) is of rich local information, which might be simply filtered with the GAP (5) in OSNet.

Iii-B Proposed Network Architecture

We employ an -branch neural network architecture, by modifying the recently-proposed OSNet baseline. Figure 3 shows the overall network architecture, which includes a backbone network, a global branch (orange colored arrows), and a local branch (blue colored arrows).

Iii-B1 Attention Modules

Compared to the OSNet, attention modules are explicitly employed in Figure 3, where both spatial attention module (SAM) and channel attention module (CAM) are used in the shared-net.

For SAM, we employ the version of [6], which was designed to capture and aggregate those semantically related pixels in the spatial domain. To further reduce the computational complexity, we use a convolution that forms a functions (or ) to reduce the number of channels to of the input .

For CAM, the squeeze-and-excitation mechanism [39] is employed with slight modifications detailed in Figure 3. Compared to the channel attention module in [6]

, it does not require to compute the channel affinity matrix and therefore can be implemented more efficiently.

Iii-B2 Shared-Net

The recently-proposed OSNet is employed as the backbone network for feature extraction. OSNet uses a lightweight network architecture for omni-scale feature learning, which is achieved by employing the factorised convolutional layer, the omni-scale residual block and the unified aggregation gate. The shared-net consists of the first 3 conv layers and 2 transition layers from OSNet. As shown in Figure 3, we insert SAM + CAM modules in both conv-2 and conv-3 layers for the shared-net.

Iii-B3 Global Branch with Global-Max-Pooling

The global branch consists of the conv4 and conv5 layers, a Global-Max-Pooling (GMP) layer to produce a 512-dimensional vector, providing a compact global feature representation for both the triplet loss and the ID-prediction loss. The use of GMP is mainly for achieving the feature diversity between the global branch and the local branch, where average pooling is known to be popular in PCB [15] and adopted in the local branch.

Iii-B4 Local Branch with Part-Level Feature Resolution

The local branch has the similar layer structure but with the average pooling (AP) in replace of GMP. To achieve feature diversify, a uniform partition strategy is employed for part-level feature resolution, and four 512-dimensional features are then concatenated for producing just one ID-prediction loss. The use of a single ID-prediction loss is unique in this paper, while PCB and its variations employed a multiple ID-prediction loss with an independent ID-prediction loss for each part.

Iii-C Loss Functions

The feature vectors from the global and local branches are concatenated as the final descriptor for the person Re-ID task. The loss function at either the global branch or the local branch is the sum of a single ID loss (softmax loss), a soft margin triplet loss [12] and a center loss [40], namely,


where are weighting factors.

Iv Experiments

Extensive experiments have been performed for evaluating the effectiveness of the proposed approach over three public person Re-ID datasets: Market1501, DukeMTMC-reID and CUHK03. The results are compared to the state-of-the-art methods.

Iv-a Datasets

The Market1501 dataset [41] has 1,501 identities collected by six cameras and a total of 32,668 pedestrian images. Following [41]. The dataset is split into a training set with 12,936 images of 751 identities and a testing set of 3,368 query images and 15,913 gallery images of 750 identities.

The DukeMTMC-reID dataset [42] contains 1,404 identities captured by more than 2 cameras and a total of 36,411 images. The training subset contains 702 identities with 16,522 images and the testing subset has other 702 identities.

The CUHK03 dataset [43] contains labeled 14,096 images and detected 14,097 images of a total of 1,467 identities captured by two camera views. With splitting just like in [41], a non-overlapping 767 identities are for training and 700 identities for testing. The labeled dataset contains 7,368 training images, 5,328 gallery, and 1,400 query images for testing, while the detected dataset contains 7,365 images for training, 5,332 gallery, and 1,400 query images for testing.

Iv-B Implementation Details

Our network is trained using a single Nvidia Tesla P100 GPU with a batch size of 64. Each identity contains 4 instance images in a batch, so there are 16 identities per batch. The backbone OSNet is initialized from the ImageNet pre-trained model. The total number of epoches is set to 120 [150], namely, 120 for both Market-1501 and DukeMTMC-reID, and 150 for CUHK03, respectively. We use the Adam optimizer with the base learning rate initialized to 3.5e-5. With a linear warm-up strategy in first 20 [40] epochs, the learning rate increases to 3.5e-4. Then, the learning rate is decayed to 3.5e-5 after 60 [100] epochs, and further decayed to 3.5e-6 after 90 [130] epochs.

For training, the input images are re-sized to and then augmented by random horizontal flip, random erasing, and normalization. The testing images are re-sized to with normalization.

Iv-C Comparison with State-of-the-art Methods

We compare our work with state-of-the-art methods, in particular emphasis on the recent remarkable works (CVPR’19 and ICCV’19) on person Re-ID, over the popular benchmark datasets Market-1501, DukeMTMC-ReID and CUHK03. All reported results are obtained without any re-ranking [44, 45] or multi-query fusion [41] techniques. The comparison results are listed in Table 1, Table 2 and Table 3. From these tables, one can observe that our proposed method performs competitively among various state-of-the-art methods, including PCB [15], IAN [5], CAMA [4],MHN [3],Pyramid [31], BagOfTricks [10], ABD-Net [6], BDB [2], SONA [34], Auto-ReID [30], OSNet [27], et al.

As shown, our PLR-OSNet has achieved the best mAP performance among various state-of-the-art methods for all the three datasets. For DukeMTMC-reID, PLR-OSNet obtained 91.6% Rank-1 accuracy and 81.2% mAP, which significantly outperforms all existing methods. For CUHK03, PLR-OSNet even outperforms SONA in both mAP and Rank-1 accuracy, which might be the best performing algorithm for CUHK03.

Besides of its strong competition in both Rank-1 and mAP performance, PLR-OSNet has a lightweight network architecture inherited from OSNet. It only has only 3.4M parameters while the recently-available Robust-ReID has 6.4M parameters.

Method mAP rank-1
KPM [46](CVPR’18) 75.3 90.1
MLFN [36](CVPR’18) 74.4 90.0
CRF [47](CVPR’18) 81.6 93.5
HA-CNN [28](CVPR’18) 75.7 91.2
PCB [15](ECCV’18) 81.6 93.8
Mancs [29] (ECCV’18) 82.3 93.1
SNL [48](ACM’18) 73.43 88.27
HDLF[49](ACM MM’18) 79.10 93.30
MGN [21](ACM MM’18) 86.9 95.7
Local CNN[50](ACM MM’18) 87.4 95.9
IAN [5] (CVPR’19) 83.1 94.4
CAMA [4](CVPR’19) 84.5 94.7
MHN [3](CVPR’19) 85.0 95.1
Pyramid [31](CVPR’19) 88.2 95.7
BagOfTricks [10] (CVPRW’19) 85.9 94.5
ABD [6] (ICCV’19) 88.28 95.6
BDB [2] (ICCV’19) 86.7 95.3
SONA [34] (ICCV’19) 88.67 95.68
Auto-ReID [30] (ICCV”19) 85.1 94.5
OSNet [27] (ICCV’19) 84.9 94.8
PLR-OSNet 88.9 95.6
TABLE I: Comparison of our proposed method with state-of-the-art methods for the Market-1501 dataset.
Method mAP rank-1
MLFN [36](CVPR’18) 62.8 81.2
GP-Re-ID [51] (CVPR’18) 72.8 85.2
HA-CNN [28](CVPR’18) 63.8 80.5
PCB [15] (ECCV’18) 69.2 83.3
Mancs (ECCV’18) 71.8 84.9
MGN [21](ACM MM’18) 78.40 88.7
Local CNN[50](ACM MM’18) 66.04 82.23
IAN [5] (CVPR’19) 73.4 87.1
CAMA [4] (CVPR’19) 72.9 85.8
MHN [3] (CVPR’19) 77.2 89.1
Pyramid [31](CVPR’19) 79.0 89.0
BagOfTricks (CVPRW’19) 76.4 86.4
ABD [6] (ICCV’19) 78.59 89.0
BDB [2] (ICCV’19) 76.0 89.0
SONA [34] (ICCV’19) 78.05 89.25
Auto-ReID [30] (ICCV’19) 75.1 88.5
OSNet[27] (ICCV’19) 73.5 88.6
PLR-OSNet 81.2 91.6
TABLE II: Comparison of our proposed method with state-of-the-art methods for the DukeMTMC-reID dataset.
Method Labeled Detected
mAP rank-1 mAP rank-1
DaRe+RE [52](CVPR’18) 61.6 66.1 59.0 63.3
MLFN [36](CVPR’18) 49.2 54.7 47.8 52.8
HA-CNN [28](CVPR’18) 41.0 44.4 38.6 41.7
PCB [15](ECCV’18) - - 57.5 63.7
Mancs (ECCV’18) 63.9 69.0 60.5 65.5
MGN [21](ACM MM’18) 67.4 68.0 66.0 68.0
MHN [3] (CVPR’19) 72.4 77.2 65.4 71.7
Pyramid [31](CVPR’19) 76.9 78.9 74.8 78.9
BDB [2] (ICCV’19) 76.7 79.4 73.5 76.4
SONA [34] (ICCV’19) 79.23 81.85 76.35 79.10
Auto-ReID [30] (ICCV’19) 73.0 77.9 69.3 73.3
OSNet [27] (ICCV’19) 67.8 72.3
PLR-OSNet 80.5 84.6 77.2 80.4
TABLE III: Comparison of our proposed method with state-of-the-art methods for the CUHK03 dataset.

Iv-D Visualization

Fig. 4: Visualization of class activation maps (CAMs) for the global branch and the local branch (including 4 part-level feature vectors). The proposed architecture allow the model to learn diverse features (marked in orange).
Fig. 5: Three Re-ID examples of PLR-OSNet and OSNet on DukeMTMC-reID. Left: query image. Upper-Right: top-10 results of PLR-OSNet. Low-Right: top-10 results of OSNet. Images in red boxes are negative results. PLR-OSNet boosts the retrieval performance

Visualization of Feature Diversity Between Two Branches: In Figure 4, we show the visualization of class activation maps (CAMs) for the global feature vector and 4 local part-level feature vectors. Note that the local branch produces 4 part-level feature vectors, corresponding to part-1, part-2, part-3 and part-4. As shown, these part-level features have some degree of diversity compared to the global features. This means that the proposed PLR-OSNet architecture allows the model to learn diverse features, which is key to the high performance of person Re-ID.

Re-ID Visual Retrieving Results: We compare PLR-OSNet with OSNet more directly from visual retrieving results. Three retrieved examples are shown in Figure 5. One can see that OSNet fails to retrieve several correct images among the top-10 results. Taking the second query as an example, PLR-OSNet is able to find correct images of the same identity in the top 10 results whilst OSNet gets 5 incorrect ones.

Iv-E Ablation Studies

Iv-E1 Benefit of Global Features

Global Features Market1501 DukeMTMC CUHK03-Labeled CUHK03-Detected
mAP rank-1 mAP rank-1 mAP rank-1 mAP rank-1
No 86.9 94.6 79.8 90.2 77.5 81.2 73.4 77.6
Yes 88.9 95.6 81.2 91.6 80.5 84.6 77.2 80.4
TABLE IV: The use of global features on the final performance
Method Market1501 DukeMTMC CUHK03-Labeled CUHK03-Detected
mAP rank-1 mAP rank-1 mAP rank-1 mAP rank-1
Multiple ID loss 85.6 94.4 77.0 89.4 79.4 83.1 74.7 78.4
Single ID loss 88.9 95.6 81.2 91.6 80.5 84.6 77.2 80.4
TABLE V: Single ID Loss vs. Multiple ID Loss
Attention Modules Market1501 DukeMTMC CUHK03-Labeled CUHK03-Detected
mAP rank-1 mAP rank-1 mAP rank-1 mAP rank-1
No 88.4 95.0 81.0 90.8 79.5 82.4 76.8 79.4
Yes 88.9 95.6 81.2 91.6 80.5 84.6 77.2 80.4
TABLE VI: The use of attention modules on the final performance

PCB employed a uniform partition strategy for producing part-level features, which did not consider any possibility of the use of global features. The proposed PLR-OSNet, however, introduces a global branch, which uses global-max-pooling for extracting global features as shown in Figure 3. With the use of global features, PLR-OSNet performs significantly better as depicted in Table IV for all the three datasets. For CUHK03-Label, PLR-OSNet achieves the Rank-1 accuracy of 84.6% with the global features, without which it simply can obtain 81.2% Rank-1 accuracy. This suggests that the global branch and the local branch reinforce each other, both contributing to the final performance.

Iv-E2 Single ID Loss vs. Multiple ID Loss

PLR-OSNet uses only single ID loss for multiple part-level features, which is sharply contrast to PCB and its variants, where each part-level feature vector is employed to drive an ID loss so that the number of ID loss is equal to the number of separated parts. The use of ID loss for each part-level feature can force it to learn the feature at each specified part with the ID-labeled dataset. The drawback, however, is that some part-level features may fail to produce any reliable ID prediction. By concatenating multiple part-level feature vector into a single feature vector, a single ID prediction is much more reliable.

With the use of multiple feature concatenation followed by a single ID loss, PLR-OSNet performs significantly better as shown in Table V for all the three datasets. For Market1501, PLR-OSNet obtains 88.9% mAP, which surpasses its counterpart (with multiple ID loss) about 3.3%.

Iv-E3 Benefit of Attention Modules

The attention modules have been widely employed in various state-of-the-art methods for person Re-ID. Therefore, we also insert these popular attention modules in the shared net as shown in Figure 3. Experiments results are shown in in Table VI for all the three datasets. Clearly, it achieves consistently improved performance for all three datasets. However, the improvement is moderate, which may be due to the existence of the inherent attention mechanisms in OSNet.

Iv-E4 Soft Margin Triplet Loss vs. Hard Margin Triplet Loss

Triplet Loss CUHK03-Labeled CUHK03-Detected
mAP rank-1 mAP rank-1
Hard Margin 75.9 78.6 72.8 76.0
Soft Margin 80.5 84.6 77.2 80.4
TABLE VII: Soft Margin Triplet Loss vs. Hard Margin Triplet Loss.

Table VII studies the impact of soft margin triplet loss on the performance of the PLR-OSNet over CUHK03. Surprisingly, there is a large performance gap between soft margin triplet loss and hard margin triplet loss. We can see that the Rank-1 accuracy gap is around 6% while the mAP gap is about 4.6%. We also do experiments over Market1501 and DukeMTMC. However, experiments show that the use of soft margin triplet loss does not produce any observable improvement over the hard margin counterpart. Therefore, it remains unknown why the soft margin triplet loss can produce significantly better results compared to the hard version for CUHK03.

V Conclusion

In this paper, we propose a new OSNet structure with part-level feature resolution for person Re-ID. With a two-branch network architecture, the proposed PLR-OSNet concatenates various uniformly-partitioned part-level feature vectors to a long vector for producing a single ID prediction loss, which is proved to be more efficient than the existing part-based methods. Extensive experiments show that PLR-OSNet achieves state-of-the-art performance on popular person Re-ID datasets, including Market1501, DukeMTMC-reID and CUHK03. In the mean time, its model size is significantly smaller than various state-of-the-art methods, thanking to the lightweight architecture of OSNet.


  • [1] Zheng, Liang and Yang, Yi and Hauptmann, Alexander G. (2016), “Person re-identification: Past, present and future,” [Online]. Available: https://arxiv.org/abs/1610.02984.
  • [2] Z. Dai, M. Chen, X. Gu, S. Zhu, and P. Tan, “Batch dropblock network for person re-identification and beyond,” in Proc. ICCV, 2019, pp. 3691–3701.
  • [3] B. Chen, W. Deng, and J. Hu, “Mixed high-order attention network for person re-identification,” in Proc. ICCV, 2019, pp. 371–381.
  • [4] W. Yang, H. Huang, Z. Zhang, X. Chen, K. Huang, and S. Zhang, “Towards rich feature discovery with class activation maps augmentation for person re-identification,” in Proc. CVPR, June 2019, pp. 1389–1398.
  • [5] R. Hou, B. Ma, H. Chang, X. Gu, S. Shan, and X. Chen, “Interaction-and-aggregation network for person re-identification,” in Proc. CVPR, 2019, pp. 9317–9326.
  • [6] T. Chen, S. Ding, J. Xie, Y. Yuan, W. Chen, Y. Yang, Z. Ren, and Z. Wang, “Abd-net: Attentive but diverse person re-identification,” in Proc. ICCV, 2019, pp. 8351–8361.
  • [7] C. Su, J. Li, S. Zhang, J. Xing, W. Gao, and Q. Tian, “Pose-driven deep convolutional model for person re-identification,” in Proc. ICCV, 2017, pp. 3960–3969.
  • [8] W. Chen, X. Chen, J. Zhang, and K. Huang, “Beyond triplet loss: A deep quadruplet network for person re-identification,” in Proc. CVPR, July 2017, pp. 1320–1329.
  • [9] S. Bai, X. Bai, and Q. Tian, “Scalable person re-identification on supervised smoothed manifold,” in Proc. CVPR, July 2017, pp. 3356–3365.
  • [10]

    H. He, Z. Zhang, H. Zhang, Z. Zhang, J. Xie, and M. Li, “Bag of tricks for image classification with convolutional neural networks,” in

    Proc. CVPR, 2019, pp. 558–567.
  • [11] Z. Zheng, L. Zheng, and Y. Yang, “A discriminatively learned cnn embedding for person reidentification,” ACM Trans. Multimedia Comput., Commun., Appl., vol. 14, no. 1, p. 13, 2018.
  • [12] Hermans, Alexander and Beyer, Lucas and Leibe, Bastian. (2017), “In defense of the triplet loss for person re-identification,” [Online]. Available: https://arxiv.org/abs/1703.07737.
  • [13] C. Su, S. Zhang, J. Xing, W. Gao, and Q. Tian, “Deep attributes driven multi-camera person re-identification,” in Proc. ECCV.   Springer, 2016, pp. 475–491.
  • [14] H. Yao, S. Zhang, R. Hong, Y. Zhang, C. Xu, and Q. Tian, “Deep representation learning with part loss for person re-identification,” IEEE Trans. Image Process., vol. 28, no. 6, pp. 2860–2871, June 2019.
  • [15] Y. Sun, L. Zheng, Y. Yang, Q. Tian, and S. Wang, “Beyond part models: Person retrieval with refined part pooling (and a strong convolutional baseline),” in Proc. ECCV, 2018, pp. 480–496.
  • [16] L. Zhao, X. Li, Y. Zhuang, and J. Wang, “Deeply-learned part-aligned representations for person re-identification,” in Proc. ICCV, Oct 2017, pp. 3239–3248.
  • [17] Dai, Zuozhuo and Chen, Mingqiang and Zhu, Siyu and Tan, Ping. (2018), “Batch feature erasing for person re-identification and beyond,” [Online]. Available: https://arxiv.org/abs/1811.07130.
  • [18] V. Kumar, A. Namboodiri, M. Paluri, and C. V. Jawahar, “Pose-aware person recognition,” in Proc. CVPR, July 2017, pp. 6797–6806.
  • [19] L. Zheng, Y. Huang, H. Lu, and Y. Yang, “Pose-invariant embedding for deep person re-identification,” IEEE Trans. Image Process, vol. 28, no. 9, pp. 4500–4509, Sep. 2019.
  • [20] X. Qian, Y. Fu, T. Xiang, W. Wang, J. Qiu, Y. Wu, Y.-G. Jiang, and X. Xue, “Pose-normalized image generation for person re-identification,” in Proc. ECCV, 2018, pp. 650–667.
  • [21] G. Wang, Y. Yuan, X. Chen, J. Li, and X. Zhou, “Learning discriminative features with multiple granularities for person re-identification,” in Proc. ACM Multimedia, 2018, pp. 274–282.
  • [22] Y. Suh, J. Wang, S. Tang, T. Mei, and K. Mu Lee, “Part-aligned bilinear representations for person re-identification,” in Proc. ECCV, 2018, pp. 402–419.
  • [23] D. Cheng, Y. Gong, S. Zhou, J. Wang, and N. Zheng, “Person re-identification by multi-channel parts-based cnn with improved triplet loss function,” in Proc. CVPR, June 2016, pp. 1335–1344.
  • [24] X. Fan, H. Luo, X. Zhang, L. He, C. Zhang, and W. Jiang, “Scpnet: Spatial-channel parallelism network for joint holistic and partial person re-identification,” in Proc. ACCV.   Springer, 2018, pp. 19–34.
  • [25] Terrance DeVries and Graham W Taylor. (2018), “Improved regularization of convolutional neural networks with cutout,” [Online]. Available: https://arxiv.org/abs/1708.04552.
  • [26] Zhong, Zhun and Zheng, Liang and Kang, Guoliang and Li, Shaozi and Yang, Yi. (2017), “Random erasing data augmentation,” [Online]. Available: https://arxiv.org/abs/1708.04896.
  • [27] K. Zhou, Y. Yang, A. Cavallaro, and T. Xiang, “Omni-scale feature learning for person re-identification,” in Proc. ICCV, 2019, pp. 3702–3712.
  • [28] W. Li, X. Zhu, and S. Gong, “Harmonious attention network for person re-identification,” in Proc. CVPR, June 2018, pp. 2285–2294.
  • [29] C. Wang, Q. Zhang, C. Huang, W. Liu, and X. Wang, “Mancs: A multi-task attentional network with curriculum sampling for person re-identification,” in Proc. ECCV, 2018, pp. 365–381.
  • [30] R. Quan, X. Dong, Y. Wu, L. Zhu, and Y. Yang, “Auto-reid: Searching for a part-aware convnet for person re-identification,” in Proc. ICCV, 2019.
  • [31] F. Zheng, C. Deng, X. Sun, X. Jiang, X. Guo, Z. Yu, F. Huang, and R. Ji, “Pyramidal person re-identification via multi-loss dynamic training,” in Proc. CVPR, June 2019.
  • [32] E. Ristani and C. Tomasi, “Features for multi-target multi-camera tracking and re-identification,” in Proc. CVPR, June 2018, pp. 6036–6046.
  • [33] W. Chen, X. Chen, J. Zhang, and K. Huang, “A multi-task deep network for person re-identification,” in Proc. AAAI, 2017.
  • [34] B. N. Xia, Y. Gong, Y. Zhang, and C. Poellabauer, “Second-order non-local attention networks for person re-identification,” in Proc. ICCV, 2019, pp. 3760–3769.
  • [35] Z. Zhang and M. Huang, “Person re-identification based on heterogeneous part-based deep network in camera networks,” IEEE Transactions on Emerging Topics in Computational Intelligence, vol. Early Access, pp. 1–10, December 2018.
  • [36] X. Chang, T. M. Hospedales, and T. Xiang, “Multi-level factorisation net for person re-identification,” in Proc. CVPR, June 2018, pp. 2109–2118.
  • [37] X. Qian, Y. Fu, Y.-G. Jiang, T. Xiang, and X. Xue, “Multi-scale deep learning architectures for person reidentification,” in Proc. ICCV, 2017.
  • [38] Y. Chen, X. Zhu, S. Gong et al., “Person re-identification by deep learning multi-scale representations,” in Proc. ICCV, 2018.
  • [39] J. Hu, L. Shen, and G. Sun, “Squeeze-and-excitation networks,” in Proc. CVPR, 2018, pp. 7132–7141.
  • [40]

    Y. Wen, K. Zhang, Z. Li, and Y. Qiao, “A discriminative feature learning approach for deep face recognition,” in

    Proc. ECCV.   Springer, 2016, pp. 499–515.
  • [41] L. Zheng, L. Shen, L. Tian, S. Wang, J. Wang, and Q. Tian, “Scalable person re-identification: A benchmark,” in Proc. ICCV, Dec 2015, pp. 1116–1124.
  • [42] E. Ristani, F. Solera, R. Zou, R. Cucchiara, and C. Tomasi, “Performance measures and a data set for multi-target, multi-camera tracking,” in Proc. ECCV.   Springer, 2016, pp. 17–35.
  • [43] W. Li, R. Zhao, T. Xiao, and X. Wang, “Deepreid: Deep filter pairing neural network for person re-identification,” in Proc. CVPR, June 2014, pp. 152–159.
  • [44] Z. Zhong, L. Zheng, D. Cao, and S. Li, “Re-ranking person re-identification with k-reciprocal encoding,” in Proc. CVPR, July 2017, pp. 3652–3661.
  • [45] M. S. Sarfraz, A. Schumann, A. Eberle, and R. Stiefelhagen, “A pose-sensitive embedding for person re-identification with expanded cross neighborhood re-ranking,” in Proc. CVPR, June 2018, pp. 420–429.
  • [46] Y. Shen, T. Xiao, H. Li, S. Yi, and X. Wang, “End-to-end deep kronecker-product matching for person re-identification,” in Proc. CVPR, June 2018, pp. 6886–6895.
  • [47] D. Chen, D. Xu, H. Li, N. Sebe, and X. Wang, “Group consistent similarity learning via deep crf for person re-identification,” in Proc. CVPR, June 2018, pp. 8649–8658.
  • [48] K. Li, Z. Ding, K. Li, Y. Zhang, and Y. Fu, “Support neighbor loss for person re-identification,” in Proc. ACM Multimedia, 2018, pp. 1492–1500.
  • [49] M. Zeng, C. Tian, and Z. Wu, “Person re-identification with hierarchical deep learning feature and efficient xqda metric,” in Proc. ACM Multimedia, 2018, pp. 1838–1846.
  • [50] J. Yang, X. Shen, X. Tian, H. Li, J. Huang, and X.-S. Hua, “Local convolutional neural networks for person re-identification,” in Proc. ACM Multimedia, 2018, pp. 1074–1082.
  • [51] Almazan, Jon and Gajic, Bojana and Murray, Naila and Larlus, Diane. (2018), “Re-id done right: towards good practices for person re-identification,” [Online]. Available: https://arxiv.org/abs/1610.02984.
  • [52] Y. Wang, L. Wang, Y. You, X. Zou, V. Chen, S. Li, G. Huang, B. Hariharan, and K. Q. Weinberger, “Resource aware person re-identification across multiple resolutions,” in Proc. CVPR, June 2018, pp. 8042–8051.