Learning Resolution-Invariant Deep Representations for Person Re-Identification

07/25/2019 ∙ by Yun-Chun Chen, et al. ∙ Umbo CV Inc. National Taiwan University 4

Person re-identification (re-ID) solves the task of matching images across cameras and is among the research topics in vision community. Since query images in real-world scenarios might suffer from resolution loss, how to solve the resolution mismatch problem during person re-ID becomes a practical problem. Instead of applying separate image super-resolution models, we propose a novel network architecture of Resolution Adaptation and re-Identification Network (RAIN) to solve cross-resolution person re-ID. Advancing the strategy of adversarial learning, we aim at extracting resolution-invariant representations for re-ID, while the proposed model is learned in an end-to-end training fashion. Our experiments confirm that the use of our model can recognize low-resolution query images, even if the resolution is not seen during training. Moreover, the extension of our model for semi-supervised re-ID further confirms the scalability of our proposed method for real-world scenarios and applications.



There are no comments yet.


page 1

page 3

page 7

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.


Aiming at matching images of the same person across different camera views, person re-identification (re-ID) [Zheng, Yang, and Hauptmann2016]

is among the active research topics in computer vision and machine learning. With a wide range of applications ranging from video surveillance to computational forensics, person re-ID has received substantial attention of communities from both academia and industry. Nevertheless, with the presence of background clutters, viewpoint and pose changes, and even occlusion, person re-ID remains a very challenging task.

Figure 1: Illustration and challenges of cross-resolution person re-identification (re-ID). In addition to recognizing images across different camera views, one also needs to match cross-resolution images.

While a number of methods [Lin et al.2017, Hermans, Beyer, and Leibe2017, Zhong et al.2018, Si et al.2018] have been proposed to address the aforementioned issues in person re-ID, these methods typically assume that the images (both gallery and query) are of similar or sufficient resolution. However, this assumption may not hold in real-world scenarios, since image resolution may vary drastically due to the distance between the camera and the person of interest. For instance, images captured by surveillance cameras (i.e., the queries to be recognized) are often of low resolution (LR) whereas the gallery ones typically have high resolution (HR). However, directly matching an LR query image against the HR gallery ones would entail a non-trivial resolution mismatch problem, as illustrated in Figure 1.

To address cross-resolution person re-ID, one can simply up-sample the LR images by leveraging super-resolution (SR) approaches like [Jiao et al.2018, Wang et al.2018b] to synthesize HR images. However, since these two tasks are addressed separately, there is no guarantee that synthesized HR outputs would result in satisfactory re-ID performances. Moreover, if the input image resolution is not seen by the SR model, then one cannot properly recover the HR outputs. Later in the experiments, we will verify the above issues.

In this paper, we propose a novel Resolution Adaptation and re-Identification Network (RAIN) for cross-resolution person re-ID. Based on the generative adversarial network (GAN) [Goodfellow et al.2014] with an end-to-end learning strategy, our RAIN is trained to extract resolution-invariant image representations, without the limitation (or assumption) of the use of LR inputs with pre-determined resolutions. More specifically, our RAIN is able to handle unseen LR images with satisfactory re-ID performances. For example, given training LR images with pixels and pixels, our model is able to recognize query images with pixels as well (prior re-ID methods requiring SR models may not properly handle LR images with unseen resolution). Finally, as image labeling is of high labor cost in real-world applications, we conduct a series of semi-supervised experiments, which supports the use and extension of our RAIN for cross-resolution person re-ID in such practical yet challenging settings.

The contributions of this paper are highlighted below:

  • We present an end-to-end trainable network that learns resolution-invariant deep representations for cross-resolution person re-ID.

  • Our advance multi-level adversarial network components in our proposed architecture effectively aligns and extracts feature representations across resolutions.

  • We demonstrate the robustness of our model in handling a range of (and even unseen) resolutions for LR query inputs, while standard SR models are required to train on images with particular resolutions.

  • Extensive experiments are performed to verify the effectiveness of our model, and confirm its use for re-ID in semi-supervised settings.

Related Work

Person re-ID has been widely studied in the literature. Most of the existing methods [Cheng et al.2016, Lin et al.2017, Kalayeh et al.2018, Si et al.2018, Chang, Hospedales, and Xiang2018, Li, Zhu, and Gong2018, Liu et al.2018, Wei et al.2018, Song et al.2018, Chen et al.2018a, Shen et al.2018] focus on tackling the challenges of matching images with viewpoint and pose variations, or those with background clutter or occlusion presented. For example, Liu et al. [Liu et al.2018] develop a pose-transferable GAN-based [Goodfellow et al.2014] framework to address image pose variations. Chen et al. [Chen et al.2018a]

integrate the conditional random field (CRF) with deep neural networks to learn more consistent multi-scale similarity metrics. The DaRe 

[Wang et al.2018a] combines the feature embeddings extracted from different convolutional layers into a single embedding to train the model in a supervised fashion. Several attention-based methods [Si et al.2018, Li, Zhu, and Gong2018, Song et al.2018] are further proposed to focus on learning the discriminative parts to mitigate the effect of background clutter. While promising results have been presented, the above approaches typically assume that all images (both query and gallery) are of the same (or similar) resolution, which might not be practical in real-world re-ID applications.

To address the challenging resolution mismatch problem, a couple of methods [Li et al.2015, Jing et al.2015, Wang et al.2016, Jiao et al.2018, Wang et al.2018b, Li et al.2019] have been recently proposed. Li et al. [Li et al.2015] present a joint learning framework that simultaneously optimizes cross-scale image domain alignment and discriminant distance metric modeling. The SLD[Jing et al.2015] learns a pair of HR and LR dictionaries and the mapping between the feature representations of HR and LR images. Wang et al. [Wang et al.2016] explore the scale-distance function space by varying the image scale of LR images when matching with HR ones. Nevertheless, the above methods employ hand-crafted descriptors, which might limit the generalization of their re-ID capability.

Driven by the recent success of convolutional neural networks (CNNs), a few CNN-based re-ID methods 

[Jiao et al.2018, Wang et al.2018b] are proposed. For example, the SING [Jiao et al.2018] comprises an SR network and a person re-ID model to address the LR re-ID problem. Wang et al. [Wang et al.2018b] propose the CSR-GAN which cascades multiple SR-GANs [Ledig et al.2017] in series to alleviate the resolution mismatch problem. Although remarkable improvements have been presented, the aforementioned methods require the learning of a separate SR model. Treating SR and re-ID as independent tasks, there is no guarantee that solving one task well would be preferable for addressing the other. Moreover, if the resolution of the LR query input is not seen during training, the CSR-GAN [Wang et al.2018b] cannot directly apply the learned SR models for synthesizing the HR images whereas the SING [Jiao et al.2018] requires to fuse the results produced by multiple learned models, each of which is specifically designed for a particular resolution. Namely, such models cannot be easily extended to cross-resolution person re-ID.

To overcome the above limitations, our method advances the architecture of the GAN and the autoencoder, which learns cross-resolution deep image representations for re-ID purposes. Our method not only allows LR queries with unseen resolution, but can be extended for solving cross-resolution re-ID in semi-supervised settings. The details of our proposed model will be discussed in the next section.

Proposed Method

Figure 2: Overview of the proposed Resolution Adaptation and re-Identification Network (RAIN). The RAIN consists of a cross-resolution feature extractor (in gray), a high-resolution decoder (in yellow), a resolution discriminator

(in orange), and a re-ID classifier

(in green). The associated loss functions (in white) are the high-resolution reconstruction loss

, adversarial loss , classification loss , and the triplet loss . Note that denotes the index of feature level.

Notations and Algorithmic Overview

For the sake of completeness, we first define the notations to be used in this paper. We assume that we have access to a set of HR images with the associated label set , where and represent the HR image and its corresponding identity label, respectively. To synthesize LR images for training purposes, we generate a synthetic image set by down-sampling each image in , followed by resizing them back to the original image size via bilinear up-sampling (i.e., ), where is the synthetic LR image associated with (with same label). Thus, the label set for is identical to .

To achieve cross-resolution person re-ID, we present an end-to-end trainable network, Resolution Adaptation and re-Identification Network (RAIN). As presented in Figure 2, our RAIN learns resolution-invariant deep representations from training HR and LR images (note that we only need to down-sample the HR training images to produce the LR ones).

As for testing, our proposed RAIN allows query images with varying resolutions; more specifically, we not only allow query images with HR or LR resolutions which are seen during training, but our model can further handle LR images with intermediate resolutions, or resolutions lower than those of the training images (i.e., those not seen during training). In the following subsections, we will detail the network components of RAIN.

Architecture of RAIN

Our proposed network, Resolution Adaptation and re-Identification Network (RAIN), includes a number of network components. The cross-resolution feature extractor encodes input images across different resolutions and produces image features for both image recovery and person re-ID. The high-resolution decoder reconstructs the encoded cross-resolution features to the HR outputs. The discriminator aligns image features across resolutions via adversarial learning, and thus enforces the learning of resolution-invariant features. Finally, the re-ID classifier is learned via classification and triplet losses.

Cross-resolution feature extractor .

Given an HR image and an LR image  111For simplicity, we would omit the subscript , denote HR and LR images as and , and represent their corresponding labels as and in this paper., we first forward and to the cross-resolution feature extractor to obtain their feature maps. In this paper, we adopt the ResNet- [He et al.2016] as the cross-resolution feature extractor , which has five residual blocks . We denote the feature maps extracted from the last activation layer of each residual block as , where and is the number of channels.

Since our goal is to perform cross-resolution person re-ID, we encourage the cross-resolution feature extractor to generate similar feature distributions when observing both and . To accomplish this, we advance the strategy of adversarial learning, and introduce a discriminator . This discriminator takes in the feature maps and as inputs to distinguish whether the input feature map is from or . Note that represents the index of the feature level and and denote the feature maps of and , respectively.

To train the cross-resolution feature extractor and the discriminator with cross-resolution input images and , we define the adversarial loss as


High-resolution decoder .

To reduce the information loss in the above feature extraction stage, we introduce a high-resolution decoder

that takes in the feature map extracted from the cross-resolution feature extractor as the input. In contrast to existing autoencoder-based methods that encourage the decoder to recover the original images given the observed latent features (i.e., self reconstruction), we explicitly enforce our HR decoder to reconstruct the HR images using features derived from the cross-resolution feature extractor . This would further allow to extract cross-resolution image features, while having focus on synthesizing the HR outputs.

To achieve the above goal, we impose an HR reconstruction loss between the outputs of the HR decoder and the corresponding HR ground truth images. Specifically, the HR reconstruction loss is defined as


Note that is the HR image corresponding to . Following [Huang et al.2018], we also use the norm to calculate the HR reconstruction loss , since it would preserve image sharpness.

Re-ID classifier .

To utilize labeled information of training data for cross-resolution person re-ID, we finally introduce a classifier

in our RAIN. The input of this classifier is the feature vector

from the global average pooling (GAP) layer on the feature map , i.e.,  GAP(), where . With person identity as ground truth information, we can compute the negative log-likelihood between the predicted label and the ground truth one-hot vector , and define the classification loss as


where is the number of identities (classes). We note that weighted classification loss [Chen et al.2017] can also be adopted to improve the identity classification performance.

To further enhance the discriminative property, we impose a triplet loss on the feature vector , which would maximize the inter-class discrepancy while minimizing intra-class distinctness. To be more specific, for each input image , we sample a positive image with the same identity label and a negative image with different identity labels to form a triplet tuple. Then, the following equations compute the distances between and /:


where , , and represent the feature vectors of images , , and , respectively.

With the above definitions, we have the triplet loss defined as


where is the margin used to define the distance difference between the distance of positive image pair and the distance of negative image pair .

We note that minimizing the triplet loss in (6) is equivalent to minimizing the intra-class distinctness in (4) while maximizing the inter-class discrepancy in (5).

Total loss.

Finally, the total loss function for training the proposed RAIN is summarized as follows:


With the above total loss, we aim to solve the min-max criterion:


In other words, to train our RAIN using training HR images (and the down-sampled LR ones), we suppress the classification loss , the triplet loss , and the HR reconstruction loss while matching feature representations across resolutions.

Rank 1 Rank 5 Rank 10 Rank 20 Rank 1 Rank 5 Rank 10 Rank 20 Rank 1 Rank 5 Rank 10 Rank 20
JUDEA [Li et al.2015] 26.2 58.0 73.4 87.0 26.0 55.1 69.2 82.3 22.0 60.1 80.8 98.1
SLD[Jing et al.2015] - - - - 20.3 44.0 62.0 78.2 18.4 44.8 61.2 83.6
SDF [Wang et al.2016] 22.2 48.0 64.0 80.0 9.25 38.1 52.4 68.0 14.3 37.5 62.5 95.2
SING [Jiao et al.2018] 67.7 90.7 94.7 97.4 33.5 57.0 66.5 76.6 33.5 72.7 89.0 98.6
CSR-GAN [Wang et al.2018b] - - - - 37.2 62.3 71.6 83.7 - - - -
Baseline (train on HR) 60.6 89.4 95.0 98.4 32.5 59.2 69.0 76.2 27.5 63.2 79.3 92.2
Baseline (train on HR & LR) 65.9 92.1 97.4 98.9 36.6 62.3 70.9 82.2 31.7 68.4 84.2 94.1
Ours (single-level) 77.6 96.2 98.5 99.3 41.2 66.3 75.6 87.1 41.5 75.3 85.6 95.8
Ours (multi-level) 78.9 97.3 98.7 99.5 42.5 68.3 79.6 88.0 42.0 77.3 89.6 98.7
Table 1: Experimental results of cross-resolution person re-ID (%). Note that the numbers in bold denote the best results.


We describe the datasets and settings for evaluation.


We perform evaluations on three benchmark datasets, including two synthetic and one real-world person re-ID datasets. We will explain how we synthesize the LR images for each dataset to perform multiple low-resolution (MLR) person re-ID.


The MLR-CUHK03 dataset is a synthetic dataset built from CUHK03 [Li et al.2014] which consists of different camera views with more than images of person identities. For each camera pair, we down-sample images of one camera by randomly selecting a down-sampling rate (i.e., the size of the down-sampled image will be ), while the image resolution of the other camera view remains the same.


The MLR-VIPeR dataset is a synthetic dataset built from VIPeR [Gray and Tao2008] which contains person-image pairs captured by two cameras. Similarly, we down-sample all the images captured by one camera view using the randomly selected down-sampling rate , while the image resolution of the other camera is fixed.


The more challenging CAVIAR dataset [Cheng et al.2011] is a genuine LR person re-ID dataset which contains images of person identities captured from two camera views. Since the images captured by the more distant camera have much lower resolution than those captured by the closer camera, this dataset is suitable for evaluating the cross-resolution person re-ID. Following [Jiao et al.2018], we discard people who only appear in the closer camera. In contrast to other synthetic datasets, this dataset inherently contains multiple realistic resolutions.

Experimental Settings and Protocols

We consider cross-resolution person re-ID where the query set contains LR images while the gallery set is composed of HR images only. We adopt the standard single-shot person re-ID setting in all of our experiments. Following [Wang et al.2016], we randomly divide the MLR-VIPeR and the CAVIAR datasets into halves for training and test set, with / training/test identity split for the MLR-CUHK03 dataset. The test (query) set is constructed with all LR images for each person identity while the gallery image set contains one randomly selected HR image for each person.

For performance evaluation, we adopt the average cumulative match characteristic and report the results recorded at ranks , , , and . We adopt the multi-level discriminator which adapts feature distributions at different feature levels. Due to the balance between efficiency and performance, we select the index of feature level with and denote our method as “Ours (multi-level)” and the variant of our method with single-level discriminator () as “Ours (single-level)”.

Evaluation and Comparisons

We compare our approach with the JUDEA [Li et al.2015], the SLD[Jing et al.2015], the SDF [Wang et al.2016], the SING [Jiao et al.2018], and the CSR-GAN [Wang et al.2018b]

. We note that our cross-resolution feature extractor is only pre-trained on the ImageNet 

[He et al.2016]. However, SING [Jiao et al.2018] and CSR-GAN [Wang et al.2018b] require their re-ID networks to be pre-trained on large-scale re-ID datasets like Market1501 [Zheng et al.2015] (which contains images of person identities).

Table 1 lists the quantitative results on the three datasets. We note that our results can be further improved by applying pre-processing or post-processing method, attention mechanisms, or re-ranking. For fair comparisons, no such techniques are applied.


Our method achieves % for single-level discriminator and % for multi-level discriminator at rank . The proposed method performs favorably against the state-of-the-art methods and outperforms the previous best competitor [Jiao et al.2018] by a large margin % for single-level discriminator and % for multi-level discriminator at rank . Our performance gains can be ascribed to the following two factors. First, unlike most existing re-ID methods, our model performs cross-resolution person re-ID and is trained in an end-to-end learning fashion. Second, the proposed approach would not suffer from the visual artifacts as our model does not leverage SR models.

Furthermore, the advantage of training on both HR and LR images, and introducing the discriminator can be observed by comparing the method “Ours (single-level)” with two baseline methods “Baseline (train on HR)” and “Baseline (train on HR & LR)”, respectively. Note that the method “Baseline (train on HR)” is considered as a naive method that only trains on HR images. A % performance drop can be observed from the method “Baseline (train on HR)”. This indicates that the resolution mismatch problem significantly alters the performance if the model is trained on HR images only. On the other hand, the method “Baseline (train on HR & LR)” trained on both HR and LR images without applying the adversarial loss still suffers a % performance drop. The result suggests that even if the model is trained with images of multiple resolutions, without the adversarial loss which aligns image features across resolutions, the resolution mismatch problem still implicitly alters the performance.

Train Test MLR-VIPeR
33.5 37.2 42.5
Table 2: Experimental results of cross-resolution person re-ID with seen and unseen resolutions on the MLR-VIPeR dataset evaluated at rank 1 (%).

(a) Colorization with respect to identity.

(b) Colorization with respect to resolution.
Figure 3: Visualization of cross-resolution feature vectors on MLR-CUHK03 via t-SNE. (a) different identities, each of which is shown in a unique color. (b) With four different down-sampling rates () are considered and shown, images with the same resolution are shown in the same color. Note that images with are not seen during training.


Our method achieves the state-of-the-art performance on all four ranks. The performance gains over the best competitor [Wang et al.2018b] at rank are % for single-level discriminator and % for multi-level discriminator.

In addition to performance comparisons, our method can reliably perform cross-resolution person re-ID and generalize well on unseen image resolutions. However, most existing methods [Jiao et al.2018, Wang et al.2018b] may not properly handle image resolutions that are not seen by their SR models or require to fuse the results produced by multiple learned models, each of which is specifically designed for a particular resolution.

We present and compare such results in Table 2. Suppose that the training set contains images with different down-sampling rates ( indicates that images remain their original sizes), if the test set contains images with down-sampling rates which have appeared in the training set, both existing methods [Jiao et al.2018, Wang et al.2018b] and our approach perform cross-resolution person re-ID properly. However, if we consider another scenario where the training set contains images with down-sampling rates , whereas the test set contains images with down-sampling rate which are not seen during training, the proposed model works properly and reliably performs cross-resolution person re-ID with satisfactory result. However, existing methods could not handle unseen resolutions properly due to the following reasons. For CSR-GAN [Wang et al.2018b], their SR models are resolution-dependent and cannot directly apply to LR images of unseen resolutions for synthesizing HR images. For SING [Jiao et al.2018], even though they can still apply their SR models on the images of unseen resolutions via fusing the results produced by several models, their SR models are specifically designed for some particular image resolutions, which will not reliably address cross-resolution person re-ID with images of unseen resolutions.


For the CAVIAR dataset, our method achieves % for single-level discriminator and % for multi-level discriminator at rank score achieving the state-of-the-art performance on all four evaluated ranks. The performance gains over the best competitor [Jiao et al.2018] measured at rank are % for single-level discriminator and % for multi-level discriminator.

Method MLR-CUHK03
Rank 1 Rank 5 Rank 10 Rank 20 mAP
Ours 78.9 97.3 98.7 99.5 74.5
Ours w/o 70.8 95.1 97.7 98.9 68.0
Ours w/o 69.1 92.2 96.6 98.7 64.1
Ours w/o 67.3 89.5 94.5 97.7 64.2
Ours w/o 65.9 92.1 97.4 98.9 62.3
Table 3: Ablation studies on the MLR-CUHK03 dataset (%).
HR LR Rank 1 mAP Rank 1 mAP Rank 1 mAP
70.9 67.7 38.9 42.6 36.3 52.9
72.3 68.8 40.3 43.2 37.9 53.1
77.2 74.8 41.5 45.4 40.1 54.6
78.9 75.9 42.5 47.0 42.0 56.3
Table 4: Effect of training images of multiple low resolutions. The bold numbers indicate the best results (%).

Ablation Studies

Loss functions.

To analyze the importance of each loss function, we conduct an ablation study on the MLR-CUHK03 dataset using the multi-level discriminator method abbreviated as “Ours”. Table 3 presents the quantitative results of the ablation experiments evaluated at ranks , , , and , and the mAP. The results show that without the classification loss or the triplet loss , our model still achieves favorable performances compared with the state-of-the-art method [Jiao et al.2018]. This is because both the classification loss and the triplet loss are introduced to control the intra-class and inter-class distances. Without either one of them, our model still has the ability to establish a well separated feature space for each person identity. However, our model suffers a % performance drop without the HR reconstruction loss at rank . The result indicates that introducing the HR decoder and imposing the HR reconstruction loss greatly reduce the information loss. In addition, without the adversarial loss , our model does not learn resolution-invariant representations and a % performance drop at rank can be observed, which indicates that our model suffers from the severe impact induced by the resolution mismatch problem. The ablation experiments show that all loss terms play crucial roles in achieving state-of-the-art performance.

Figure 4: Example of the top-ranked HR gallery images of the MLR-CUHK03 dataset which are matched by the LR query input. Images bounded in green and red rectangles denote correct and incorrect matches, respectively.

Effect of training images of multiple low resolutions.

We conduct experiments on all three adopted datasets with different combinations of down-sampling rates and present the results in Table 4. We observe that when our model (bottom row) is trained with LR images of multiple down-sampling rates, it achieves the best results compared with our three variant methods (first three rows), each of which is trained with LR images of a single down-sampling rate. On the other hand, from the third variant method (the third row), the result demonstrates that our method reliably handles unseen but intermediate resolutions ().

Visualization of cross-resolution feature vector .

We now visualize the feature vectors on the test set of the MLR-CUHK03 dataset in Figures 2(a) and 2(b) via t-SNE.

In Figure 2(a), we select different person identities, each of which is indicated by a color. We observe that the projected feature vectors are well separated, which suggests that sufficient re-ID ability can be exhibited by our model. On the other hand, for Figure 2(b), we colorize each image resolution with a color in each identity cluster (four different down-sampling rates ). It can be observed that the projected feature vectors of the same identity but different down-sampling rates are all well clustered. We note that images with down-sampling rate are not presented in the training set.

The above visualizations demonstrate that our model learns resolution-invariant representations, and is able to generalize well to unseen image resolution (e.g., ) for cross-resolution person re-ID.

Top ranked gallery images.

Given an LR query image with down-sampling rate (the leftmost column), we present the first top-ranked HR gallery images in Figure 4. We compare our method (bottom row) with two baseline methods “Baseline (Train on HR)” (top row) and “Baseline (Train on HR and LR)” (middle row). The green and red boundaries indicate correct and incorrect matches, respectively. From the results in the top row of this figure, we observe that the method “Baseline (Train on HR)” only achieves out of correct matches. When trained with images of various resolutions, the method “Baseline (Train on HR and LR)” improves the matching results to out of correct matches. Finally, our method achieves out of correct matches, which again verify the effectiveness and robustness of our model. Note that the resolution () of the query image is not seen during training.

Figure 5: Semi-supervised cross-resolution person re-ID on the MLR-CUHK03 dataset (%).

Semi-Supervised Cross-Resolution Re-ID

To show that, even only a portion of the dataset are with labels (i.e., computing the classification loss and the triplet loss is only applicable for such data), the unique design of our RAIN would still exhibit sufficient ability in learning cross-resolution person re-ID image features, we conduct a series of semi-supervised experiments.

We increase the amount of labeled data by % each time (i.e., %, %, %, %, %, and % labeled data) and record the performance at rank as presented in Figure 5. Note that the unlabeled data can still compute the HR reconstruction loss and the adversarial loss. We compare our method with two baseline methods “Baseline (train on HR)” and “Baseline (train on HR & LR)”. From Figure 5, we observe that without any labeled information, our method achieves % at rank . When the amount of labeled data is increased to %, our model results in % at rank , which is only slightly worse than the performance of SING [Jiao et al.2018] (%) recorded with % labeled data. When we further increase the amount of labeled data to %, our model achieves % at rank , which outperforms the result of SING [Jiao et al.2018] recorded with fully labeled data by %.

With the experiments, we confirm that the unique design and the integration of cross-resolution feature extractor (with resolution adversarial learning), HR decoder, and classification components, would allow one to learn resolution-invariant representations for cross-resolution person re-ID, even if only a portion of the image data are labeled. Thus, the use of our RAIN for real-world re-ID problems can be supported.


We have presented an end-to-end trainable network, Resolution Adaptation and re-Identification Network (RAIN), which is able to learn resolution-invariant representations for cross-resolution person re-ID. The novelty of this network lies in the use of adversarial learning for deriving latent image features across image resolutions, with an autoencoder-like architecture which preserves the image representation ability. Utilizing image labels, the classification components further exploit the discriminative property for re-ID purposes. From our experiments, we confirm that our model performs favorably against state-of-the-art cross-resolution person re-ID methods. We also verify that our model is able to handle LR query inputs with varying image resolutions, even if such resolutions are not seen during training. Finally, the extension to semi-supervised re-ID further supports the use of our proposed model for solving practical cross-resolution re-ID tasks. In the future, we hope to leverage attention mechanisms by localizing [Chen and Hsu2019] or segmenting [Lin et al.2019] identities to better focus on classifying foreground objects. We also hope our proposed method can facilitate other computer vision tasks such as semantic matching [Chen et al.2018b], object co-segmentation [Chen et al.2019b], and domain adaptation [Chen et al.2019a].

Acknowledgements. This work is supported in part by Umbo Computer Vision and the Ministry of Science and Technology of Taiwan under grant MOST 107-2634-F-002-010.


  • [Chang, Hospedales, and Xiang2018] Chang, X.; Hospedales, T. M.; and Xiang, T. 2018. Multi-level factorisation net for person re-identification. In CVPR.
  • [Chen and Hsu2019] Chen, Y.-C., and Hsu, W. H. 2019. Saliency aware: Weakly supervised object localization. In ICASSP.
  • [Chen et al.2017] Chen, Y.-C.; Li, Y.-J.; Tseng, A.; and Lin, T. 2017. Deep learning for malicious flow detection. In IEEE PIMRC.
  • [Chen et al.2018a] Chen, D.; Xu, D.; Li, H.; Sebe, N.; and Wang, X. 2018a. Group consistent similarity learning via deep crf for person re-identification. In CVPR.
  • [Chen et al.2018b] Chen, Y.-C.; Huang, P.-H.; Yu, L.-Y.; Huang, J.-B.; Yang, M.-H.; and Lin, Y.-Y. 2018b. Deep semantic matching with foreground detection and cycle consistency. In ACCV.
  • [Chen et al.2019a] Chen, Y.-C.; Lin, Y.-Y.; Yang, M.-H.; and Huang, J.-B. 2019a. Crdoco: Pixel-level domain transfer with cross-domain consistency. In CVPR.
  • [Chen et al.2019b] Chen, Y.-C.; Lin, Y.-Y.; Yang, M.-H.; and Huang, J.-B. 2019b. Show, match and segment: Joint learning of semantic matching and object co-segmentation. arXiv.
  • [Cheng et al.2011] Cheng, D. S.; Cristani, M.; Stoppa, M.; Bazzani, L.; and Murino, V. 2011. Custom pictorial structures for re-identification. In BMVC.
  • [Cheng et al.2016] Cheng, D.; Gong, Y.; Zhou, S.; Wang, J.; and Zheng, N. 2016. Person re-identification by multi-channel parts-based cnn with improved triplet loss function. In CVPR.
  • [Goodfellow et al.2014] Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; and Bengio, Y. 2014. Generative adversarial nets. In NIPS.
  • [Gray and Tao2008] Gray, D., and Tao, H. 2008. Viewpoint invariant pedestrian recognition with an ensemble of localized features. In ECCV.
  • [He et al.2016] He, K.; Zhang, X.; Ren, S.; and Sun, J. 2016. Deep residual learning for image recognition. In CVPR.
  • [Hermans, Beyer, and Leibe2017] Hermans, A.; Beyer, L.; and Leibe, B. 2017. In defense of the triplet loss for person re-identification. arXiv.
  • [Huang et al.2018] Huang, X.; Liu, M.-Y.; Belongie, S.; and Kautz, J. 2018.

    Multimodal unsupervised image-to-image translation.

    In ECCV.
  • [Jiao et al.2018] Jiao, J.; Zheng, W.-S.; Wu, A.; Zhu, X.; and Gong, S. 2018. Deep low-resolution person re-identification. In AAAI.
  • [Jing et al.2015] Jing, X.-Y.; Zhu, X.; Wu, F.; You, X.; Liu, Q.; Yue, D.; Hu, R.; and Xu, B. 2015. Super-resolution person re-identification with semi-coupled low-rank discriminant dictionary learning. In CVPR.
  • [Kalayeh et al.2018] Kalayeh, M. M.; Basaran, E.; Gökmen, M.; Kamasak, M. E.; and Shah, M. 2018. Human semantic parsing for person re-identification. In CVPR.
  • [Ledig et al.2017] Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A. P.; Tejani, A.; Totz, J.; Wang, Z.; et al. 2017. Photo-realistic single image super-resolution using a generative adversarial network. In CVPR.
  • [Li et al.2014] Li, W.; Zhao, R.; Xiao, T.; and Wang, X. 2014. Deepreid: Deep filter pairing neural network for person re-identification. In CVPR.
  • [Li et al.2015] Li, X.; Zheng, W.-S.; Wang, X.; Xiang, T.; and Gong, S. 2015. Multi-scale learning for low-resolution person re-identification. In ICCV.
  • [Li et al.2019] Li, Y.-J.; Chen, Y.-C.; Lin, Y.-Y.; Du, X.; and Wang, Y.-C. F. 2019. Recover and identify: A generative dual model for cross-resolution person re-identification. In ICCV.
  • [Li, Zhu, and Gong2018] Li, W.; Zhu, X.; and Gong, S. 2018. Harmonious attention network for person re-identification. In CVPR.
  • [Lin et al.2017] Lin, Y.; Zheng, L.; Zheng, Z.; Wu, Y.; and Yang, Y. 2017. Improving person re-identification by attribute and identity learning. arXiv.
  • [Lin et al.2019] Lin, J.-Y.; Wu, M.-S.; Chang, Y.-C.; Chen, Y.-C.; Chou, C.-T.; Wu, C.-T.; and Hsu, W. H. 2019. Learning volumetric segmentation for lung tumor. IEEE ICIP VIP Cup Tech. report.
  • [Liu et al.2018] Liu, J.; Ni, B.; Yan, Y.; Zhou, P.; Cheng, S.; and Hu, J. 2018. Pose transferrable person re-identification. In CVPR.
  • [Shen et al.2018] Shen, Y.; Li, H.; Xiao, T.; Yi, S.; Chen, D.; and Wang, X. 2018. Deep group-shuffling random walk for person re-identification. In CVPR.
  • [Si et al.2018] Si, J.; Zhang, H.; Li, C.-G.; Kuen, J.; Kong, X.; Kot, A. C.; and Wang, G. 2018. Dual attention matching network for context-aware feature sequence based person re-identification. In CVPR.
  • [Song et al.2018] Song, C.; Huang, Y.; Ouyang, W.; and Wang, L. 2018.

    Mask-guided contrastive attention model for person re-identification.

    In CVPR.
  • [Wang et al.2016] Wang, Z.; Hu, R.; Yu, Y.; Jiang, J.; Liang, C.; and Wang, J. 2016. Scale-adaptive low-resolution person re-identification via learning a discriminating surface. In IJCAI.
  • [Wang et al.2018a] Wang, Y.; Wang, L.; You, Y.; Zou, X.; Chen, V.; Li, S.; Huang, G.; Hariharan, B.; and Weinberger, K. Q. 2018a. Resource aware person re-identification across multiple resolutions. In CVPR.
  • [Wang et al.2018b] Wang, Z.; Ye, M.; Yang, F.; Bai, X.; and Satoh, S. 2018b. Cascaded sr-gan for scale-adaptive low resolution person re-identification. In IJCAI.
  • [Wei et al.2018] Wei, L.; Zhang, S.; Gao, W.; and Tian, Q. 2018. Person transfer gan to bridge domain gap for person re-identification. In CVPR.
  • [Zheng et al.2015] Zheng, L.; Shen, L.; Tian, L.; Wang, S.; Wang, J.; and Tian, Q. 2015. Scalable person re-identification: A benchmark. In ICCV.
  • [Zheng, Yang, and Hauptmann2016] Zheng, L.; Yang, Y.; and Hauptmann, A. G. 2016. Person re-identification: Past, present and future. arXiv.
  • [Zhong et al.2018] Zhong, Z.; Zheng, L.; Zheng, Z.; Li, S.; and Yang, Y. 2018. Camera style adaptation for person re-identification. In CVPR.