DARI: Distance metric And Representation Integration for Person Verification

by   Guangrun Wang, et al.
NetEase, Inc

The past decade has witnessed the rapid development of feature representation learning and distance metric learning, whereas the two steps are often discussed separately. To explore their interaction, this work proposes an end-to-end learning framework called DARI, i.e. Distance metric And Representation Integration, and validates the effectiveness of DARI in the challenging task of person verification. Given the training images annotated with the labels, we first produce a large number of triplet units, and each one contains three images, i.e. one person and the matched/mismatch references. For each triplet unit, the distance disparity between the matched pair and the mismatched pair tends to be maximized. We solve this objective by building a deep architecture of convolutional neural networks. In particular, the Mahalanobis distance matrix is naturally factorized as one top fully-connected layer that is seamlessly integrated with other bottom layers representing the image feature. The image feature and the distance metric can be thus simultaneously optimized via the one-shot backward propagation. On several public datasets, DARI shows very promising performance on re-identifying individuals cross cameras against various challenges, and outperforms other state-of-the-art approaches.



There are no comments yet.


page 1


Deep Feature Learning with Relative Distance Comparison for Person Re-identification

Identifying the same individual across different scenes is an important ...

In Defense of the Triplet Loss for Person Re-Identification

In the past few years, the field of computer vision has gone through a r...

Discriminative Feature Learning with Foreground Attention for Person Re-Identification

The performance of person re-identification (Re-ID) seriously depends on...

Consistent Cross-view Matching for Unsupervised Person Re-identification

Most existing unsupervised person re-identificationmethods focus on lear...

Cross-media Similarity Metric Learning with Unified Deep Networks

As a highlighting research topic in the multimedia area, cross-media ret...

Multiple Target Tracking by Learning Feature Representation and Distance Metric Jointly

Designing a robust affinity model is the key issue in multiple target tr...

Deep Bayes Factor Scoring for Authorship Verification

The PAN 2020 authorship verification (AV) challenge focuses on a cross-t...
This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.


Figure 1: Typical examples of person verification across cameras. Each column corresponds to one individual, and the large variations exist between the two examples due to the light, pose and view point changes.

Distance/similarity measure between images plays a fundamental role in many computer vision applications, e.g., person verification

[Hirzer et al.2012], matching heterogeneous data [Zhai et al.2012], and multimedia retrieval [Chechik et al.2012]

. Conventional Mahalanobis distance metric learning models, which aim to seek a linear transformation by pulling the distances between similar pairs while pushing the distances between dissimilar pairs, are theoretically appealing in producing discriminative similarity or distance measure from the given training data

[Weinberger, Blitzer, and Saul2005a]. However, these approaches usually are performed in the original data space or the hand-engineered feature space (i.e. representation), and thus are limited in capturing variety of image appearance and handling complicated nonlinear manifold.

In this paper, we investigate the possible interaction between feature learning and distance metric learning, and address the very challenging task of person verification (i.e. matching individuals across cameras). Figure 1 shows some examples of this task, where the pedestrians across cameras distinctly vary in appearance and structure caused by pose, lighting and view angle changes. The proposed framework is built based on the convolutional neural network (CNN) [LeCun et al.1989], which jointly optimizes the human representation as well as the proper distance metric to robustly match individuals against various real challenges. We call this framework DARI (i.e. Distance metric And Representation Integration).

We aim at preserving similarity of the same person while discriminating the different individuals, and thus define DARI in the form of maximizing relative distance. Specifically, we organize the training images into triplet units, and each unit contains one person image and the matched/mismatch references. For all triplet units, the objective of DARI is to maximize the distance disparity between the matched pairs and the mismatched pairs. In particular, DARI seeks optimal solutions for both feature representation and distance metric, while the existing approaches of person verification [Xu et al.2013, Liu et al.2012]

focuses on only either of the two components. We adopt the deep CNN architecture to extract the discriminative features from the input images, where the convolutional layers, max-pooling operators, and one full connection layer are stacked up. And the Mahalanobis distance matrix is applied with the generated feature as the matching metric. Due to the positive semi-definite requirement for the Mahalanobis metric, directly optimizing the metric matrix is computational intensive. We propose to decompose the Mahalanobis matrix inspired by 

[Mignon and Jurie2012], and further factorize the matrix into a fully-connected layer on the top of our deep architecture. In this way, the distance metric is seamlessly integrated with the image feature represented by the other layers of neural networks. The joint optimization can be then efficiently achieved via the standard backward propagation. Therefore, by means of the nonlinearity learning of deep neural networks, DARI is capable of representing the complicated transformation to identify the people in the wild.

To scale up our approach to the large amount of training data, we implement the training in a batch-process fashion. In each round of training, we randomly select a relatively small number (say

) of images, and use them to organize the triplet units. By taking the triplets as the inputs, we update the model parameters by the stochastic gradient descent (SGD) algorithm 

[LeCun et al.1998]. Another arising issue is that the triplet organization cubically enlarges the number (say ) of training samples, as one image can be included into more than one triplet. To overcome it, we calculate the gradients on the images instead of the produced triplets, and thus reduce the computation cost by making it only depends on the number of the selected images.

The key contribution of this paper is a novel end-to-end framework that naturally fuses the concept of feature learning and metric learning via the deep neural networks. To the best of our knowledge, such an approach is original to the community. On several challenging benchmarks for person verification (e.g., CUHK03 [Li et al.2014],CUHK01[Li, Zhao, and Wang2012] and iLIDS [Zheng, Gong, and Xiang2013]), our DARI framework demonstrates superior performances over other state-of-the-art approaches.

Related Work

A number of approaches, e.g., local metric learning and kernelized metric learning, have been suggested to learn multiple or nonlinear metrics from training data with complicated nonlinear manifold structure. In local metric learning, local metrics can be learned independently for each region or by considering the data manifold structure [Noh, Zhang, and Lee2010, Wang, Kalousis, and Woznica2012, Weinberger, Blitzer, and Saul2005b]. In kernelized metric learning, a kernel function is exploited to implicitly embed instances into the reproducing kernel Hilbert space (RKHS), and a Mahalanobis distance metric is then learned in the RKHS space [Wang et al.2011]. Actually, kernelized method with Gaussian RBF kernel can also be treated as local learning approach. As pointed out in [Bengio2009], local learning are also shallow models, and generally are insufficient in coping with highly varying appearance and deformations of images. Another efficient local distance metric learning [Yang et al.2006] was also proposed for classification and retrieval. To handle heterogeneous data, [Xiong et al.2012]

propose a method using a random forest-based classifier to strengthen the distance function with implicit pairwise position dependence.

On the other hand, deep convolutional models have been intensively studied and achieved extremely well performance. Compared with the multiple layer perceptron, CNN contains much less parameters to be learned, and can be efficiently trained using stochastic gradient descent. With the increasing of large scale training data and computational resources, deeper CNN and novel regularization methods had been developed, and deep CNN has gained great success in many visual recognition tasks, e.g., image classification

[Krizhevsky, Sutskever, and Hinton2012a], object detection [Szegedy, Toshev, and Erhan2013], and scene labeling [Pinheiro and Collobert2014].

Despite the success of deep learning in variety of vision tasks, little studies were conducted on metric learning with deep architecture. Chopra et al.

[Chopra, Hadsell, and LeCun2005]

suggested a energy-based model (EBM) for discriminative similarity metric learning for image pairs. Stacked restricted Boltzmann machines (RBMs) had also been exploited to learn nonlinear transformation for data visualization and supervised embedding

[Min et al.2010]. Cai et al. [Cai et al.2010]

proposed a deep nonlinear metric learning method by combining logistic regression and independent subspace analysis. Hu et al.

[Hu, Lu, and Tan2014] adopted the forward multi-layer neural network to learn deep metric for hand-crafted features. Compared with these approaches, the proposed DARI model considers the prominence of CNN in capturing salient and incorporates the Mahalanobis distance with the generated image features into one optimization target for distance metric and representation integration.

One approach close to ours was proposed by Wang et al. [Wang et al.2014]

, which addresses the triplet-based similarity learning for image retrieval. However, our work have significant differences with that work. First, we derive our formulation from a novel angle, i.e. integrating feature learning and distance metric learning. Second, our learning method has advantage in the triplet generation and the batch-based gradient descent learning. Specifically, given

training triplets containing distinct images (), their algorithm optimizes with forward and backward propagations, while only rounds is required for our approach because we derive to calculate the gradient over the images. Last, our deep architecture is specifically designed (only two conv layers are used) and we train our model from scratch, while they utilized the Alex’s model [Krizhevsky, Sutskever, and Hinton2012b]

that is pre-trained on the ImageNet.


Primal Formulation

Given a fixed feature representation, metric learning is to learn a distance function by satisfying the constraint according to the label information of samples. Here we define our formulation via relative distance comparison based on the triplet-based constraint. As is discussed in [Ding et al.2015], the triplet models allows the images of one identity lying on a manifold while maximizing the margin between within-class pairs from between-class pairs, and tends to result in better tradeoff between adaptability and discriminability.

More precisely, the relative distance constraint is defined with a set of triplet units , in which is a pair of matched images (images of the same individual) and contains two mismatched images from the labeled image set with denoting the label. Let denote the metric matrix and denote the feature representations of the th image learned by the feature network with the network parameters . Then Mahalanobis distance between and using the CNN features can be written as follows:


where denotes the feature difference between the image and . For each training triplet in , the desired distance should satisfy: . Let denote , we turn this relative constraints into the minimization of the following hinge-loss like objective function where acts as a regularization term as in [Shen et al.2012].

In the following, we use to denote for notation simplicity. By the definition of , we get the following objective functions:

Figure 2: Illustration of our learning formulation. It tends to maximize the distance disparity between the matched pair and mismatched pair for each triplet, in terms of optimization.

An intuitive solution to Eqn. (2) is to iteratively optimize and with either of them fixed. It is, however, computationally expensive, as the PSD projection is necessarily imposed once is updated to ensure to be positive semi-definite. In this work, to overcome this issue we propose an unified solution by incorporating into the deep neural networks.

Deep Architecture

We factorize the metric matrix into as satisfies the positive semi-definite constraint. The distance measure in Eqn. (1) can be then transformed as,


The objective function in Eqn. (2) can be then reformulated with , as


Thus, we can take as a linear transform on the output of the CNN-based feature representation. In literature, [Weinberger, Blitzer, and Saul2005a] [Mignon and Jurie2012] also proposed to decompose the Mahalanobis distance matrix for simplifying the distance metric learning. They attempt to pursue a low-dimensional projection in Euclidean space which embeds the distance metric while ensuring the PSD constraint. However, their solution are complicated requiring additional hypothesis. In this work, we implement a fully connected layer to represent , which is stacked over the layers representing image features, making the distance metric tightly combined with the deep neural network. Specifically, we treat

as the neuron weight of the layer, and the network can represent

by taking as the input. Then, the feature learning and the distance metric can be thus jointly optimized in an end-to-end way.

In the following, we denote for notation simplicity. As the regularization term on will be automatically implemented by the built-in weight decay mechanisms in CNN networks, we neglect this part in the objective function.


Integrating the metric learning and feature learning into one CNN network yields several advantages. First, this leads to a good property of efficient matching. In particular, for each sample stored in a database, we can precomputed its feature representation and the corresponding decomposed Mahalanobis distance matrix. Then the similarity matching in the testing stage can be very fast. Second, it integrates feature learning and metric learning by building an end-to-end deep architecture of neural networks.

As discussed above, our model defined in Eqn. (5) jointly handles similarity function learning and feature learning. This integration is achieved by building a deep architecture of convolutional neural networks, which is illustrated in Figure 2

. Our deep architecture is composed of two sub-networks: feature learning sub-network and metric learning sub-network. The feature learning sub-network contains two convolution-RELU-pooling layers and one fully-connected layer. Both the pooling layers are max-pooling operations with the size of

and the stride size is set as

pixels. The first convolutional layer includes kernels of size with a stride of pixels. The second convolutional layer contains filters of size and the filter stride step is set as

pixel. A fully-connected layer is followed and it outputs a vector of

dimensions. We further normalize the output vector of this fully-connected layer before it is fed to the metric learning sub-network by , where , denote the value of the

th neuron before and after normalization respectively. Accordingly, the back propagation process accounts for the normalization operation using the chain rule during calculation of the partial derivatives. The metric learning sub-network includes only one fully-connected layer. The neural layer outputs

. In this way, the distance metric is tightly integrated with the feature representations, and they can be jointly optimized during the model training.

Learning Algorithm

Given a labeled dataset with classes (persons) and each class has images, then the number of all possible meaningful triplets is . Even for a dataset of moderate size, it is intractable to load all these triplets into the limited memory for the model training. To overcome this issue, we apply batch learning to optimize the parameters, in which the network parameters are updated by the gradient derived only from a small part of all the triplets in each iteration.

Batch Process

In the batch learning process, we need to generate a subset of triplets from all the possible triplets in each iteration. The simplest method is to generate triplets randomly. However, this method makes the number of distinct images be approximately three times the number of the generated triplets because each triplet contains three images, and the likelihood of two triplets sharing the same image is very low. This triplet generation method is very inefficient because there are only a few distance constraints placed on the selected images in each iteration. Instead, to capitalize on the strength of relative distance comparison model, a more reasonable triplet generation method would be one that satisfies the two following conditions:

1. In each iteration, large number of triplets are generated from small number of images to ensure the selected images can be loaded to the memory while rich distance constraints are posed on these images;

2. When increased numbers of iterations are executed, all the possible relative distance constraints between any two classes should be considered in the training process.

These two principles lead to our proposed triplet generation scheme as follows. In each iteration, we select a fixed number of classes (persons), and construct the triplets only using these selected classes. More precisely, for each image in each class, we randomly construct a certain number of triplets with the matched reference coming from the same class and the mismatched references coming from the remaining selected classes. The complete mini-batch learning process is presented in Algorithm 1.

0:    Training images ;
0:    Network Parameters
1:  while  do
2:     ;
3:     Randomly select a set of classes (persons) from the training set;
4:     Construct a set of triplets from the selected classes;
5:     Calculate the gradient for the generated triplets using Algorithm 2;
7:  end while
Algorithm 1 Learning DARI with batch-process

Parameter Optimization

Under the mini-batch training framework, a key step is to calculate the gradient for the triplet set in each iteration. A straight method is to calculate the gradient for each triplet according to the loss function, and sum these gradients to get the overall gradient. But with this approach three separate memory units and a network propagation would be needed for every triplet. This is inefficient as there will be duplicated network propagation for the same image, recalling that for each batch we generate triplets from a known subset of images. We now show that there exists an optimized algorithm in which the computational load mainly depends on the number of distinct images rather than the number of the triplets.

It would be difficult to write the objective function in Eqn. (5) directly as the sum of image-based loss items because it takes the following form (for notation simplicity, we use to denote in the rest of the paper):

Fortunately, because the loss function for a specific triplet is defined by the outputs of the images in this triplet, the total loss can also be considered as follows, where represents the set of all the distinct images in the triplets and denote the size of the distinct images in the triplets.

By the derivative rule, we have the following equations, where represents the network parameters, represents the feature maps of the image at the layer and denote the partial derivative derived from image .


Eqn. 6 shows that the overall gradient is the sum of the image-based terms (image-based gradient). Eqn. 7 shows that the partial derivative with respect to the feature maps of each image can be calculated recursively. With Eqn. 6 and Eqn. 7, the gradients with respect to the network parameters can be obtained by summing the image based gradients using the network back propagation algorithm. The central premise is that we have computed the partial derivative of the output layer’s activation for every image, which can be easily obtained from Eqn. 5. Algorithm 2 gives the detailed process. This optimized algorithm has two obvious merits:

1. We can conveniently use exiting deep learning implementations such as Caffe

111http://caffe.berkeleyvision.org/ to train our model.

2. The number of network propagation executions can be reduced to the number of distinct images in the triplets, a crucial advantage for large scale datasets.

0:    Training triplets ;
0:    The gradient of network parameters:
1:  Collect all the distinct images in
2:  for all  do
3:     Calculate by forward propagation;
4:  end for
5:  for all  do
6:     ;
7:     for all triplet  do
8:         if  then
9:            if = then
11:            else if = then
13:            else if = then
15:            end if
16:         end if
17:     end for
18:     Set the partial derivative with respect to the outputs using
19:     Calculate using back propagation;
20:     Sum the partial derivative: +=;
21:  end for
Algorithm 2 Calculating gradients for optimization


Datasets and Implementation details. We conduct our experiments using three challenging human verification datasets, i.e. CUHK03[Li et al.2014], CUHK01[Li, Zhao, and Wang2012] and iLIDS[Zheng, Gong, and Xiang2013] . All the images are resized to 250

100 for the experiment. The weights of the filters and the full connection parameters are initialized from two zero-mean Gaussian distributions with standard deviation 0.01 and 0.001 respectively. The bias terms were set with the constant 0. During the training, we select 60 persons to construct 4800 triplets in each iteration. Before feeding to the network, the images are mirrored with 0.5 probability and cropped to the size 230

80 at the center with a small random perturbation to augment the training data. We implement our learning algorithm based on the Caffe framework, where we revise the data layer and loss layer to generate the triplets and apply our loss function. We execute the code on a PC with GTX780 GPU and quad-core CPU. And stop the training process when there are less than 10 triplets whose distance constraints are violated, i.e. the distance between the matched pair is greater than the distance between the mismatched pair.

Evaluation Protocol. We adopt the widely used cumulative match curve (CMC) approach [Gray, Brennan, and Tao2007] for quantitative evaluation. We follow the standard setting to randomly partition each dataset into training set and test set without overlap. In each testing, the test set is further divided into a gallery set and a probe set without overlap for 10 times. A rank rate is obtained for each time, and we use the average rate as the final result.

Component analysis. In order to demonstrate how the joint optimization of distance metric with feature representation contributes to performance, we implement a simplified model for comparison by discarding the distance metric learning(i.e. the last neural layer). In this implementation, we only optimize CNN-based feature representation by the back-propagation method.

Experiments on CUHK03 Dataset. This benchmark [Li et al.2014] is the largest one up to date, which contains 14096 images of 1467 pedestrians collected from 5 different pairs of camera views, making it an ideal place for deep learning. Each person is observed by two disjoint camera views and has an average of 4.8 images in each view. We follow the standard setting of using CUHK03 to randomly partition this dataset for 10 times without overlap, and a training set (including 1367 persons) and a test set (including 100 persons) are obtained. In each testing, the testing set is further randomly divided into a gallery set of 100 images (i.e. one image per person) and a probe set (including images of individuals from different camera views in contrast to the gallery set) without overlap for 10 times.

Figure 3: Quantitative results on the three datasets: (a) CUHK03, (b) CUHK01 and (c) iLIDS dataset. Our DARI framework leads superior performances over existing state-of-the-arts overall. Note that “Ours-nj” represents a simplified version of our model, i.e. discarding the joint optimization of distance metric and CNN-based feature.

We introduce several types of state-of-the-arts in this experiment. First, we adopt Local Distance Metric Learning (LDM) [Guillaumin, Verbeek, and Schmid2009], the learning-to-rank method (RANK) [McFee and Lanckriet2010] for comparison, which learn distance metrics based on a fixed feature representation. Following their implementation, the handcrafted features of dense color histograms and dense SIFT uniformly sampled from patches are adopted. Two methods especially designed for person re-identification are introduced in this experiment: KISSME [Kostinger et al.2012], eSDC [Zhao, Ouyang, and Wang2013]. Moreover, we compare with a recently proposed deep learning method, DRSCH [Zhang et al.2015], FPNN [Li et al.2014] and IDLA [Ahmed, Jones, and Marks2015]. DRSCH [Zhang et al.2015] learns hashing code with regularized similarity for image retrieval and person re-identification. FPNN [Li et al.2014] learns pairs of filters to extract person representation and IDLA [Ahmed, Jones, and Marks2015] is also recently proposed deep learning method for person re-identification.

The results are shown in Fig. 3 (a). It is encouraging to observe that our approach achieves a new state-of-the-art on CUHK03. Note that without the joint optimization of distance metric and representation, the performance (i.e., “ Ours-nj”) degenerates from 55.4% to 45.4%.

Experiments on CUHK01 Dataset.

CUHK01 contains 971 individuals, each of which has two samples captured by two disjoint camera views. We partition this dataset into a training set and a testing set exactly following [Li et al.2014][Ahmed, Jones, and Marks2015]: 100 persons are used for testing and the remaining 871 persons for training. Each person has two images for each view and we randomly select one into the gallery set. Single-shot is adopted in the evaluation of CMC curve.

In addition to comparing with the methods adopted in the experiment on CUHK03, we introduce a recently proposed method which also addresses the interaction of representation learning and metric Learning (denoted as L + X) [Liao et al.2015].

Fig. 3 (b) shows the comparison of our DARI framework with other approaches. DARI achieves a new state of the art, with a rank-1 recognition rate of 65.95%. The gain of the joint optimization of distance metric and CNN-based feature is also clear on this dataset, 65.95% over 57.7%.

Cross-dataset Evaluation

The iLIDS dataset [Zheng, Gong, and Xiang2013] was constructed from video images captured in a busy airport arrival hall. It has 119 pedestrians, totaling 479 images. We conduct a challenging task, i.e. cross-dataset task using this dataset, which accords with the real-world surveillance applications. Specifically, we randomly split this dataset into a gallery set and a probe set: the gallery contains only one image of each pedestrian and the remaining images form the probe set. Our model is trained on CUHK03 and tested on this iLIDS dataset without fine-tuning the parameters.

We compare our DARI with several existing methods such as Xing’s [Xing et al.2002], and MCC [Globerson and Roweis2005]. They all use an ensemble of color histograms and texture histograms as the feature representation. Note that the results reported by these competing methods are generated by a different setting: both of the training and the testing data are from this dataset.

Fig. 3 (c) shows the quantitative results. Our superior performance over other approaches demonstrate the good generalization power of DARI. On the other hand, without incorporating Mahalanobis distance matrix, the performance (i.e. “Ours-nj” in ) clearly degenerates from 42.8% to 36.9%, which highlights the significance of the joint optimization of feature representation and distance metric. In the following, we further evaluate our approach under different implementation setting on iLIDS.

Data Augmentation Strategy. We crop the center of the images with random perturbation to augment the training data. This mechanism can effectively alleviate the over-fitting problems. Without this augmentation scheme, the top 1 performance drop by about 30 percent relatively.

Triplet Generation Scheme. We compared two generation strategy. In the first strategy, we select 60 persons for each iteration and only construct 60 triplets for these persons. In the second strategy, we select the same number of persons while constructing 4800 triplets for these persons. As expected by our analysis, the learning process of the first strategy is much slower than the second strategy and when the learning process of the second strategy converges in 7000 iterations, the performance of the first strategy only achieves about 70 percent of the second strategy.


We have presented a novel deep learning framework incorporating Mahalanobis distance matrix with convolutional neural networks. In future work, we will extend our approach for larger scale heterogeneous data, thereby exploring new applications.


  • [Ahmed, Jones, and Marks2015] Ahmed, E.; Jones, M.; and Marks, T. K. 2015. An improved deep learning architecture for person re-identification. In CVPR. IEEE.
  • [Bengio2009] Bengio, Y. 2009. Learning deep architectures for ai. Foundations and Trends

    in Machine Learning

  • [Cai et al.2010] Cai, X.; Wang, C.; Xiao, B.; Chen, X.; and Zhou, J. 2010. Deep nonlinear metric learning with independent subspace analysis for face verification. In ACM MM, 749–752.
  • [Chechik et al.2012] Chechik, G.; Sharma, V.; Shalit, U.; and Bengio, S. 2012. Large scale online learning of image similarity through ranking. JMLR 11:1109–1135.
  • [Chopra, Hadsell, and LeCun2005] Chopra, S.; Hadsell, R.; and LeCun, Y. 2005. Learning a similarity metric discriminatively, with application to face verification. In CVPR, 539–546.
  • [Ding et al.2015] Ding, S.; Lin, L.; Wang, G.; and Chao, H. 2015. Deep feature learning with relative distance comparison for person re-identification. Pattern Recognition.
  • [Globerson and Roweis2005] Globerson, A., and Roweis, S. T. 2005. Metric learning by collapsing classes. In NIPS, 451–458.
  • [Gray, Brennan, and Tao2007] Gray, D.; Brennan, S.; and Tao, H. 2007. Evaluating appearance models for recognition, reacquisition, and tracking. In PETS. Citeseer.
  • [Guillaumin, Verbeek, and Schmid2009] Guillaumin, M.; Verbeek, J.; and Schmid, C. 2009. Is that you? metric learning approaches for face identification. In ICCV, 498–505.
  • [Hirzer et al.2012] Hirzer, M.; Roth, P. M.; Köstinger, M.; and Bischof, H. 2012. Relaxed Pairwise Learned Metric for Person Re-identification. In ECCV, 780–793.
  • [Hu, Lu, and Tan2014] Hu, J.; Lu, J.; and Tan, Y.-P. 2014. Discriminative deep metric learning for face verification in the wild. In CVPR.
  • [Kostinger et al.2012] Kostinger, M.; Hirzer, M.; Wohlhart, P.; Roth, P. M.; and Bischof, H. 2012. Large scale metric learning from equivalence constraints. In CVPR, 2288–2295. IEEE.
  • [Krizhevsky, Sutskever, and Hinton2012a] Krizhevsky, A.; Sutskever, I.; and Hinton, G. E. 2012a. Imagenet classification with deep convolutional neural networks. In Pereira, F.; Burges, C.; Bottou, L.; and Weinberger, K., eds., NIPS. 1097–1105.
  • [Krizhevsky, Sutskever, and Hinton2012b] Krizhevsky, A.; Sutskever, I.; and Hinton, G. E. 2012b. Imagenet classification with deep convolutional neural networks. In NIPS, 1097–1105.
  • [LeCun et al.1989] LeCun, Y.; Boser, B.; Denker, J.; Henderson, D.; Howard, R.; Hubbard, W.; and Jackel, L. 1989. Backpropagation applied to handwritten zip code recognition. Neural Computation 1:541–551.
  • [LeCun et al.1998] LeCun, Y.; Bottou, L.; Bengio, Y.; and Haffner, P. 1998. Gradient-based learning applied to document recognition. Proceedings of the IEEE 86(11):2278–2324.
  • [Li et al.2014] Li, W.; Zhao, R.; Xiao, T.; and Wang, X. 2014. Deepreid: Deep filter pairing neural network for person re-identification. In CVPR, 152–159. IEEE.
  • [Li, Zhao, and Wang2012] Li, W.; Zhao, R.; and Wang, X. 2012. Human reidentification with transferred metric learning. In ACCV, 31–44.
  • [Liao et al.2015] Liao, S.; Hu, Y.; Zhu, X.; and Li, S. Z. 2015. Person re-identification by local maximal occurrence representation and metric learning. In CVPR. IEEE.
  • [Liu et al.2012] Liu, C.; Gong, S.; Loy, C. C.; and Lin, X. 2012. Person re-identification: what features are important? In ECCV, 391–401. Springer.
  • [McFee and Lanckriet2010] McFee, B., and Lanckriet, G. R. 2010. Metric learning to rank. In ICML, 775–782.
  • [Mignon and Jurie2012] Mignon, A., and Jurie, F. 2012. Pcca: A new approach for distance learning from sparse pairwise constraints. In CVPR, 2666–2672. IEEE.
  • [Min et al.2010] Min, R.; van der Maaten, L.; Yuan, Z.; Bonner, A.; and Zhang, Z. 2010. Deep supervised t-distributed embedding. In ICML.
  • [Noh, Zhang, and Lee2010] Noh, Y.-K.; Zhang, B.-T.; and Lee, D. 2010. Generative local metric learning for nearest neighbor classification. In Lafferty, J.; Williams, C.; Shawe-Taylor, J.; Zemel, R.; and Culotta, A., eds., NIPS. 1822–1830.
  • [Pinheiro and Collobert2014] Pinheiro, P., and Collobert, R. 2014. Recurrent convolutional neural networks for scene labeling. In ICML, 82–90.
  • [Shen et al.2012] Shen, C.; Kim, J.; Wang, L.; and Van Den Hengel, A. 2012. Positive semidefinite metric learning using boosting-like algorithms. JMLR 98888(1):1007–1036.
  • [Szegedy, Toshev, and Erhan2013] Szegedy, C.; Toshev, A.; and Erhan, D. 2013. Deep neural networks for object detection. In Burges, C.; Bottou, L.; Welling, M.; Ghahramani, Z.; and Weinberger, K., eds., NIPS. Curran Associates, Inc. 2553–2561.
  • [Wang et al.2011] Wang, J.; Do, H. T.; Woznica, A.; and Kalousis, A. 2011. Metric learning with multiple kernels. In Shawe-Taylor, J.; Zemel, R.; Bartlett, P.; Pereira, F.; and Weinberger, K., eds., NIPS. 1170–1178.
  • [Wang et al.2014] Wang, J.; Leung, T.; Rosenberg, C.; Wang, J.; Philbin, J.; Chen, B.; Wu, Y.; et al. 2014. Learning fine-grained image similarity with deep ranking. In CVPR.
  • [Wang, Kalousis, and Woznica2012] Wang, J.; Kalousis, A.; and Woznica, A. 2012. Parametric local metric learning for nearest neighbor classification. In Pereira, F.; Burges, C.; Bottou, L.; and Weinberger, K., eds., NIPS. 1601–1609.
  • [Weinberger, Blitzer, and Saul2005a] Weinberger, K. Q.; Blitzer, J.; and Saul, L. K. 2005a. Distance metric learning for large margin nearest neighbor classification. In NIPS, 1473–1480.
  • [Weinberger, Blitzer, and Saul2005b] Weinberger, K. Q.; Blitzer, J.; and Saul, L. K. 2005b. Distance Metric Learning for Large Margin Nearest Neighbor Classification. JMLR 10:207–244.
  • [Xing et al.2002] Xing, E. P.; Jordan, M. I.; Russell, S.; and Ng, A. Y. 2002. Distance metric learning with application to clustering with side-information. In NIPS, 505–512.
  • [Xiong et al.2012] Xiong, C.; Johnson, D.; Xu, R.; and Corso, J. J. 2012. Random forests for metric learning with implicit pairwise position dependence. In Proceedings of the 18th ACM SIGKDD, 958–966. ACM.
  • [Xu et al.2013] Xu, Y.; Lin, L.; Zheng, W.-S.; and Liu, X. 2013. Human re-identification by matching compositional template with cluster sampling. In ICCV, 3152–3159. IEEE.
  • [Yang et al.2006] Yang, L.; Jin, R.; Sukthankar, R.; and Liu, Y. 2006. An efficient algorithm for local distance metric learning. In AAAI, volume 2.
  • [Zhai et al.2012] Zhai, D.; Chang, H.; Shan, S.; Chen, X.; and Gao, W. 2012. Multiview metric learning with global consistency and local smoothness. TIST 3:53:1–22.
  • [Zhang et al.2015] Zhang, R.; Lin, L.; Zhang, R.; Zuo, W.; and Zhang, L. 2015. Bit-scalable deep hashing with regularized similarity learning for image retrieval. TIP.
  • [Zhao, Ouyang, and Wang2013] Zhao, R.; Ouyang, W.; and Wang, X. 2013. Unsupervised salience learning for person re-identification. In CVPR, 3586–3593. IEEE.
  • [Zheng, Gong, and Xiang2013] Zheng, W.; Gong, S.; and Xiang, T. 2013. Associating groups of people. In BMVC. IEEE.