1 Introduction
Principal Component Pursuit (PCP) as proposed in [7, 8] and its variants e.g. [2, 25, 33, 36, 3, 5]
are the current methods of choice for recovering a lowrank subspace from a set of grossly corrupted and possibly incomplete highdimensional data. PCP employs the nuclear norm and the
norm (convex surrogates of the rank and sparsity constraints, respectively) in order to approximate the original norm regularised rank minimisation problem. In particular, under certain conditions (such as the restricted isometry property [6]), the relaxation gap is zero and rank minimisation is equivalent to nuclear norm minimisation. However, these conditions rarely hold for realworld visual data and PCP thus occasionally yields degenerate or suboptimal solutions. To alleviate this, it is advantageous for PCP to take into account of domaindependent prior knowledge [13], i.e. side information [32].The use of side information has been studied in the context of matrix completion [9, 34] and compressed sensing [17]. Recently, side information has been applied to the PCP framework in the noiseless case [21, 10]. In particular, an errorfree orthogonal column space was used to drive a PCPbased deformable image alignment algorithm [21]. More generally, [10] used both a column and a row space as side information and the algorithm had to recover the weights of their interaction. The main limitation of such methods is that they require a set of clean, noisefree data samples in order to determine the column and/or row spaces of the lowrank component. Clearly, these data are are difficult to find in practice.
In this paper, we investigate the idea of using a noisy
approximation of the lowrank component to guide PCP. Knowledge regarding the lowrank component, albeit noisy, is available in many applications. In background subtraction, we may find some frames of the video that do not contain changes and therefore may be used to accurately estimate the background. Another example concerns the problem of disentangling identity and expression components in expressive faces, where the lowrank component is roughly similar to the neutral face. Note that side information which has the same form as the source is already subject to widespread usage. Watermark detection methods require a reference image to identify ownership
[11]. Automated photo tagging explores visually similar social images [31]. Locality preserving projection can be enhanced by exploiting similar pairs of patterns [1]. Spatial and temporal correlation can improve signal recovery algorithms in compressive imaging [26]. In contentbased image retrieval, historical feedback log data can help retrieve semantically relevant images
[35]. Lowresolution images can help adapt a highresolution compressive sensing system [29]. Nearaccurate fingerprint or DNA can be used as side information to hack a biometric authentication system [14].Our contributions are summarised as follows:

A novel convex program is proposed to use side information, which is a noisy approximation of the lowrank component, within the PCP framework with a provably convergent solver.

Furthermore, we extend our proposed PCP model using side information to exploit prior knowledge regarding the column and row spaces of the lowrank component in a more general algorithmic framework.

We demonstrate the applicability and effectiveness of the proposed approaches in several applications, namely background subtraction, facial image denoising as well as face recognition and facial expression classification.

We also show that our proposed methods can mitigate the transductive constraint of RPCA. With side information, training can be performed on fewer samples and hence reducing the computational cost.
Notations Lowercase letters denote scalars and uppercase letters denote matrices, unless otherwise stated. For norms of matrix , is the Frobenius norm; is the nuclear norm; and is the maximum absolute value among all matrix entries. Moreover, represents tr() for real matrices . Additionally, is the
th largest singular value of a matrix and
is the singular value at the th percentile.2 Related work
The problem of incorporating side information in estimating lowrank components can be stated as follows. Suppose that there is a matrix with rank min() and a sparse matrix with entries of arbitrary magnitude. If we are provided with the data matrix
(1) 
and additional side information, how can we recover the lowrank component and the sparse noise accurately by taking advantage of the side information?
One the first methods for incorporating side information was proposed in the context of deformable face alignment [21]. The RAPS algorithm assumes that we have available an orthogonal column space , where , and
(2) 
A generalisation of the above was proposed as Principal Component Pursuit with Features (PCPF) in [10] where further row spaces were assumed to be available with , and
(3) 
[23, 24] incorporate structural knowledge into RPCA by adding spectral graph regularisation. Given the graph Laplacian of each data similarity graph, Robust PCA on Graphs (RPCAG) and Fast Robust PCA on Graphs (FRPCAG) add an additional tr term to the PCP objective for the lowrank component . The main drawback of the above mentioned models is that the side information needs to be accurate and noiseless, which is not trivial in practical scenarios.
3 Robust Principal Component Analysis Using Side Information
In this section, the proposed RPCA models with side information are introduced. In particular, we propose to incorporate the side information into PCP by using the trace distance of the difference between the lowrank component and the noisy estimate, which is reasonable if their difference is of low rank. However, we show empirically (Section 4) that it also works if the difference is fullrank. This may be attributed to the fact that the trace distance is a natural distance metric between two dissimilar distributions from KolmogorovSmirnov statistics [18]. Besides that, this is a generalisation of the compressed sensing with side information where the norm has been used in order to measure the distance of the target signal with prior information [17].
3.1 The PCPS model
Assuming that a noisy estimate of the lowrank component of the data is available, we propose the following model of PCP using side information (PCPS):
(4) 
where are parameters that weigh the effects of side information and noise sparsity.
The proposed PCPS can be revamped to generalise the previous attempt of PCPF by the following objective of PCPS with features (PCPSF):
(5) 
where are bilinear mappings for the recovered lowrank matrix and side information respectively. Note that the lowrank matrix is recovered from the optimal solution () to objective (5) via . If side information is not available, PCPSF reduces to PCPF by setting to zero. If the features are not present either, PCP can be restored by fixing both of them at identity. However, when only the side information is accessible, objective (5) is transformed back into PCPS.
3.2 The algorithm
If we substitute for and orthogonalise and , the optimisation problem (5) is identical to the following convex but nonsmooth problem:
(6) 
which is amenable to the multiblock alternating direction method of multipliers (ADMM).
The corresponding augmented Lagrangian of (6) is:
(7) 
where and are Lagrange multipliers and is the learning rate.
The ADMM operates by carrying out repeated cycles of updates till convergence. During each cycle, are updated serially by minimising (7) with other variables fixed. Afterwards, Lagrange multipliers are updated at the end of each iteration. Direct solutions to the single variable minimisation subproblems rely on the shrinkage and the singular value thresholding operators [7]. Let serve as the shrinkage operator, which naturally extends to matrices, , by applying it to matrix elementwise. Similarly, let be the singular value thresholding operator on real matrix , with
being the singular value decomposition (SVD) of
.Minimising (7) w.r.t. at fixed is equivalent to the following:
(8) 
where . Its solution is shown to be . Furthermore, for ,
(9) 
where , whose update rule is , and for ,
(10) 
where with a closedform solution . Finally, Lagrange multipliers are updated as usual:
(11) 
(12) 
The overall algorithm is summarised in Algorithm 1.
3.3 Complexity and convergence
Orthogonalisation of the features
via the GramSchmidt process has an operation count of
and respectively. The update in Step is the most costly step of each iteration in Algorithm 1. Specifically, the SVD required in the singular value thresholding action dominates with complexity.It has been recently established that for a 3block separable convex minimisation problem, the direct extension of the ADMM achieves global convergence with linear convergence rate if one block in the objective is substrongly monotonic [27]. In our case, it can be shown that processes such substrong monotonicity. We have also used the fast continuation technique to increase incrementally for accelerated superlinear performance. The cold start initialisation strategies for variables and Lagrange multipliers are described in [4]. Besides, we have scheduled to be updated first. As for stopping criteria, we have employed the KarushKuhnTucker (KKT) feasibility conditions. Namely, within a maximum number of iterations, when the maximum of and dwindles from a predefined threshold , the algorithm is terminated, where signifies values at the ^{th} iteration.
4 Experimental results
In this section, we illustrate the enhancement made by side information through both numerical simulations and realworld applications. First, we compare the recoverability of our proposed algorithms with stateoftheart methods for incorporating features or dictionaries, i.e. PCPF [10] and RAPS [21] on synthetic data as well as the baseline PCP [7] when there are no features available. Second, we show how powerful side information can be for the task of object segmentation in video preprocessing. Third, we demonstrate that side information is instructive in the lowdimensionality face modeling from images of different illuminations. Last, we reveal that the more accurately reconstructed expressions in the light of side information lead to better emotion classification.
For RAPS, clean subspace is used instead of the observation itself as the dictionary in LRR [15]
. PCP is solved via the inexact ALM and the heuristics for predicting the dimension of principal singular space is not adopted here due to its lack of validity on uncharted real data. We also include Partial Sum of Singular Values (PSSV)
[19] in our comparison for its stated advantage in view of the limited number of expression observations available.4.1 Parameter calibration
The process of tuning the algorithmic parameters for various models is described in the supplementary material. Although theoretical determination of and is beyond the scope of this paper, we nevertheless provide empirical guidance based on extensive experiments. for a general matrix of dimension from PCP works well for both of our proposed models. depends on the quality of the side information. When the side information is accurate, a large should be selected to capitalise upon the side information as much as possible, whereas when the side information is improper, a small should be picked to sidestep the dissonance caused by the side information. Here, we have discovered that a value of works best with synthetic data and a value of is suited for public video sequences. It is worth emphasising again that prior knowledge of the structural information about the data yields more appropriate values for and .
4.2 Phase transition on synthetic datasets
We now focus on the recoverability problem, i.e. recovering matrices of varying ranks from errors of varying sparsity. True lowrank matrices are created via , where matrices
have independent elements drawn randomly from a Gaussian distribution of mean
and variance
so is the rank of . Next, we generate error matrices , which possess nonzero elements located randomly within the matrix. We consider two types of entries for : Bernoulli and , where is the projection operator and is the support set of . thus becomes the simulated observation. For each pair, three observations are constructed. The recovery is successful if for all these three problems,(13) 
from the recovered . In addition, let be the SVD of . Feature
is formed by randomly interweaving column vectors of
with arbitrary orthonormal bases for the null space of , while permuting the expanded columns of with random orthonormal bases for the kernel of forms feature . Hence, the feasibility conditions are fulfilled: , , where is the column space operator.Entrywise corruptions. For these trials, we construct the side information by directly adding small Gaussian noise to each element of : ,
. As a result, the standard deviation of the error in each element is
of that among the elements themselves. On average, the Frobenius percent error, , is . Such side information is genuine in regard to the fact that classical PCA with accurate rank is not able to eliminate the noise [22]. For , Figures 1(a.I) and 1(a.II) plot results from PCPF, RAPS and PCPSF. On the other hand, the situation with no available features is investigated in Figures 1(a.III) and 1(a.IV) for PCP and PCPS. The frontier of PCPF has been advanced by PCPSF everywhere for both sign types. Especially at low ranks, errors with much higher density can be removed. Without features, PCPS surpasses PCP by and large with significant expansion at small sparsity for both cases.Deficient ranks. Now we first make a new matrix by retaining only the singular values from to in . Then, side information is constructed according to , aka hard thresholding. As rank increases, Frobenius percent error of decreases from to sublinearly. Figures 1(b.I) and 1(b.II) show results from PCPF, RAPS and PCPSF where is again kept at . The corresponding cases with no features are presented in Figures 1(b.III) and 1(b.IV) for PCP and PCPS. Notwithstanding the most spurious side information, PCPSF and PCPS have reclaimed the largest region unattainable by PCPF and PCP respectively for the two signs.
Distorted singular values. Here, we produce the matrix by adding Gaussian noise to singular values in : for all . Next, side information is formed by . The mean Frobenius percent error in is . With relaxed to , recoverability diagrams for PCPF, RAPS, PCPSF and PCP, PCPS are drawn in Figures (c.I), (c.II) and (c.III), (c.IV). We observe substantial growth of recoverability for PCPS over PCP across the full range of ranks. And with features, there is still omniscient gain in recoverablity for PCPSF against PCPF, which is marked at low ranks.
We remark that in unrecoverable areas, PCPS and PCPSF still obtain much smaller values of . In view of the marginal improvement of RAPS contrasted with PCPF and PCPSF, we will not consider it any longer. Results from RPCAG and PSSV are worse than PCP (see the supplementary material). FRPCAG fails to recover anything at all.
4.3 Face denoising under variable illumination
It has been previously proved that a convex Lambertian surface under distant and isotropic lighting has an underlying model that spans a 9D linear subspace. Albeit faces can be described as Lambertian, it is only approximate and harmonic planes are not real images due to negative pixels. In addition, theoretical lighting conditions cannot be realised and there are unavoidable occlusion and albedo variations. It is thus more natural to decompose facial image formation as a lowrank component for face description and a sparse component for defects. What is more, we suggest that further boost to the performance of facial characterisation can be gained by leveraging an image which faithfully represents the subject.
We consider images of a fixed pose under different illuminations from the extended Yale B database for testing. Ten subjects were randomly chosen and all 64 images were studied for each person. For singleperson experiments, observation matrices were formed by vectorising each image and the side information was chosen to be the average of all images, tiled to the same size as the observation matrix for each subject. For the multiperson experiment, both singleperson observation and side information matrices were concatenated into matrices respectively.
For PCPF and PCPSF to run, we learn the feature dictionary following an approach by Vishal et al. [20]. In a nutshell, the feature learning process can be treated as a sparse encoding problem. More specifically, we simultaneously seek a dictionary and a sparse representation such that:
(14) 
where is the number of atoms, ’s count the number of nonzero elements in each sparsity code and is the sparsity constraint factor. This can be solved by the KSVD algorithm. Here, feature is the dictionary , feature corresponds to a similar solution using the transpose of the observation matrix as input and the sparse codes are irrelevant. For implementation details, we set to , to and used iterations. Because KSVD could not converge in reasonable time for the multiperson experiment, we resorted to classical PCA applied to the observation matrix to obtain features of dimension .
As a visual illustration, two challenging cases are exhibited in Figure 2 (PSSV, RPCAG, FRPCAG do not improve upon PCP and are shown in the supplementary material). For subject 2, it is clearly evident that PCPS and PCPSF outperform the best existing methods through the complete elimination of acquisition faults. More surprisingly, PCPSF even manages to restore the flash in the pupils that is not present in the side information. For subject 33, PCPS indubitably reconstructs a more vivid left eye than that from PCP which is only discernible. With that said, PCPSF still prevails by uncovering more shadows, especially around the medial canthus of the left eye, and revealing a more distinct crease in the upper eyelid as well a more translucent iris. We also notice that results from the singleperson experiment outdo their counterparts from the multiperson experiment. Thence, we will focus on a single subject alone.
To quantitatively verify the improvement made by our proposed approaches, we examine the structural information contained within the deionised eigenfaces. Singular values of the recovered lowrank matrices from all algorithms are plotted in Figure 3. Singular values decease most sharply for PCPSF followed by PCPS. By the theoretical limit, they are orders of magnitude smaller than those values from other methods. This validates our proposed approaches.
We further unmask the strength of PCPS by considering the stringent side information made of the average of other subjects. Surprisingly, PCPS still manages to remove the noise recovering an authentic image (see Figure 2 (d)).
4.4 Background subtraction from surveillance video
In automated video analytics, object detection is instrumental in object tracking, activity recognition and behaviour understanding. Practical applications include surveillance, traffic control, robotic operation, etc., where foreground objects can be people, vehicles, products and so forth. Background subtraction segments moving objects by calculating the pixelwise difference between each video frame and the background. For a static camera, the background is almost static, while the foreground objects are mostly moving. Consequently, a decomposition into a lowrank component for the background and a sparse component for foreground objects is a valid model for such dynamics. Indeed, if the only change in the background is illumination, then the matrix representation of vectorised backgrounds has a rank of . It has been demonstrated that PCP is quite effective for such a lowrank matrix analysis problem [7]. Nevertheless, through the application of our proposed algorithm to such a backgroundforeground separation scenario, we show that useful side information can help achieve better background restoration.
One video sequence from the PETS 2006 dataset and one from the I2R dataset were utilised for evaluation. Each consists of scenes at a hall where people walk intermittently. 200 consecutive frames of resolution grayscale images were stacked by columns into a observation matrix from the first video and 200 frames of images from the second video were stacked into another observation matrix. Two side information arrays comprised columns that are copies of a vectorised photo which contains an empty hallway. To commence object detection, PCP and PCPS were first run to extract the backgrounds. Then objects were recovered by calculating the absolute values of the difference between the original frame and the estimated background. Since parameters for dictionary learning need exhaustive search, we will not be comparing PCPF and PCPSF for what follows.
We quantitatively compare the performance of the competing methods according to the weighted Fmeasure [16] against manually annotated bounding boxes provided as the ground truth. The resulting scores for each frame are presented in Figure 5. From the consistently higher precision statistics, the merit of PCPS over PCP is confirmed. For qualitative reference, representative images of the recovered background and foreground from all methods are listed in Figure 4 (For space reasons, we have only included the most noticeable sector. See the supplementary material for whole images.). PCP and its variants only partially detect infrequent moving objects, people who stop moving for extended periods of time, leaving ghost artifacts in the background. In contrast, PCPS segments a fairly sharp silhouette of slowly moving objects to produce a much cleaner background, promoting its novelty.
To further unravel of the robustness of our propositions, shortened videos from PETS and Airport consisting of 60 frames are analysed via PCPS. Figures 4 (c,d) & (o,p) show that PCPS with less input can achieve comparative or better results than PCP with more input. This suggests that the transductive constraint of RPCA no longer applies because with the help of side information we can run PCPS on fewer frames rather than the entire collection every time new observation arrives. The speedups for PETS and Airport are and respectively.
4.5 Face and facial expression recognition
Recent research has established that an expressive face can be treated as a neutral face plus a sparse expression component [28], which is identityindependent due to its constituent local nonrigid motions, i.e. action units. This is central to computer vision as it enables human emotion classification from such visual cues. We will demonstrate how the accurate reconstruction of facial expressions guided by side information ameliorates classification analysis.
To begin with, evaluation was effected on the CMU MultiPIE dataset. Aligned and cropped images of frontal pose and normal lighting from 54 subjects were used. We batchprocessed each subject forming a observation matrix to extract expressions: Neutral, Smile, Surprise, Disgust, Scream and Squint. For each subject, side information was offered by a sextet of neutral face repetitions. Archetypal expressions recovered by PCP, PCPS, PSSV, RPCAG are laid out in Figure 6 (the restricted number of expressions disallows FRPCAG). It is noteworthy that local appearance changes separated by PCPS are the most salient which paves the way for better classification. We avail ourselves of the multiclass RBFkernel SVM and the SRC [30] to map expressions to emotions. 9fold crossvalidation results are reported in Table 1. PCPS leads PCP by a fair margin with PSSV, RPCAG underperforming PCP.
Algorithm  PCP  PSSV  PCPS  RPCAG 

Nonlinear SVM  78.40  74.69  79.94  77.16 
SRC  79.01  74.38  82.72  79.01 
Lastly, the CK+ dataset was incorporated to assess the joint face and expression recognition capabilities of various algorithms. Each test image is sparsely coded via a dictionary of both identities and universal expressions (Anger, Disgust, Fear, Happiness, Sadness and Surprise). The least resulting reconstruction residual thereupon determines its identity or expression. We refer readers to [12] for the exact problem setup and implementation details. Table 2 collects the computed recognition rates. Altough RPCAG and FRPCAG are superior than PCP as expected, PCPS performs distinctly better than all others.
Algorithm  PCP  PSSV  PCPS  RPCAG  FRPCAG 

Identity  87.35  87.05  95.23  89.77  90.98 
Expression  49.24  45.30  67.50  58.26  57.73 
5 Conclusion
In this paper, we have, for the first time, assimilated side information of the same format as observation into the framework of Robust Principal Component Analysis based on trace norms. Existing extensions of subspace features have also been successfully amalgamated in a convex fashion. Extensive experiments have shown that our algorithms not only perform better where Robust PCA is effective but also remain potent when Robust PCA fails. Directions for future research include generalising to the tensor case and to components of multiple scales.
Appendix A Parameter calibration
In order to tune the algorithmic parameters, we first conduct a benchmark experiment as follows: a lowrank matrix is generated from , where have entries from a distribution; a sparse matrix is generated by randomly setting entries to zero with others taking values of
with equal probability.
If is set as the leftsingular vectors of and is set as the rightsingular vectors of , then a scaling ratio , a tolerance threshold and a maximum step size to avoid illconditioning can bring PCP, RAPS, PCPF to convergence with a recovered of rank , a recovered of sparsity and an accuracy on the order of . Hereafter, we will adopt these parameter settings for PCP, RAPS, PCPF and will apply them to PCPS and PCPSF as well. PSSV also uses these parameter settings as done similarly in [19].
For RPCAG and FRPCAG, the graphs are built using nearest neighbors. Using Euclidean distances, each sample is connected to 10 nearest neighbors with weight , where is the Euclidean distance between the two samples and is the average of . Weight between unconnected samples is set to 0. Having obtained such weight matrix , we can calculate the normalised graph Laplacian , where is the diagonal degree matrix. The tolerance threshold for RPCAG and FRPCAG are all set to for reasons of consistency. We choose for a general matrix of dimension as suggested in [23, 24]. For simulation experiments, in RPCAG is given by the minimiser (at ) of on the benchmark problem (Figure 7). And for realworld datasets, is set to 10 following [23]. For FRPCAG, we take which is searched over on the benchmark problem (Figure 8). The resulting minimiser (at ) of is used in both simulation and realworld experiments.
To find and in PCPS, a parameter sweep in the space for perfect side information () is shown in Figure 9 (a) and for observation as side information () in Figure 9 (b) to impart a lower bound and a upper bound respectively. It can be easily seen that from PCP works well in both cases. Conversely, depends on the quality of the side information. At , the minimiser of occurs at for noisy side information. This value of together with is used in simulation experiments for both PCPS and PCPSF. For public video sequences, increasing the value of to can produce visual results that are noticeable to the naked eye.
Appendix B Simulation Results
Appendix C Realworld applications
c.1 Data sources
The datasets used herein are listed below:
The Extended Yale Face Database B: http://vision.ucsd.edu/~iskwak/ExtYaleDatabase/ExtYaleB.html.
Performance Evaluation of Tracking and Surveillance Workshop 2006: http://www.cvg.reading.ac.uk/PETS2006/data.html.
I2R Dataset: http://perception.i2r.astar.edu.sg/bk_model/bk_index.html.
The CMU MultiPIE Face Database: http://www.cs.cmu.edu/afs/cs/project/PIE/MultiPie/MultiPie/Home.html.
The Extended CohnKande Dataset (CK+): http://www.consortium.ri.cmu.edu/ckagree/.
c.2 Face denoising
Illustration of face denoising ability of PSSV, RPCAG, FRPCAG is presented in Figure 12. The average running times of different algorithms for a single subject and multiple subjects are summarised in Table 3 ^{1}^{1}1All experiments were performed on a 3.60GHz quadcore computer with 16GB RAM running MATLAB R2016a..
Algorithm  Time  

Single Subject  Multiple Subjects  
KSVD (X)  9 min  — 
KSVD (Y)  78 min  — 
PCP  12s  5 min 
PCPS  27s  12 min 
PCPF  16s  9 min 
PCPSF  19s  8 min 
PSSV  13s  5 min 
kNN (X)  7s  4 min 
kNN (Y)  1s  8s 
RPCAG  2min  17 min 
FRPCAG  8s  1 min 
c.3 Background Subtraction
Recovered images of the background and the foreground from all methods are listed in Figure 13 for Airport and Figure 14 for PETS. The running times of different algorithms for Airport and PETS are summarised in Table 4.
Algorithm  Time  

Airport  PETS  
PCP  52s  17 min 
PCPS  2 min  36 min 
PSSV  51s  17 min 
kNN (X)  52s  2h 
kNN (Y)  1s  24s 
RPCAG  7 min  3h 
FRPCAG  11s  34s 
PCP (60 frames)  52s  3 min 
PCPS (60 frames)  20s  7 min 
Appendix D Derivations
Here we give deviations of the various equivalent subproblems for the algorithm quoted in the text:
Appendix E Further comments
One might suggest that a potentially better and more direct approach in using the side information is to subtract the side information. That is, do RPCA on , where is the data and is the noisy side information, to obtain with .
We argue that this is not correct for the following reasons:

The rank of is no smaller than , which does not make the problem any simpler than the original one.

When is merged into , the additional information provided by is lost and the features can on longer be applied.

When includes fullrank noise on , is not lowrank anymore. This violates the assumption of RPCA.
To verify our claim, we perform the Airport experiment again, but with different side information than that used in the paper. We collect 200 different frames of relatively clean backgrounds and stack them into the side information . Comparison of the suggestion with PCPS and PCP is shown in Figure 15, 16 and 17. It is clearly visible that the lowrank structure cannot be recovered by the suggestion and spurious noises are introduced in the segmentation, whereas PCPS works impeccably segmenting accurately the foreground moving objects leaving a clean background.
References

[1]
S. An, W. Liu, and S. Venkatesh.
Exploiting side information in locality preserving projection.
IEEE Conference on Computer Vision and Pattern Recognition
, 2008. 
[2]
A. Aravkin, S. Becker, V. Cevher, and P. Olsen.
A variational approach to stable principal component pursuit.
Conference on Uncertainty in Artificial Intelligence
, pages 32–41, 2014.  [3] B. Bao, G. Liu, C. Xu, and S. Yan. Inductive robust principal component analysis. IEEE Transactions on Image Processing, 21(8):3794 – 3800, 2012.
 [4] S. Boyd, N. Parikh, E. Chu, B. Peleato, and J. Eckstein. Distributed optimization and statistical learning via the alternating direction method of multipliers. Foundations and Trends in Machine Learning, 3(1):1–122, 2011.
 [5] R. Cabral, F. De la Torre, J. Costeira, and A. Bernardino. Unifying nuclear norm and bilinear factorization approaches for lowrank matrix decomposition. International Conference on Computer Vision, 2013.
 [6] E. Candès. The restricted isometry property and its implications for compressed sensing. Comptes Rendus Mathematique, 346(9):589–592, 2008.
 [7] E. Candès, X. Li, Y. Ma, and J. Wright. Robust principal component analysis? Journal of the ACM, 58(3):11:1–11:37, 2011.
 [8] V. Chandrasekaran, S. Sanghavi, P. Parrilo, and A. Willsky. Ranksparsity incoherence for matrix decomposition. SIAM Journal on Optimization, 21(2):572––596, 2011.
 [9] K. Chiang, C. Hsieh, and I. Dhillon. Matrix completion with noisy side information. Advances in Neural Information Processing Systems, 2015.
 [10] K. Chiang, C. Hsieh, and I. Dhillon. Robust principal component analysis with side information. International Conference on Machine Learning, 2016.
 [11] I. Cox, J. Kilian, F. Leighton, and T. Shamoon. Secure spread spectrum watermarking for multimedia. IEEE Transactions on Image Processing, 6(12):1673–1687, 1997.
 [12] C. Georgakis, Y. Panagakis, and M. Pantic. Discriminant incoherent component analysis. IEEE Transactions on Image Processing, 25(5):2021–2034, 2016.
 [13] J. Jiao, T. Courtade, K. Venkat, and T. Weissman. Justification of logarithmic loss via the benefit of side information. IEEE Transactions on Information Theory, 61(10):5357–5365, 2015.
 [14] W. Kang, D. Cao, and N. Liu. Deception with side information in biometric authentication systems. IEEE Transactions on Information Theory, 61(3):1344–1350, 2015.
 [15] G. Liu, Z. Lin, S. Yan, J. Sun, Y. Yu, and Y. Ma. Robust recovery of subspace structures by lowrank representation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(1):171–184, 2013.
 [16] R. Margolin, L. ZelnikManor, and A. Tal. How to evaluate foreground maps? IEEE Conference on Computer Vision and Pattern Recognition, 2014.
 [17] J. Mota, N. Deligiannis, and M. Rodrigues. Compressed sensing with side information: Geometrical interpretation and performance bounds. IEEE Global Conference on Signal and Information Processing, 2014.
 [18] M. Nielsen and I. Chuang. Quantum computation and quantum information. Cambridge University Press, 2010.
 [19] T. Oh, Y. Tai, J. Bazin, H. Kim, and I. Kweon. Partial sum minimization of singular values in robust pca: Algorithm and applications. IEEE Transactions on Pattern Analysis and Machine Intelligence, 38(4):744–758, 2016.
 [20] V. Patel, T. Wu, S. Biswas, P. Phillips, and R. Chellappa. Dictionarybased face recognition under variable lighting and pose. IEEE Transactions on Information Forensics and Security, 7(3):954–965, 2012.
 [21] C. Sagonas, Y. Panagakis, S. Zafeiriou, and M. Pantic. RAPS: Robust and efficient automatic construction of personspecific deformable models. IEEE Conference on Computer Vision and Pattern Recognition, 2014.

[22]
A. Shabalin and A. Nobel.
Reconstruction of a lowrank matrix in the presence of gaussian
noise.
Journal of Multivariate Analysis
, 118:67–76, 2013.  [23] N. Shahid, V. Kalofolias, X. Bresson, M. Bronstein, and P. Vandergheynst. Robust principal component analysis on graphs. International Conference on Computer Vision, 2015.
 [24] N. Shahid, N. Perraudin, V. Kalofolias, G. Puy, and P. Vandergheynst. Fast robust PCA on graphs. IEEE Journal of Selected Topics in Signal Processing, 10(4):740–756, 2016.
 [25] F. Shang, Y. Liu, J. Cheng, and H. Cheng. Robust principal component analysis with missing data. ACM International Conference on Information and Knowledge Management, pages 1149–1158, 2014.
 [26] V. Stanković, L. Stanković, and S. Cheng. Compressive image sampling with side information. International Conference on Image Processing, 2009.
 [27] H. Sun, J. Wang, and T. Deng. On the global and linear convergence of direct extension of admm for 3block separable convex minimization models. Journal of Inequalities and Applications, (227):227, 2016.
 [28] S. Taheri, V. Patel, and R. Chellappa. Componentbased recognition of facesand facial expressions. IEEE Transactions on Affective Computing, 4:360–371, 2013.
 [29] G. Warnell, S. Bhattacharya, R. Chellappa, and T. Başar. Adaptiverate compressive sensing using side information. IEEE Transactions on Image Processing, 24(11):3846–3857, 2015.
 [30] J. Wright, A. Yang, A. Ganesh, S. Sastry, and Y. Ma. Robust face recognition via sparse representation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 31(2):210–227, 2009.
 [31] L. Wu, S. Hoi, R. Jin, J. Zhu, and N. Yu. Distance metric learning from uncertain side information with application to automated photo tagging. ACM Multimedia Conference, 2009.
 [32] A. Wyner and J. Ziv. The ratedistortion function for source coding with side information at the decoder. IEEE Transactions on Information Theory, 22(1):1–10, 1976.

[33]
H. Xu, C. Caramanis, and S. Sanghavi.
Robust pca via outlier pursuit.
IEEE Transactions on Information Theory, 58(5):3047–3064, 2012.  [34] M. Xu, J. R, and Z. Zhou. Speedup matrix completion with side information: Application to multilabel learning. Advances in Neural Information Processing Systems, 2013.
 [35] L. Zhang, L. Wang, and W. Lin. Conjunctive patches subspace learning with side information for collaborative image retrieval. IEEE Transactions on Image Processing, 21(8):3707–3720, 2012.
 [36] Z. Zhou, X. Li, J. Wright, E. Candès, and Y. Ma. Stable principal component pursuit. IEEE International Symposium on Information Theory, 2010.
Comments
There are no comments yet.