Performance Guarantees for ReProCS Correlated Low-Rank Matrix Entries Case

Similar documents
Recursive Sparse Recovery in Large but Structured Noise - Part 2

Online (Recursive) Robust Principal Components Analysis

SEPARATING SPARSE AND LOW-DIMENSIONAL SIGNAL SEQUENCES FROM TIME-VARYING UNDERSAMPLED PROJECTIONS OF THEIR SUMS

Nearly Optimal Robust Subspace Tracking

Correlated-PCA: Principal Components Analysis when Data and Noise are Correlated

Robust Principal Component Analysis

Recursive Sparse Recovery in Large but Correlated Noise

PCA and Robust PCA for Modern Datasets: Theory, Algorithms, and Applications

Recovery of Low-Rank Plus Compressed Sparse Matrices with Application to Unveiling Traffic Anomalies

New Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit

Stability of LS-CS-residual and modified-cs for sparse signal sequence reconstruction

Exact Reconstruction Conditions and Error Bounds for Regularized Modified Basis Pursuit (Reg-Modified-BP)

Stability of Modified-CS over Time for recursive causal sparse reconstruction

Conditions for Robust Principal Component Analysis

Analysis of Robust PCA via Local Incoherence

Exact Low-rank Matrix Recovery via Nonconvex M p -Minimization

A new method on deterministic construction of the measurement matrix in compressed sensing

of Orthogonal Matching Pursuit

Linear Subspace Models

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

Robust PCA. CS5240 Theoretical Foundations in Multimedia. Leow Wee Kheng

COMPRESSED SENSING IN PYTHON

arxiv: v2 [cs.lg] 20 Mar 2017

Structured matrix factorizations. Example: Eigenfaces

Robust Component Analysis via HQ Minimization

Lecture: Face Recognition and Feature Reduction

Non-convex Robust PCA: Provable Bounds

CS 229r: Algorithms for Big Data Fall Lecture 19 Nov 5

Applications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices

On the l 1 -Norm Invariant Convex k-sparse Decomposition of Signals

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Observability of a Linear System Under Sparsity Constraints

CoSaMP. Iterative signal recovery from incomplete and inaccurate samples. Joel A. Tropp

Compressed Sensing and Robust Recovery of Low Rank Matrices

Compressed Sensing and Sparse Recovery

A New Estimate of Restricted Isometry Constants for Sparse Solutions

Sparse Solutions of an Undetermined Linear System

Robust PCA via Outlier Pursuit

arxiv: v5 [math.na] 16 Nov 2017

Recovering overcomplete sparse representations from structured sensing

Design of Projection Matrix for Compressive Sensing by Nonsmooth Optimization

Compressive Sensing and Beyond

Exact Recoverability of Robust PCA via Outlier Pursuit with Tight Recovery Bounds

Introduction How it works Theory behind Compressed Sensing. Compressed Sensing. Huichao Xue. CS3750 Fall 2011

GROUP-SPARSE SUBSPACE CLUSTERING WITH MISSING DATA

Compressed sensing. Or: the equation Ax = b, revisited. Terence Tao. Mahler Lecture Series. University of California, Los Angeles

Strengthened Sobolev inequalities for a random subspace of functions

Solution-recovery in l 1 -norm for non-square linear systems: deterministic conditions and open questions

Sketching for Large-Scale Learning of Mixture Models

Matrix-Tensor and Deep Learning in High Dimensional Data Analysis

Lecture: Face Recognition and Feature Reduction

Particle Filtered Modified-CS (PaFiMoCS) for tracking signal sequences

Robust Principal Component Analysis Based on Low-Rank and Block-Sparse Matrix Decomposition

Joint Direction-of-Arrival and Order Estimation in Compressed Sensing using Angles between Subspaces

Orthogonal Matching Pursuit for Sparse Signal Recovery With Noise

Greedy Signal Recovery and Uniform Uncertainty Principles

GoDec: Randomized Low-rank & Sparse Matrix Decomposition in Noisy Case

Sparse Sensing in Colocated MIMO Radar: A Matrix Completion Approach

Chapter 3 Transformations

SUBSPACE CLUSTERING WITH DENSE REPRESENTATIONS. Eva L. Dyer, Christoph Studer, Richard G. Baraniuk

The convex algebraic geometry of rank minimization

A NEW FRAMEWORK FOR DESIGNING INCOHERENT SPARSIFYING DICTIONARIES

A Cross-Associative Neural Network for SVD of Nonsquared Data Matrix in Signal Processing

Uniform Uncertainty Principle and signal recovery via Regularized Orthogonal Matching Pursuit

GREEDY SIGNAL RECOVERY REVIEW

Thresholds for the Recovery of Sparse Solutions via L1 Minimization

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

PCA with random noise. Van Ha Vu. Department of Mathematics Yale University

Solving Corrupted Quadratic Equations, Provably

1 Principal Components Analysis

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Robust PCA via Outlier Pursuit

Dense Error Correction for Low-Rank Matrices via Principal Component Pursuit

arxiv: v1 [cs.it] 26 Oct 2018

CoSaMP: Greedy Signal Recovery and Uniform Uncertainty Principles

Uniqueness Conditions for A Class of l 0 -Minimization Problems

Latent Variable Graphical Model Selection Via Convex Optimization

Linear dimensionality reduction for data analysis

Tractable Upper Bounds on the Restricted Isometry Constant

CSC 576: Variants of Sparse Learning

First Efficient Convergence for Streaming k-pca: a Global, Gap-Free, and Near-Optimal Rate

Combining Sparsity with Physically-Meaningful Constraints in Sparse Parameter Estimation

L26: Advanced dimensionality reduction

Sparse Legendre expansions via l 1 minimization

Introduction to Machine Learning

Near Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing

c Springer, Reprinted with permission.

Recovery of Sparse Signals from Noisy Measurements Using an l p -Regularized Least-Squares Algorithm

Stability (over time) of regularized modified CS (noisy) for recursive causal sparse reconstruction

Signal Recovery From Incomplete and Inaccurate Measurements via Regularized Orthogonal Matching Pursuit

INDUSTRIAL MATHEMATICS INSTITUTE. B.S. Kashin and V.N. Temlyakov. IMI Preprint Series. Department of Mathematics University of South Carolina

MODIFIED DISTRIBUTED ITERATIVE HARD THRESHOLDING

SUBSPACE CLUSTERING WITH DENSE REPRESENTATIONS. Eva L. Dyer, Christoph Studer, Richard G. Baraniuk. ECE Department, Rice University, Houston, TX

Low Rank Matrix Completion Formulation and Algorithm

SPARSE signal representations have gained popularity in recent

Online PCA for Contaminated Data

Particle Filtering for Large-Dimensional State Spaces With Multimodal Observation Likelihoods Namrata Vaswani

Principal Component Analysis -- PCA (also called Karhunen-Loeve transformation)

Fast Angular Synchronization for Phase Retrieval via Incomplete Information

Transcription:

Performance Guarantees for ReProCS Correlated Low-Rank Matrix Entries Case Jinchun Zhan Dep. of Electrical & Computer Eng. Iowa State University, Ames, Iowa, USA Email: jzhan@iastate.edu Namrata Vaswani Dep. of Electrical & Computer Eng. Iowa State University, Ames, Iowa, USA Email: namrata@iastate.edu Abstract To be considered for an IEEE Jack Keil Wolf ISIT Student Paper Award. Online or recursive robust PCA can be posed as a problem of recovering a sparse vector, S t, and a dense vector, L t, which lies in a slowly changing low-dimensional subspace, from M t := S t + L t on-the-fly as new data comes in. For initialization, it is assumed that an accurate knowledge of the subspace in which L 0 lies is available. In recent works, Qiu et al proposed and analyzed a novel solution to this problem called recursive projected compressed sensing or ReProCS. In this work, we relax one limiting assumption of Qiu et al s result. Their work required that the L t s be mutually independent over time. However this is not a practical assumption, e.g., in the video application, L t is the background image sequence and one would expect it to be correlated over time. In this work we relax this and allow the L t s to follow an autoregressive model. We are able to show that under mild assumptions and under a denseness assumption on the unestimated part of the changed subspace, with high probability (w.h.p.), ReProCS can exactly recover the support set of S t at all times; the reconstruction errors of both S t and L t are upper bounded by a time invariant and small value; and the subspace recovery error decays to a small value within a finite delay of a subspace change time. Because the last assumption depends on an algorithm estimate, this result cannot be interpreted as a correctness result but only a useful step towards it. I. INTRODUCTION Principal Components Analysis (PCA) is a widely used dimension reduction technique that finds a small number of orthogonal basis vectors, called principal components (PCs), along which most of the variability of the dataset lies. Often, for time series data, the PCs space changes gradually over time. Updating it on-the-fly (recursively) in the presence of outliers, as more data comes in is referred to as online or recursive robust PCA [], [2]. As noted in earlier work, an outlier is well modeled as a sparse vector. With this, as will be evident, this problem can also be interpreted as one of recursive sparse recovery in large but structured (lowdimensional) noise. A key application where the robust PCA problem occurs is in video analysis where the goal is to separate a slowly changing background from moving foreground objects [3], [4]. If we stack each frame as a column vector, the background is well modeled as being dense and lying in a low dimensional subspace that may gradually change over time, while the moving foreground objects constitute the sparse outliers [4]. Other applications include detection of brain activation patterns from functional MRI sequences or detection of anomalous behavior in dynamic networks [5].There has been a large amount of earlier work on robust PCA, e.g. see [3]. In recent works [4], [6], the batch robust PCA problem has been posed as one of separating a low rank matrix, L t := [L 0, L,..., L t ], from a sparse matrix, S t := [S 0, S,..., S t ], using the measurement matrix, M t := [M 0, M,..., M t ] = L t + S t. A novel convex optimization solution called principal components pursuit (PCP) has been proposed and shown to achieve exact recovery under mild assumptions. Since then, the batch robust PCA problem, or what is now also often called the sparse+lowrank recovery problem, has been studied extensively. We do not discuss all the works here due to limited space. Online or recursive robust PCA can be posed as a problem of recovering a sparse vector, S t, and a dense vector, L t, which lies in a slowly changing low-dimensional subspace, from M t := S t + L t on-the-fly as new data comes in. For initialization, it is assumed that an accurate knowledge of the subspace in which L 0 lies is available. In recent works, Qiu et al proposed and analyzed a novel solution to this problem called ReProCS [7], [8], [9]. Contribution: In this work, we relax one limiting assumption of Qiu et al s result. Their work required that the L t s be mutually independent over time. However this is not a practical assumption, e.g., in the video application, L t is the background image sequence and one would expect it to be correlated over time. In this work we relax this and allow the L t s to follow an autoregressive (AR) model. We are able to show that, under mild assumptions and a denseness assumption on the currently unestimated subspace, with high probability (w.h.p.), ReProCS (with subspace change model known) can exactly recover the support set of S t at all times; the reconstruction errors of both S t and L t are upper bounded by a time invariant and small value; and the subspace recovery error decays to a small value within a finite delay of a subspace change time. The last assumption depends on an algorithm estimate and hence this result also cannot be interpreted as a correctness result but only a useful step towards it. To the best of our knowledge, the result of Qiu et al and this follow-up work are among the first results for any recursive (online) robust PCA approach, and also for recursive sparse recovery in large but structured (low-dimensional) noise. Other very recent work on algorithms for recursive / online

robust PCA includes [0], [], [2], [3], [4], [5]. In [3], [4], two online algorithms for robust PCA (that do not model the outlier as a sparse vector but only as a vector that is far from the data subspace) have been shown to approximate the batch solution and do so only asymptotically. A. Notation For a set T {, 2,..., n}, we use T to denote its cardinality. We use T c to denote its complement w.r.t. {, 2,... n}, i.e. T c := {i {, 2,... n} : i / T }. We use the interval notation, [t, t 2 ], to denote the set of all integers between and including t to t 2, i.e. [t, t 2 ] := {t, t +,..., t 2 }. For a vector v, v i denotes the ith entry of v and v T denotes a vector consisting of the entries of v indexed by T. We use v p to denote the l p norm of v. The support of v, supp(v), is the set of indices at which v is nonzero, supp(v) := {i : v i 0}. We say that v is s-sparse if supp(v) s. For a matrix B, B denotes its transpose, and B its pseudoinverse.we use B 2 := max x 0 Bx 2 / x 2 to denote the induced 2-norm of the matrix. Also, B is the nuclear norm (sum of singular values) and B max denotes the maximum over the absolute values of all its entries. For a Hermitian matrix, B, we use the notation B EV = D UΛU to denote the eigenvalue decomposition of B. Here U is an orthonormal matrix and Λ is a diagonal matrix with entries arranged in decreasing order. We use λ max (B) and λ min (B) denote its maximum and minimum eigenvalues. We use I to denote an identity matrix of appropriate size. For an index set T and a matrix B, B T is the sub-matrix of B containing columns with indices in the set T. For a tall matrix P, span(p ) denotes the subspace spanned by the column vectors of P. The notation [.] denotes an empty matrix. Definition.: We refer to a tall matrix P as a basis matrix if it satisfies P P = I. Definition.2: The s-restricted isometry constant (RIC) [5], δ s, for an n m matrix Ψ is the smallest real number satisfying ( δ s ) x 2 2 Ψ T x 2 2 (+δ s ) x 2 2 for all sets T with T s and all real vectors x of length T. It is easy to see that max T : T s (Ψ T Ψ T ) 2 δ s (Ψ). Definition.3: Define the time interval I j,k := [t j + (k )α, t j + kα ] for k =,... K and I j,k+ := [t j + Kα, t j+ ]. II. PROBLEM DEFINITION The measurement vector at time t, M t, satisfies M t := L t + S t () where S t is a sparse vector and L t is a dense vector that satisfies the model given below. Denote by P 0 a basis matrix for L ttrain = [L 0, L,, L ttrain ], i.e., span(p 0 ) = span(l ttrain ). We are given an accurate enough estimate ˆP 0 for P 0, i.e., (I ˆP 0 ˆP 0 )P 0 2 is small. The goal is ) to estimate both S t and L t at each time t > t train, and 2) to estimate span(l t ) every so often. A. Signal Model Signal Model 2. (Model on L t ): ) We assume that L t = P (t) a t with P (t) = P j for all t j t < t j+, j = 0,, 2 J, where P j is an n r j basis matrix with r j min(n, (t j+ t j )). At the change times, t j, P j changes as P j = [P j P j,new ]. Here, P j,new is an n c j,new basis matrix with P j,new P j = 0. Thus r j = r j + c j,new. We let t 0 = 0 and t J+ equal the sequence length or t J+ =. 2) The vector of coefficients, a t := P (t) L t, satisfies the following autoregressive model a t = ba t + ν t where b < is a scalar, E[ν t ] = 0, ν t s are mutually independent over time t and the entries of any ν t are pairwise uncorrelated, i.e. E[(ν t ) i (ν t ) j ] = 0 when i j. Definition 2.2: Define the covariance matrices of ν t and a t to be the diagonal matrices Λ ν,t := Cov[ν t ] = E(ν t ν t) and Λ a,t := Cov[a t ] = E(a t a t). Then clearly, Λ a,t = b 2 Λ a,t + Λ ν,t (2) Also, for t j t < t j+, a t is an r j length vector which can be split as a t = P at, j L t = a t,new where a t, := P j L t is an r j length vector and a t,new := P j,new L t is a c j,new length vector. Thus, for this interval, L t can be rewritten as at, L t = [P j P j,new ] = P a j a t, + P j,new a t,new t,new Also, Λ a,t can be split as (Λa,t ) Λ a,t = 0 0 (Λ a,t ) new where (Λ a,t ) = Cov[a t, ] and (Λ a,t ) new = Cov[a t,new ] are diagonal matrices. Assumption 2.3: Assume that L t satisfies Signal Model 2. with ) 0 c j,new c for all j (thus r j r max := r 0 + Jc) 2) ν t ( b)γ 3) 4) max j max ν t,new ( b)γ new,k, (3) t I j,k where γ new,k = min(v k γ new, γ ) with a v > ( b 2 )λ λ min (Λ ν,t ) λ max (Λ ν,t ) ( b 2 )λ +, and λ λ min (Λ a,0 ) λ max (Λ a,0 ) ( b 2 )λ +, where 0 < λ < λ + < 5) ( b 2 )λ new λ min ((Λ ν,t ) new ) λ max ((Λ ν,t ) new ) ( b 2 )λ + new and

λ new λ min ((Λ a,tj ) new ) λ max (Λ a,tj ) new ) ( b 2 )λ + new With the above assumptions, clearly, the eigenvalues of Λ a,t lie between λ and λ + and those of Λ a,t,new lie between λ new and λ + new. Thus the condition numbers of any Λ a,t and of Λ a,t,new are bounded by f := λ+ λ and g := λ+ new λ new respectively. Define the following quantities for S t. Definition 2.4: Let T t := {i : (S t ) i 0} denote the support of S t. Define S min := min min (S t ) i, and s := max T t t>t train i T t t III. THE REPROCS ALGORITHM AND ITS PERFORMANCE A. Recursive Projected CS GUARANTEE We summarize the ReProCS algorithm in Algorithm 2 [8]. This calls the projection PCA algorithm in the subspace update step. Given a data matrix D, a basis matrix P and an integer r, projection-pca (proj-pca) applies PCA on D proj := (I P P )D, i.e., it computes the top r eigenvectors (the eigenvectors with the largest r eigenvalues) of D proj D proj. Here is the number of column vectors in D. This is summarized in Algorithm. If P = [.], then projection-pca reduces to standard PCA, i.e. it computes the top r eigenvectors of DD. Algorithm projection-pca: Q proj-pca(d, P, r) ) Projection: compute D proj (I P P )D EV D 2) PCA: compute D proj D proj = Λ 0 Q Q Q where Q is an n r 0 Λ Q basis matrix and is the number of columns in D. B. Main Result The following definition is needed for Theorem 3.2. Definition 3.: ) Let r := r 0 + (J )c. 2) Define η = max{ cγ2 3) Define K(ζ) := λ, cγ2 new,k + λ + new log(0.85cζ) log 0.6, k =, 2,, K}. 4) Define ξ 0 (ζ) := cγ new + ζ( r + c) 5) With K = K(ζ), define α add := (log 6KJ + log n) 8 92 2 min(.2 4K γnew, 4 γ ) 4 ζ 2 (λ ) 2 Theorem 3.2: Consider Algorithm 2. Assume that Assumption 2. holds with b 0.4. Assume also that the initial subspace estimate is accurate enough, i.e. (I ˆP 0 ˆP 0 )P 0 r 0 ζ, with ( ) 0 4.5 0 4 ζ min r 2, r 2, f r 3 γ 2 where f := λ+ λ Algorithm 2 Recursive Projected CS (ReProCS) Parameters: algorithm parameters: ξ, ω, α, K, model parameters: t j, r 0, c (set as in Theorem 3.2 or as in [8, Section IX-B] when the model is not known) Input: M t, Output: Ŝ t, ˆL t, ˆP(t) Initialization: Compute ˆP0 proj- PCA([L, L 2,, L ttrain ], [.], r 0 ) and set ˆP(t) ˆP 0. Let j, k. For t > t train, do the following: ) Estimate T t and S t via Projected CS: a) Nullify most of L t : compute Φ (t) I ˆP (t ) ˆP (t ), compute y t Φ (t) M t b) Sparse Recovery: compute Ŝt,cs as the solution of min x x s.t. y t Φ (t) x 2 ξ c) Support Estimate: compute ˆT t = {i : (Ŝt,cs) i > ω} d) LS Estimate of S t : compute (Ŝt) ˆTt = ((Φ t ) ˆTt ) y t, (Ŝt) ˆT c = 0 t 2) Estimate L t : ˆL t = M t Ŝt. 3) Update ˆP (t) : K Projection PCA steps (Figure ). a) If t = t j + kα, i) ˆPj,new,k proj- PCA ( [ˆLtj +(k )α,..., ˆL ] tj+kα, ˆP j, c ). ii) set ˆP (t) [ ˆP j ˆPj,new,k ]; increment k k +. Else i) set ˆP(t) ˆP (t ). b) If t = t j +Kα, then set ˆP j [ ˆP j ˆPj,new,K ]. Increment j j +. Reset k. 4) Increment t t + and go to step. If the following conditions hold: ) the algorithm parameters are set as ξ = ξ 0 (ζ), 7ξ ω S min 7ξ, K = K(ζ), α α add (ζ), 2) slow subspace change holds: t j+ t j Kα; (3) holds with v =.2; and 4ξ 0 (ζ) S min, 3) denseness holds: max j κ 2s (P j ) κ + 2s, = 0.3 and max j κ 2s (P j,new ) κ + 2s,new = 0.5 where κ s (B) := max T s I T basis(b) 2 is the denseness coefficient introduced in [8], 4) matrices D j,new,k := (I ˆP j ˆP j ˆP j,new,k ˆP j,new,k )P j,new and Q j,new,k := (I P j,new P j,new ) ˆP j,new,k satisfy κ s (D j,new,k ) κ + s := 0.5, κ 2s (Q j,new,k ) κ + 2s := 0.5, 5) the condition number of the covariance matrix of a t,new is bounded, i.e., g g + = 2, 6) and η.7,

(Subspace change time) First Projection PCA Second Projection PCA K times Projection PCA PCA Fig. : The K Projection PCA steps. then, with probability at least ( n 0 ), ) at all times, t, ˆT t = T t and e t 2 = L t ˆL t 2 = Ŝt S t 2 0.8 c0.72 k γ new +.2 ζ( r + 0.023 c) 2) the subspace error SE (t) := (I ˆP (t) ˆP (t) )P (t) 2 satisfies (r 0 + (j )c)ζ + 0.5cζ +0.6 k SE (t) if t I j,k, k K (r 0 + jc)ζ if t I j,k+ { 0 2 ζ + 0.6 k if t I j,k, k K 0 2 ζ if t I j,k+ 3) the error e t = Ŝt S t = L t ˆL t also satisfies bounds similar to those on SE (t) given above [see [6]]. Proof: The full proof is given in [6]. We give a brief outline in Sec III-D. C. Discussion The above result says the following. Consider Algorithm 2. Assume that the initial subspace error is small enough. If the algorithm parameters are appropriately set, if slow subspace change holds, if the subspaces are dense, if the condition number of Cov[a t,new ] is small enough, and if the currently unestimated part of the newly added subspace is dense enough (this is an assumption on the algorithm estimates), then, w.h.p., we will get exact support recovery at all times. Moreover, the sparse recovery error will always be bounded by 0.8 cγ new plus a constant times ζ. Since ζ is very small, γ new S min, and c is also small, the normalized reconstruction error for recovering S t will be small at all times. In the second conclusion, we bound the subspace estimation error, SE (t). When a subspace change occurs, this error is initially bounded by one. The above result shows that, w.h.p., with each projection PCA step, this error decays exponentially and falls below 0.0 ζ within K projection PCA steps. The third conclusion shows that, with each projection PCA step, w.h.p., the sparse recovery error as well as the error in recovering L t also decay in a similar fashion. The above result allows the a t s, and hence the L t s, to be correlated over time; it models the correlation using an AR model which is a frequently used practical model. Even with this more general model as long as the AR parameter, b 0.4, we are able to get almost exactly the same result as that of Qiu et al [8, Theorem 4.]. The α needed is a little larger. Also, the only extra assumption needed is a small enough upper bound on η which is the ratio of the maximum magnitude entry of any ν t to the maximum variance. This is true for many types of probability distributions. For example if the i th entry of ν t is ±q i with equal probability independent of all others then η =. If each entry is zero mean uniform distributed (with different spreads) then η = 3. Like [8], we still need a denseness assumption on D new,k and Q new,k both of which are functions of algorithm estimates ˆP j and ˆP j,new,k. Because of this, our result cannot be interpreted as a correctness result but only a useful step towards it. As explained in [8], from simulation experiments, this assumption does hold whenever there is some support changes every few frames. In future work, we hope to be able to replace it with an assumption on the support change of S t s. Also, like [8], the above result analyzes an algorithm that assumes knowledge of the model parameters c, γ new and the subspace change times t j. Requiring kmowledge of c and γ new is not very restrictive. However it also needs to know the subspace change times t j and this is somewhat restrictive. One approach to try to remove this requirement is explained in [8]. As explained in [8], under slow subspace change, it is quite valid to assume that the condition number of the new directions, g, is bounded, in fact if at most one new direction could get added, i.e. if c =, then we would always have g =. On the other hand, notice that we do not need any bound on f. This is important and needed to allow E[ L t 2 2] to be large (L t is the large but structured noise in case S t is the signal of interest) while still allowing slow subspace change (needs small γ new and hence small λ γ new ). Notice that E[ L t 2 2] rλ +. D. Proof Outline The first step in the proof is to analyze the projected sparse recovery step and show exact support recovery conditioned on the fact that the subspace has been accurately recovered in the previous projection-pca interval. Exact support recovery along with the LS step allow us to get an exact expression for the recovery error in estimating S t and hence also for that of

L t. This exact expression is the key to being able to analyze the subspace recovery. For subspace recovery, the first step involves bounding the subspace recovery error in terms of sub-matrices of the true matrix, t Φ ˆL (t) t ˆL t Φ (t), and the perturbation in it, t Φ (t)(ˆl t ˆL t L t L t)φ (t), using the famous sin θ theorem [7]. This result bounds the error in the eigenvectors of a matrix perturbed by a Hermitian perturbation. The second step involves obtaining high probability bounds on each of the terms in this bound using the matrix Azuma inequality [8]. The third step involves using the assumptions of the theorem to show that this bound decays roughly exponentially with k and finally falls below cζ within K proj-pca steps. The most important difference w.r.t. the result of [8] is the following. Define the random variable X j,k := [ν 0, ν,, ν tj +kα ]. In the second step, we need to bound the minimum or maximum singular values of sub-matrices of terms of the form t f (X j,k )a t a tf 2 (X j,k ) where f (.), f 2 (.) are functions of the random variable X j,k. In Qiu et al [8], one could use a simple corollary of the matrix Hoeffding inequality [8] to do this because there, the terms of this summation were conditionally independent given X j,k. However, here they are not. We instead need to first use the AR model to rewrite things in terms of sub-matrices of t f (X j,k )ν t f 3 (ν 0, ν,, ν t ) f 2 (X j,k ). Notice that even now, the terms of this summation are not conditionally independent given X j,k. However, conditioned on X j,k, this term is now in a form for which the matrix Azuma inequality [8] can be used. Notice that the ReProCS algorithm does not need knowledge of b. If b were known, one could modify the algorithm to do proj-pca on (ˆL t bˆl t ) s. With this one could use the exact same proof strategy as in [8]. IV. SIMULATION In this section, we compare ReProCS with PCP using simulated data that satisfies the assumed signal model. The data was generated as explained in [8, Section X-C] except that here we generate correlated a t s using a t = ba t +ν t with b = 0.5 and with ν t, being uniformly distributed between [ γ, γ ] and ν t,new uniformly distributed between [ γ new,k, γ new,k ]. Also we set t train = 40, S min = 2, S max = 3, s = 7, r 0 = 2, n = 200, J = 2, v =. and the support T t was constant for every set of 50 frames and then changed by index. Other parameters were the same as those in [8, Section X-C]. By running 00 Monte Carlo simulations, we got the result shown in Figure 2. REFERENCES [] D. Skocaj and A. Leonardis, Weighted and robust incremental method for subspace learning, in IEEE Intl. Conf. on Computer Vision (ICCV), vol. 2, Oct 2003, pp. 494 50. [2] Y. Li, L. Xu, J. Morphett, and R. Jacobs, An integrated algorithm of incremental and robust pca, in IEEE Intl. Conf. Image Proc. (ICIP), 2003, pp. 245 248. [3] F. D. L. Torre and M. J. Black, A framework for robust subspace learning, International Journal of Computer Vision, vol. 54, pp. 7 42, 2003. log 0 ( Ŝt St 2/ St 2) 0.5.5 2 2.5 3 3.5 ReProCS Recursive PCP 4 0 500 000 500 2000 2500 time Fig. 2: Comparing recovery error of PCP implemented at the time instants shown by the triangles and of ReProCS. [4] E. J. Candès, X. Li, Y. Ma, and J. Wright, Robust principal component analysis? Journal of ACM, vol. 58, no. 3, 20. [5] M. Mardani, G. Mateos, and G. Giannakis, Dynamic anomalography: Tracking network anomalies via sparsity and low rank, J. Sel. Topics in Sig. Proc., Feb 203. [6] V. Chandrasekaran, S. Sanghavi, P. A. Parrilo, and A. S. Willsky, Rank-sparsity incoherence for matrix decomposition, SIAM Journal on Optimization, vol. 2, 20. [7] C. Qiu and N. Vaswani, Real-time robust principal components pursuit, in Allerton Conference on Communication, Control, and Computing, 200. [8] C. Qiu, N. Vaswani, B. Lois, and L. Hogben, Recursive robust pca or recursive sparse recovery in large but structured noise, under revision for IEEE Tran. Info. Th., also at arxiv: 2.3754[cs.IT], shorter versions in ICASSP 203 and ISIT 203. [9] C. Qiu and N. Vaswani, Recursive sparse recovery in large but correlated noise, in 48th Allerton Conference on Communication Control and Computing, 20. [0] J. He, L. Balzano, and A. Szlam, Incremental gradient on the grassmannian for online foreground and background separation in subsampled video, in IEEE Conf. on Comp. Vis. Pat. Rec. (CVPR), 202. [] C. Hage and M. Kleinsteuber, Robust pca and subspace tracking from incomplete observations using l0-surrogates, arxiv:20.0805 [stat.ml], 203. [2] A. E. Abdel-Hakim and M. El-Saban, Frpca: Fast robust principal component analysis for online observations, in Pattern Recognition (ICPR), 202 2st International Conference on. IEEE, 202, pp. 43 46. [3] J. Feng, H. Xu, and S. Yan, Online robust pca via stochastic optimization, in Adv. Neural Info. Proc. Sys. (NIPS), 203. [4] J. Feng, H. Xu, S. Mannor, and S. Yan, Online pca for contaminated data, in Adv. Neural Info. Proc. Sys. (NIPS), 203. [5] E. Candes and T. Tao, Decoding by linear programming, IEEE Trans. Info. Th., vol. 5(2), pp. 4203 425, Dec. 2005. [6] J. Zhan and N. Vaswani, Correlated low rank coefficients in recursive robust pca, http://home.engineering.iastate.edu/ jzhan/ CorrReProCSLong.pdf. [7] C. Davis and W. M. Kahan, The rotation of eigenvectors by a perturbation. iii, SIAM Journal on Numerical Analysis, vol. 7, pp. 46, Mar. 970. [8] J. A. Tropp, User-friendly tail bounds for sums of random matrices, Foundations of Computational Mathematics, vol. 2, no. 4, pp. 389 434, 202.