Elsevier

Neurocomputing

Volume 386, 21 April 2020, Pages 165-178
Neurocomputing

Robust principal component analysis with intra-block correlation

https://doi.org/10.1016/j.neucom.2019.12.092Get rights and content

Abstract

This paper proposes a novel optimization program for solving the Robust Principle Component Analysis (RPCA) problem, which decomposes a data matrix into a conventional low-rank part plus a particular block-sparse residual. This kind of block-sparse residual often scattered in the source signal scope as contaminants, and often existed in many practical applications, such as an ordinary imaging system, a Hyper Spectral Imaging system, EEG and MEG, and types of physiological signals. Different from most currently existing approaches, the study perceived especially a highly spatial correlation among the inner structure of the neighbouring pixels in this contiguously block-sparse residual. The high intra-block correlation is then introduced as prior information to deal the governing optimization problem. In order to enhance the block-sparsity and maintain the local smoothness simultaneously, a localized low-rank promoting method is introduced with a theoretical guarantee. An efficient solving algorithm is designed accordingly with a convergence analysis by adopting the classical Alternating Direction Method of Multipliers (ADMM) framework. In addition to the theoretical model derivation, several synthetic simulations together with a real application on image denoising experiment have been conducted to validate the proposed model. As expected, the models outperforms significantly the compared state-of-the-arts.

Introduction

In the last decade, classical PCA [1] is the most widely used tool for dimensionality reduction, but is highly sensitive to grossly corrupted observations. Unfortunately, gross errors are now ubiquitous in practical applications such as face recognition [2], image inpainting [3], and video surveillance [4]. Robust Principle Component Analysis (RPCA) [5], [6], as an extension of PCA, have been proposed to seek the low-dimensional structure from highly corrupted measurements with arbitrary magnitude entries. This approach possessed of a polynomial-time algorithm receives much attention in many problems of Low-Rank and Sparse Decomposition (LRSD), including background subtraction [7], latent semantic indexing [8] and image/video denoising [9], etc.

Suppose we have a large data matrix MRm×n which can generally be decomposed as M=L+S, as shown in Fig. 1 (up row). The RPCA method aims to recover the low-rank matrix LRm×n from the real observation matrix MRm×n with the premise of modelling the corrupted noise parts with arbitrary distribution and magnitude as a sparse matrix SRm×n. Mathematically, it can be realized by solving the following optimization minimization [5], [10]:minL,Srank(L)+λS0s.t.M=L+S,where rank( · ) works as a sparsity regularization of matrix singular values; ‖ · ‖0 denotes the number of the nonzero elements, which is used as the regularization term to promote sparsity; and λ > 0 is denoted as the relatively balanced parameter between the two term. Unfortunately, minimizing the optimization function directly is an NP-hard problem, i.e., it is difficult to be solved within polynomial time. Hence, in order to deal with this issue, we consider a surrogate problem which can be formulated as computing the minimizer ofminL,SL*+λg(S)s.t.M=L+S,where L*=iδi(L) denotes the nuclear norm [11] and δi(L) represents the ith singular value of L (sorted in decreasing order). In terms of the surrogate functions g(S) of the sparse penalty term S0, the general option of classical RPCA is l1-norm S1=i=1mj=1n|Sij| [5]. However, the l1-norm is only a loose approximation of the l0-norm, which will result in a suboptimal solution to the original l0-minimization in reality. Several more efficient surrogate functions g(S) are hence proposed for extending the classical RPCA model, such as lp-norm (0 < p < 1) [12], Smoothly Clipped Absolute Deviation (SCAD) [13], Exponential-Type Penalty (ETP) [14] and generalized Weighted-based Integral Convolution Penalty (g-WICP) [15], etc.

It is well known that the formula of l0-norm treats each entry (pixel) in the sparse matrix independently, thus the classical RPCA method is extremely insensitive to column/row outliers. And we further find that the block-sparse structural information does exist abundantly in real circumstances. For example, the background (see Fig. 2(b)) of an image (see Fig. 2(a)) which randomly selected from Wallflower dataset [16] can be easily captured by the low-rank matrix, while the corresponding foreground object (see Fig. 2(c)) can be regarded as a block-sparse corruption for it is spatially contiguous and occupies a portion of the scene. Under the condition of block corruption, a l2,0-norm which stands for the number of nonzero l2-norm of the column was proposed to replace the original l0-norm to manage the block corruption of the sparse matrix. Therefore the new optimization problem for decomposing an observation matrix into a low-rank part plus a block-sparse residual can be reformulated as:minL,SL*+λS2,0s.t.M=L+S.An illustration for RPCA with a block-sparse residual is shown in the bottom row of Fig. 1 to contrast significantly to that with snow-sparse residual. For this block-sparse penalty term, the l2,1-norm S2,1=i=1mj=1n|Sij2| was first adopted in [17] to approximate the l2,0-norm for detecting outliers with column-wise sparsity. The l2,p-norm has also been proved more exact than the l2,1-norm when imposed on the reconstruction error to handle outlier pursuit [12]. Related works of low-rank and block-sparse matrix decomposition are currently following up with some researches. In [18], Tang and Nehorai proposed a RPCA-LBD detector to separate the primary components from the outliers, which conducted a good result when the sparse pattern involved in an entire column. With the potential for outlier detection applications, this approach then was wildly applied to many practical problems, such as salient motion detection [19] and imaging processing [20]. Moreover, a great progress in applying one alternative block-sparse RPCA to the topic of background modelling was achieved too. By Liu et al. [21] who proposed a novel approach, namely Low-rank and Structured sparse Decomposition (LSD), for foreground detection, and attained the algorithm successfully through the method of Augmented Lagrange Multiplier (ALM). The essential of such an issue is still regarded as the decomposition of an underlying image into a low-rank matrix and a block-sparse outlier matrix.

Although there were few previous works considered to include the prior information about the high-correlated structure in the outlier matrix. Apart from block sparsity, the images in real applications also often have the tendency to own extra useful structural compositions, which can be explored to improve the reconstruction performance. As shown in Fig. 3(a), it is obvious to find that there are several non-zero structural blocks in the actual foreground image (Fig. 3(b)). Upon this observation, a small non-zero region which owned such kind of structural block was randomly picked to examine the correlations among the compositions within the region. A correlation graph is hence plotted along the rows as shown in Fig. 3(c). The graph concludes visually that the inner structure among the neighbouring elements of these columns possesses highly spatial correlation. Such kind of observations are also found in many other practical scenarios. For example, when HyperSpectral Imaging (HSI) [22] technique is applied to collect approximately continuous spectral information of ground objects, the block-sparse structures with correlated coefficients are regularly appeared since the object generally covers a continuous portion of the scene. Source localization in EEG and MEG [23] falls into this category too, i.e., the coefficients in a certain source may be highly correlated. The other related examples involved several physiological signals, e.g. blood pressure, MECG, and temperature [24], which are all regarded to be typically spatially continuous signals with intra-block correlation. So far, there were several algorithms applied to solve the intra-block correlation problem. Typically, a method named Localized lOw rank Promoting (LOOP) was proposed in [25] for recovery of the block-sparse signals with intra-block correlated elements, in which the neighboring coefficients of the sparse signal were organized to form a number of 2 × 2 matrices. Clearly, these matrices were mostly not full-ranked. A log-determinant function was hence adopted to approximately suppress the matrices in low rank, and guaranteed its corresponding performance. The intra-block high-correlated entries act as the pieces of prior information, which guarantees the recovery of the block-sparse signals even without requiring the knowledge of the sizes and locations of the blocks.

In this work, a new approach of low-rank and block-sparse decomposition motivated by the LOOP method, named RPCA-LOOP, is proposed. Different from other RPCA methods, this new proposed model specially deals with the pixels of matrices in a correlative manner with their neighboring pixel values. To separate the principal components from the block-sparse outliers with the prior high correlation, an efficient optimization approach which can be mainly summarized as two steps, has been designed. In the first step, since the block-sparsity penalty term in the original optimization problem is replaced by a non-convex logarithmic function, many effective algorithms, applied to solve the traditional l1-norm convex minimization problem, are not appropriate to be used here. Instead of directly minimizing the objective function, Majorization Minimization (MM) approach [26] was adopted to solve the non-convex minimization model. The MM approach suggests that iteratively minimizing a simple convex upper-bound can push the objective function downhill to the optimum. The related plausibility and operability also have been demonstrated in [27], [28]. Thus, the original model can be reformulated as a convex optimization surrogate. In the second step, to solve this convex minimization program, an effective iterative scheme based on the ADMM framework is employed to converge to the minimum of the original problem, and the optimization subproblems both have closed-form solutions. Moreover, a detailed convergence analysis of ADMM with respect to our new problem is presented to complete the study.

The main contributions of the paper are summarized as follows:

  • A novel and effective model, called RPCA-LOOP, is proposed, which not only is an ordinary decomposition for both the low-rank component and the block-sparsity component, but also exploits spatial high-correlation structures existing in the columns of the block-sparsity component which can provide better performance in handling many practical applications.

  • A MM approach is employed to approximate the original non-convex minimization model by a convex surrogate optimization problem, which can be solved with the unique closed-form solution by the polynomial-time iterative algorithms, i.e. ADMM. Particularly, the theoretical convergence analysis is also discussed by a clear proof in detail.

  • A series of numerical experiments together with image denoising applications are taken to demonstrate the superiority of the proposed RPCA-LOOP.

The remainder of paper is organized as follows. In Section 2, we give some notations and preliminaries which are used in later section. Section 3 presents an efficient algorithm to solve RPCA problem with a block-sparse residual by adopting the ADMM. In Section 4, the convergence analysis of the new proposed algorithm is given in detail. Experimental results are presented in Section 5 to demonstrate the effectiveness of our method. Section 6 summarizes the conclusion and future work.

Section snippets

Notations and preliminaries

Convention of notations and preliminaries adopted for the article are given as follows. Whereas boldface capital letters, e.g. A are adopted for representing matrices; boldface lowercase letters, e.g. a are for vectors. A matrix Imn represents a mn × mn identity matrix. For a matrix ARm×n, the (i, j)th entry of A is conventionally denoted as Aij, where R denotes the fields of real numbers; The notation det(A) is reserved for the determinant of the matrix A or alternatively for a scalar of its

Proposed method

In this section, a novel regularization model for RPCA problem with a block-sparse residual together with its algorithms will be illustrated. Firstly, the assumptions that the matrix L is low-rank and S has only a few non-zero blocks in its columns are adopted as fundamental restrictions of this paper. Aside from these conditions, intrablock highly correlated nonzero coefficients are also intrinsically existed in the processed matrix S, although the number and the locations of non-zero blocks

Convergence analysis

Although the original constrained objective function of the RPCA-LOOP model (9) is non-smooth, the Algorithm 1 to solve the augmented Lagrange function (10) can still be proved to has the same excellent convergence property. More precisely, we will establish the formal statements below to further explain the validity of the proposed method. Our proof adapts the proof approach of [30] to our new complicated model. Before going on, the optimal value P* of the optimization problem (9) is defined as

Numerical experiment

In this section, to verify the effectiveness of our proposed method in practical applications, we conduct a series of tests including simulation experiments and real-image denoising. In particular, since this work is the extension of RPCA problem, we compare the performance of the proposed method directly with three existing RPCA approaches: RPCA-ALM [6], RPCA-LBD [18] and RPCA-LSD [21]. In addition, all the numerical experiments are conducted in MATLAB R2016b on a desktop computer with a AMD

Conclusion and future work

In this paper, we study the Robust Principal Component (RPCA) problem with a block-sparse residual which aims to combine highly-correlation priors to recover a low rank matrix and a sparse matrix from their sum. Our proposed method characterizes the intrablock correlated entries by organizing the sparse matrix to form a number of 2 × 2 matrices and adopting the log-determinant function to promote their block-sparsity. To solve the new model, we design an efficient iterative algorithm by using

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Acknowledgment

This work was supported by National Natural Science Foundation of China (Grant nos. 61673015, 61273020), Fundamental Research Funds for the Central Universities (Grant nos. XDJK2018C076, SWU1809002) and China Postdoctoral Science Foundation (Grant no. 2018M643390), Graduate Student Scientific Research Innovation Projects in Chongqing (no. CYB19083).

Can Jiang received the B.S. degree from the School of Mathematics and Statistics, Southwest University, Chongqing, China, in 2017. Currently, she is pursuing the M.S. degree with the School of Mathematics and Statistics, Southwest University. Her research focuses on compressed sensing and image processing.

References (33)

  • H. Yong et al.

    Robust online matrix factorization for dynamic background subtraction

    IEEE Trans. Pattern Anal. Mach. Intell.

    (2017)
  • K. Min et al.

    Decomposing background topics from keywords by principal component pursuit

    Proceedings of the Nineteenth ACM international conference on Information and knowledge management

    (2010)
  • C. Lu et al.

    Tensor robust principal component analysis: exact recovery of corrupted low-rank tensors via convex optimization

    Proceedings of the IEEE Conference on Computer Vision and Pattern recognition,(CVPR)

    (2016)
  • V. Chandrasekaran et al.

    Rank-sparsity incoherence for matrix decomposition

    SIAM J. Optim.

    (2011)
  • M. Fazel

    Matrix rank minimization with applications

    (2002)
  • J. Fan et al.

    Variable selection via nonconcave penalized likelihood and its oracle properties

    J. Am. Stat. Assoc.

    (2001)
  • Cited by (0)

    Can Jiang received the B.S. degree from the School of Mathematics and Statistics, Southwest University, Chongqing, China, in 2017. Currently, she is pursuing the M.S. degree with the School of Mathematics and Statistics, Southwest University. Her research focuses on compressed sensing and image processing.

    Feng Zhang received the M.S. degree from the School of Mathematics and Statistics, Southwest University, Chongqing, China, in 2017. He is currently pursuing the Ph.D. degree with the School of Mathematics and Statistics, Southwest University. His research focuses on compressed sensing and tensor sparsity.

    Jianjun Wang received the B.S. degree in mathematical education from Ningxia University in 2000 and the MS degree in fundamental mathematics in 2003 from Ningxia University, China. And Ph.D. degree in applied mathematics was obtained from the Institute for Information and System Science, Xi’an Jiaotong University in Dec. 2006. He is currently a professor in the College of Artifical Intelligence at Southwest University of China. His research focuses on machine learning, data mining, neural networks and sparse learning.

    Chan-Yun Yang received his B.S. and M.S. degrees from National Taiwan University, Taiwan, in 1985 and 1989, respectively, and got his Ph.D. degree of Bio-Industrial Mechatronics Engineering from National Taiwan University in 2001. He is currently a professor in the Department of Electrical Engineering at National Taipei University. His research interests include machine learning, robotic path planning, and complex system modeling.

    Wendong Wang received the M.S. degree from the School of Mathematics and Statistics, Southwest University, Chongqing, China, in 2014. And the Ph.D. degree from the School of Computer and Information Science, Southwest University, in 2017. Currently, he is studying as Postdoctoral Fellow at the school of mathematics and statistics, Southwest University. His research interests include compressed sensing and machine learning.

    View full text