Elsevier

Signal Processing

Volume 164, November 2019, Pages 284-294
Signal Processing

Cauchy greedy algorithm for robust sparse recovery and multiclass classification

https://doi.org/10.1016/j.sigpro.2019.06.006Get rights and content

Highlights

  • We present a robust greedy algorithm called CauchyMP (Cauchy Matching Pursuit) for robust sparse recovery.

  • We generalize CauchyMP for robust block and quaternion sparse recovery.

  • We devise a CauchyMP based classifier for robust multiclass classification.

  • Experimental results show that the proposed methods improve related competing algorithms with notable performance gains in presence of various gross corruption and outliers.

Abstract

Greedy algorithms have attracted considerable interest for sparse signal recovery (SSR) due to their appealing efficiency and performance recently. However, conventional greedy algorithms utilize the ℓ2 norm based loss function and suffer from severe performance degradation in the presence of gross corruption and outliers. Furthermore, they cannot be directly applied to the recovery of quaternion sparse signals due to the noncommutativity of quaternion multiplication. To alleviate these problems, we propose a robust greedy algorithm referred as Cauchy matching pursuit (CauchyMP) for SSR and extend it for quaternion SSR. By leveraging the Cauchy estimator and generalizing it to the quaternion space to measure the residual error, our method can robustly recover the sparse signal in both real and quaternion space from noisy data corrupted by various severe noises and outliers. To tackle the resulting quaternion optimization problem, we develop an efficient half-quadratic optimization algorithm by introducing two quaternion operators. In addition, we have also devised a CauchyMP based classifier termed CauchyMPC for robust multiclass classification. The experiments on both synthetic and real-world datasets validate the efficacy and robustness of the proposed methods for SSR, block SSR, quaternion SSR and multiclass classification.

Introduction

Sparse representation (SR) has shown great potential in a variety of problems in signal processing and computer vision in the past decade [1]. For instance, SR methods have achieved great success in signal recovery [6], face recognition [1], and image restoration [8]. To be specific, given a dictionary matrix DRm×n(m<n), SR aims to recover the target sparse signal x0Rn from its compressed measurement vectory=Dx0+n,where nRm denotes the noise vector.

According to the mechanisms of inducing sparsity, existing SR approaches can be roughly divided into two categories: ℓ1 minimization based approaches [5] and greedy algorithms [2]. A natural SR approach is to optimize the following problem related to the ℓ0 constraintminxyDx22subjecttox0K.where K denotes the sparsity parameter. However, the problem above is generally NP-hard due to the discontinuity and discrete nature of the ℓ0. This makes it intractable to solve the problem (2) directly.To reduce such problem, ℓ1 minimization based approaches consider the following surrogate optimization problemminxyDx22+λx1,which is termed as Lasso [5]. Here the nonnegative scalar λ is the regularization parameter. While Lasso enjoys striking theoretical properties under appropriate conditions, most ℓ1 minimization based approaches require heavy computation burden [6].

Unlike Lasso, greedy algorithms (GA) iteratively identify the indexes of nonzero entries of x0 and estimate the sparse vector. Due to the low complexity and competitive performance, greedy algorithms have attracted increasing interest in recent years. Orthogonal matching pursuit (OMP) [2] is perhaps the most popular GA method because of its simplicity. Specifically, OMP identifies a column of the dictionary D in each iteration and estimates the sparse vector using the selected atoms in previous iterations. To improve the efficiency of OMP, the generalized OMP (GOMP) [2] selects multiple informative atoms in each iteration. The regularized OMP (ROMP) [4] algorithm tries to keep the advantages of both Lasso and OMP, i.e., the strong theoretical guarantees of Lasso and the high efficiency of OMP using a regularization rule. Analogously, the compressive sampling matching pursuit (CoSaMP) [3] also enhances OMP with an additional pruning step to provide strong theoretical guarantees that OMP cannot.

Despite their empirical success, most existing greedy algorithms explore the squared ℓ2 norm as the loss function, which depends on the Gaussianity assumption of the noise distribution and sensitive to outliers. A violation of this assumption, e.g., missing entries, impulsive noise or random occlusions in face image data, may lead to severe performance degradation. To reduce the limitation, various robust SR approaches have been developed recently. The first category of robust SR methods aim to improve the robustness of lasso against outliers and gross corruptions. For instance, Carrillo and Barner [7] exploit the Lorentzian-norm to measure the residual error and introduce a geometric optimization problem for robust SR. Recent studies [8] have shown that leveraging the ℓ1 norm based loss function in Lasso can lead to much better robustness compared with the conventional ℓ2 norm and Lorentzian-norm. The resulting objective function isminxyDx1+λx1.To solve the 11 norm optimization problem above, many effective algorithms have been devised such as YALL1 [9]. In [10], the generalized ℓp-norm, 0 ≤ p < 2 is also adopted as the loss function for the residual error and the authors developed an alternating direction method based algorithm termed Lp-ADM for the corresponding optimization problem. Since these robust SR methods are still based on ℓ1 norm regularization, these methods have heavy computational burden.

The second category of robust SR methods attempt to improve the robustness of greedy algorithms while keeping high efficiency. In [11], Razavi et al. attempt to robustify conventional greedy algorithms such as OMP by drawing on robust statistics and replacing least squares regression in OMP with robust regression. The robust OMP (RobOMP) [11] first calculates the so-called residual pseudo-values and selects a new atom which has the largest correlation with the residual pseudo-values in each iteration. In [12], Zeng et al. generalize conventional GA algorithms such as MP and OMP from inner product space to ℓp space (p > 0) for robust SR. The robust version of MP and OMP are called ℓp-MP and ℓp-OMP, respectively.

Another weakness of most existing SR methods is that they are designed for real or complex sparse recovery and cannot be directly applied to quaternion sparse signal recovery (QSSR). Because the product of two quaternions are noncommunicative in general, i.e., q˙1q˙2q˙2q˙1. In addition, the definition and computation of the derivative (or gradient) of quaternion matrix function are much more complicated than those in Rn or Cn [13], [14]. This greatly increases the difficulty to tackle the quaternion optimization problems for QSSR.

In fact, quaternion has been widely used in various applications, including but not confined to, vector-sensor array signal processing [15], color face recognition [13], color image denoising, superresolution and inpainting [14]. Recent advances on quaternion image analysis [13], [14] show that quaternions are well adapted to color images by encoding the color channels into the three imaginary parts. In [14], Xu et al. extend OMP to the quaternion space and devise the quaternion OMP (QOMP) algorithm with application to color image restoration. In our previous work [13], we propose the quaternion Lasso (QLasso) model with quaternion ℓ1 minimization for QSSR and color face recognition. However, both QOMP and QLasso rely on the quaternion ℓ2 norm based loss function and may be sensitive to gross corruption and outliers.

In this paper, we develop a robust greedy algorithm referred as Cauchy Matching Pursuit (CauchyMP) by exploiting and generalizing the Cauchy estimator for robust sparse signal recovery (SSR) and quaternion SSR. By devising the half-quadratic theory [16] based optimization algorithm, CauchyMP can be viewed as an adaptive weighted OMP approach. The intuition behind CauchyMP is that it adaptively assign large weights on clean entries of y and small weights on noisy or outlying entries of y. Accordingly, the impact of corrupted entries and outliers can be well alleviated. The contributions of this work are summarized as below.

  • 1.

    We present a CauchyMP algorithm for robust sparse signal recovery. In the presence of gross corruption and outliers, CauchyMP can improve many prior greedy algorithms with notable performance gains.

  • 2.

    We generalize CauchyMP and devise the quaternion CauchyMP (QCauchyMP) algorithm for the recovery of quaternion sparse signals. Since the product of quaternions is noncommunicative in general, previous robust SR approaches cannot be directly applied to quaternion sparse signal recovery.

  • 3.

    We devise a CauchyMP based classifier termed CauchyMPC for robust multiclass classification and establish the theoretical analysis. Compared to the original sparse representation-based classification (SRC) [1], the proposed approach has better robust property and is more efficient.

The key differences between prior robust greedy algorithms (e.g., RobOMP and ℓp-OMP) and the proposed method lie in the following two aspects. First, the steps of identifying a new atom by these methods are different. Concretely, RobOMP identifies a new atom which has the largest correlation with the residual pseudo-values eψ and ℓp-OMP selects the atom that has the largest ℓp correlation with the residual. While CauchyMP selects a new atom most correlated with the residual in a reweighted versionjk=argmaxj=1,,n|dj,wk1r|,where wk1 is the weight vector indicating the importance of each entry of r. For real vectors u,vRm, the inner product is defined as u,v=uTv, while for quaternion vectors u˙,v˙Hm, the inner product is defined by u˙,v˙=u˙Hv˙. Here u˙H=[u˙¯1,,u˙¯m]T denotes the conjugate transpose of u˙ and u˙¯i is the conjugate of the quaternion u˙i. Specifically, if the ith entry ri of r is severely corrupted, it will receive small weight (the ith entry of wk1, which is estimated adaptively) and the impact of noisy entries can be effectively suppressed. Thus the identification step of CauchyMP has clear and intuitive explanation. Second, the steps of estimating the sparse signal by these methods are also distinct. Specifically, both RobOMP and ℓp-OMP use the Iteratively Re-weighted Least Squares (IRLS) algorithm while CauchyMP estimates the sparse signal by the half-quadratic theory based optimization algorithm with guaranteed convergence.

The rest of this paper is organized as follows. In Section 2.2, we present the proposed methods for sparse signal recovery (SSR), block SSR, quaternion SSR and multiclass classification, respectively. Section 3 presents the experiments. Finally, Section 4 concludes the paper.

In this work, scalars, vectors and matrices are represented using italic letters (e.g., x), boldface lowercase letters (e.g., x), and boldface capital letters (e.g., X), respectively. For each vector xRn and an index set J{1,2,,n}, xj denotes its jth entry and xJ denotes a subvector of x containing entries indexed by the set J. Analogously, for a matrix XRm×n, XJ denotes a submatrix of X containing columns of X indexed by the set J. Table 1 summarizes the key notations and acronyms used in this paper.

Section snippets

The proposed approach

This section is arranged as follows. Firstly, we introduce the Cauchy estimator and propose a novel robust greedy algorithm referred to as Cauchy Matching Pursuit (CauchyMP) for robust SSR. Secondly, we generalize CauchyMP to the recovery of block sparse signal by exploiting the block sparsity. Thirdly, we extend CauchyMP and develop the quaternion CauchyMP (QCauchyMP) for the recovery of quaternion sparse signals. Finally, we devise a CauchyMP based classifier for robust multiclass

Experiments

In this section, we evaluate the performance of the proposed methods for SSR, block SSR, quaternion SSR and multiclass classification, respectively.

Conclusion

This paper presents a novel greedy algorithm referred as CauchyMP for robust sparse signal recovery in real and quaternion space and multiclass classification. Specifically, CauchyMP leverages the robust Cauchy estimator to recover the sparse signal, which can tolerate noisy data contaminated with various severe noise. The experiments demonstrate the efficacy of the proposed methods for sparse recovery and classification.

Disclosure of conflicts of interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Acknowledgments

This work was supported by the National Natural Science Foundation of China under Grant Nos. 61806027, 61702057, 61672114 and 11771130.

References (27)

  • D. Needell et al.

    CoSaMP: iterative signal recovery from incomplete and inaccurate samples

    Appl. Comp. Harmonic Anal.

    (2009)
  • X. Zhang et al.

    Quaternion-valued robust adaptive beamformer for electromagnetic vector-sensor arrays with worst-case constraint

    Signal Process.

    (2014)
  • I. Mizera et al.

    Breakdown points of cauchy regression-scale estimators

    Stat. Probab. Lett.

    (2002)
  • J. Wright et al.

    Robust face recognition via sparse representation

    IEEE Trans. Pattern Anal. Mach. Intell.

    (2009)
  • J. Wang et al.

    Generalized orthogonal matching pursuit

    IEEE Trans. Signal Process.

    (2012)
  • D. Needell et al.

    Signal recovery from incomplete and inaccurate measurements via regularized orthogonal matching pursuit

    IEEE J. Sel. Topics Signal Process.

    (2010)
  • R. Tibshirani

    Regression shrinkage and selection via the lasso

    J. Roy. Stat. Soc. Ser. B

    (1996)
  • D. Donoho et al.

    Sparse solution of underdetermined systems of linear equations by stagewise orthogonal matching pursuit

    IEEE Trans. Inf. Theory

    (2012)
  • R. Carrillo et al.

    Lorentzian iterative hard thresholding: robust compressed sensing with prior information

    IEEE Trans. Signal Process.

    (2013)
  • D.S. Pham et al.

    Efficient algorithms for robust recovery of images from compressed data

    IEEE Trans. Image Process.

    (2013)
  • J. Yang et al.

    Alternating direction algorithms for ℓ1-problems in compressive sensing

    SIAM J. Sci. Comput.

    (2011)
  • F. Wen et al.

    Robust sparse recovery in impulsive noise via ℓp-ℓ1 optimization

    IEEE Trans. Signal Process.

    (2017)
  • S. Razavi et al.

    Robust greedy algorithms for compressed sensing

    Proceedings of the 20th European Signal Process Conference (EUSIPCO)

    (2012)
  • Cited by (5)

    • Correlation-based local detection for deceptive interference mitigation in multi-parameter modulated radar

      2022, Signal Processing
      Citation Excerpt :

      Under the precondition that the signal can be represented sparsely in some transform domain, the signal can be recovered by few observation so as to acquire the effective target information [15,16]. There are numerous studies regarding the recovery algorithms, like the greedy algorithm [17–19], convex or non-convex optimization [20,21] and the signal statistics-based method [22], which demonstrate the remarkable performance in signal recovery. Some of them, like forward-backward pursuit (FBP) [23] and compressed sampling matching pursuit (CoSaMP) algorithm [24], can even reconstruct the signal in a relatively low signal-noise-ratio (SNR) case.

    • Model recovery for multi-input signal-output nonlinear systems based on the compressed sensing recovery theory

      2022, Journal of the Franklin Institute
      Citation Excerpt :

      The CS recovery algorithms include the convex optimization algorithms, the greedy algorithms and the combination algorithms. The greedy algorithms have received much attention for the identification problem due to their computationally simple [23,24]. The matching pursuit algorithm, the orthogonal matching pursuit (OMP) algorithm and their variations all are typical greedy algorithms.

    • Atom selection strategy for signal compressed recovery based on sensing information entropy

      2021, ISA Transactions
      Citation Excerpt :

      Expectedly, researchers also dedicate to pursuing a trade-off between the reconstructed precision and the recovery speed. The existing recovery algorithms mainly include greedy algorithm [7–9], Bayesian learning algorithm [10–12] and machine learning algorithm [13,14]. The application of the last two methods is limited due to the enormous computation and rigorous hardware condition.

    View full text