Elsevier

Pattern Recognition Letters

Volume 149, September 2021, Pages 44-50
Pattern Recognition Letters

Robust subspace clustering network with dual-domain regularization

https://doi.org/10.1016/j.patrec.2021.06.009Get rights and content

Abstract

The field of deep subspace clustering has advanced rapidly in recent years. Ideas such as self-expression and self-supervision have led to innovative network design and improved clustering performance. However, it is observed that the nonlinear low-dimensional manifold constraint is valid in not only the ambient but also the latent feature spaces. Meanwhile, the issue of robustness has been largely overlooked in the literature of deep subspace clustering despite previous studies on robust model-based clustering. Based on these two observations, we present a robust subspace clustering network (RSCN) based on a novel hybrid loss function with dual-domain regularization. On the one hand, we propose to replace the existing L2 loss by a robust hybrid function inspired by half-quadratic minimization; on the other hand, we come up with a novel strategy of sparsity regularization in the dual domain (both ambient and feature space). To the best of our knowledge, this is the first attempt to incorporate dual manifold constraints into deep subspace clustering. Experimental results show that our new network outperforms the existing state-of-the-art on several widely-studied datasets such as Extended Yale B, COIL20, and COIL100. The performance gain of our RSCN over several other competing approaches improves dramatically in the presence of noise contamination.

Introduction

Subspace clustering - a problem of clustering high-dimensional data into multiple groups satisfying a subspace constraint – has been extensively studied in the past decades (please refer to the tutorial and reviewing articles [1], [7], [27], [36]). Early works [1], [27] focused on the curse of dimensionality and tackled the challenge by feature transformation techniques [27]. Subspace clustering can be viewed as an extension of feature selection targeting at finding clusters in well-structured subspaces [1]. Inspired by the ideas of sparse coding or sparse representation [5], [43], sparse subspace clustering (SSC) has been proposed to cluster data in [6], [7], [40] based on the observation that each point in a union of subspaces admits a sparse representation with respect to the dictionary formed by all other data points. This line of research was further developed by connecting with low-rank matrix decomposition [28], information theoretic formulation [12], orthogonal matching pursuit [47] and elastic net theory [46]. The framework of SSC has been extended into kernel SSC in [29] and structured SSC in [19].

Most recently, breakthroughs in deep learning [8], [17] have made a splash in the field of subspace clustering, leading to a flurry of deep subspace clustering networks [3], [4], [15], [30], [31], [32], [48], [50]. In [32], a deep learning model named PARTY was developed to progressively transform the input data into nonlinear latent space; in addition to being adaptive to the local and global subspace structure simultaneously, PARTY incorporates a sparsity prior in order to preserve the sparse reconstruction over the whole data set. In [15], the self-expressiveness property - originated from SSC [6] – was introduced into the architecture of deep convolutional auto-encoder, which produced a novel deep subspace clustering network (DSCN). In [50], the concept of adversarial learning [9] was leveraged into subspace clustering, leading to a deep adversarial subspace clustering (DASC) network. DASC consists of a subspace clustering generator and a quality-verifying discriminator, which learn against each other in a generative adversarial network (GAN)-like manner [9]. In [48], an end-to-end trainable framework consisting of three modules (feature learning, subspace clustering and self-supervision as shown in Fig. 1) was developed and demonstrate state-of-the-art performance on several well-known data sets. Most recently, multiscale fusion and multilevel representation have been studied for subspace clustering in [4], [16] respectively.

The motivation behind this paper is two-fold. On the one hand, it has been observed that not only the original high-dimensional image data reside on a nonlinear low-dimensional manifold in the ambient space, but also their feature representations lie on a manifold in the latent space [45]. Such observation has inspired the regularization strategy constructed on a dual graph aiming at simultaneously exploiting the geometric structures of both data manifold and feature manifold. Unfortunately, similar ideas have not been considered for the class of deep subspace clustering networks. On the other hand, the issue of robustness – even though it has been well addressed in conventional frameworks such as robust PCA (RPCA) [2], [42] and robust subspace clustering [34], has not been formally addressed under the framework of deep subspace clustering (with the only exception of [31]). The presence of outliers - either due to errors in observation data or the deviation from a linear subspace model - could have severe impact on the performance of subspace clustering techniques on real-world data [48].

Inspired by previous works of nonconvex penalty function [44] and half-quadratic minimization [49], we propose to replace the L1/L2 norm in the original DSCN by a robust loss function consisting of both L1 and L2 terms. In particular, the L1-regularization term reflects the sparsity constraint on modeling errors, which is well aligned with the ideas in previous model-based approaches such as RPCA [2], [42]. Meanwhile, we suggest that such a hybrid loss function can be combined with dual-domain regularization for greater flexibility and robustness. As far as we know, this work represents the first attempt to leverage a hybrid robust loss function involving dual domain (both ambient domain and feature space) regularization into a deep learning-based subspace clustering framework. By contrast, the recent work on deep subspace clustering [31] considers L1-norm in the ambient domain only.

We have validated the proposed robust subspace clustering network (RSCN) with dual domain regularization on several well-known datasets including Extended Yale B, COIL20, and COIL100. Experimental results have demonstrated that the proposed method outperforms existing state-of-the-art methods including DSC-Net [15] and Self-Supervised Convolutional Subspace Clustering Network [48] on most datasets. The gain (in terms of error rate reduction) improves as the number of different subjects increases, which verifies the effectiveness of the proposed network architecture. To further verify the robustness of RSCN to noise interference, we have artificially contaminated the dataset by additive white Gaussian noise. The performance gain of our RSCN over other competing methods is dramatically larger in the presence of noise contamination, which justifies the robustness of our approach.

Section snippets

Related work

Deep subspace clustering. Existing works on subspace clustering can be classified into two categories [48]: in the original space and in the feature space. The former includes the classic spectral clustering methods such as [36], [37], [38]. These methods often divide the task of subspace clustering into two subproblems: one is to learn a data affinity matrix; the other is to find the segmentation by spectral clustering. The so-called self-expressive property has been critical to the

Self-expressiveness via sparse coding

For a data point xi in high-dimensional space X, self-expressiveness [15], [48] refers to the property that xi can be expressed by the linear combination of other data points in a union of subspaces. In other words, it is assumed that data points in the same subspace are highly correlated; while data points in different subspaces are at most weakly correlated. Therefore, self-expressiveness represents a strategy of exploiting the correlation among data points in a subspace to facilitate the

Experimental results

To evaluate the performance of our RSCN, we have conducted extensive experiments on three standard datasets: one face image dataset (Extended Yale B) and two object image data sets (COIL20 and COIL 100). We have compared our proposed algorithm with the following baseline methods including Low Rank Representation (LRR) [22], Low Rank Subspace Clustering (LRSC) [37], Sparse Subspace Clustering (SSC) [7], Kernel Sparse Subspace Clustering (KSSC) [29], SSC by Orthogonal Matching Pursuit(SSC-OMP)

Conclusion

In this paper, we have presented a robust subspace clustering network (RSCN) consisting of a hybrid loss function and a dual domain regularization strategy. RSCN can supervise the joint training of self-supervision (implemented by spectral clustering) and self-expression (implemented by an auto-encoder) modules in an end-to-end manner. Moreover, inspired by dual-domain regularization, we have proposed a composite/hybrid loss function of L1 and Frobenius norms in both ambient domain and feature

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Acknowledgments

This work was supported in part by the National Key R&D Program of China under Grant 2018AAA0101400 and the Natural Science Foundation of China under Grant 61991451, Grant 61632019, Grant 61621005, and Grant 61836008.

References (50)

  • X. Guo et al.

    Robust low-rank subspace segmentation with finite mixture noise

    Pattern Recognit.

    (2019)
  • R. Vidal et al.

    Low rank subspace clustering (lRSC)

    Pattern Recognit. Lett.

    (2014)
  • R. Agrawal et al.

    Automatic subspace clustering of high dimensional data for data mining applications

    (1998)
  • E.J. Candès et al.

    Robust principal component analysis?

    J. ACM

    (2011)
  • Y. Chen et al.

    Stochastic sparse subspace clustering

    Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition

    (2020)
  • Z. Dang et al.

    Multi-scale fusion subspace clustering using similarity constraint

    Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition

    (2020)
  • M. Elad

    Sparse and Redundant Representations: From Theory to Applications in Signal and Image Processing

    (2010)
  • E. Elhamifar et al.

    Sparse subspace clustering

    2009 IEEE Conference on Computer Vision and Pattern Recognition

    (2009)
  • E. Elhamifar et al.

    Sparse subspace clustering: algorithm, theory, and applications

    IEEE Trans. Pattern Anal. Mach.Intell.

    (2013)
  • I. Goodfellow et al.

    Deep Learning

    (2016)
  • I. Goodfellow et al.

    Generative adversarial nets

    Advances in Neural Information Processing Systems

    (2014)
  • B.D. Haeffele et al.

    A critique of self-expressive deep subspace clustering

    International Conference on Learning Representations (ICLR)

    (2021)
  • R. He et al.

    Information theoretic subspace clustering

    IEEE Trans. Neural Netw. lLearn.Syst.

    (2015)
  • R. Heckel et al.

    Robust subspace clustering via thresholding

    IEEE Trans. Inf. Theory

    (2015)
  • P. Ji et al.

    Efficient dense subspace clustering

    IEEE Winter Conference on Applications of Computer Vision

    (2014)
  • P. Ji et al.

    Deep subspace clustering networks

    Advances in Neural Information Processing Systems

    (2017)
  • M. Kheirandishfard et al.

    Multi-level representation learning for deep subspace clustering

    Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision

    (2020)
  • Y. LeCun et al.

    Deep learning

    Nature

    (2015)
  • K.-C. Lee et al.

    Acquiring linear subspaces for face recognition under variable lighting

    IEEE Trans. Pattern Anal. Mach.Intell.

    (2005)
  • C.-G. Li et al.

    Structured sparse subspace clustering: a joint affinity learning and subspace clustering framework

    IEEE Trans. Image Process.

    (2017)
  • X. Li et al.

    Robust subspace clustering by cauchy loss function

    IEEE Trans. Neural Netw. Learn.Syst.

    (2018)
  • Y. Li et al.

    Robust subspace clustering with independent and piecewise identically distributed noise modeling

    Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition

    (2019)
  • G. Liu et al.

    Robust subspace segmentation by low-rank representation

    Proceedings of the 27th International Conference on Machine Learning (ICML-10)

    (2010)
  • L. Van der Maaten et al.

    Visualizing data using T-SNE

    J. Mach. Learn. Res.

    (2008)
  • S.A. Nene, S.K. Nayar, H. Murase, et al., Columbia object image library...
  • Cited by (3)

    Editor: ”Jiwen Lu”.

    View full text