Elsevier

Neurocomputing

Volume 216, 5 December 2016, Pages 342-350
Neurocomputing

A novel multi-view clustering method via low-rank and matrix-induced regularization

https://doi.org/10.1016/j.neucom.2016.08.014Get rights and content

Abstract

Multi-view clustering algorithms have shown promising performance in various applications over the last few decades. Most of them, however, do not adequately take noises and correlation among multiple views into account, which may degrade the clustering performance. In this paper, we propose a novel multi-view clustering method to address these issues. In specific, we construct a low-rank consensus matrix and a sparse error matrix from each similarity matrix corresponding to each view. Furthermore, a matrix-induced regularization term is incorporated to reduce the redundancy and enhance the diversity among different views. The augmented Lagrangian multiplier algorithm is adopted to solve the resultant optimization problem. Comprehensive experiments are conducted to verify the effectiveness of the proposed algorithm. Results demonstrate that our algorithm outperforms several state-of-the-art ones on both synthetic and benchmark data sets.

Introduction

Multi-view clustering, which aims to partition data points into groups with similar ones by applying information from different views, has shown great performance in recent years. Multi-view data are common in real-life problems. For example, web data can be represented in features extracted from text and hyperlinks, whereas books and papers can be translated into different languages. Several multi-view clustering algorithms have been proposed. However, identifying the underlying relationships between data points is difficult because of noises and redundant information.

Robust clustering methods have been proposed to handle noises. The idea of low-rank has shown its powerfulness on robustness. Ye et al. [25] propose a robust late fusion method that decomposes each original score matrix from individual models into a common rank-2 matrix and sparse deviation errors. Pan et al. [22] utilize a low-rank constraint for rank aggregation. Similarly, Xia et al. [24] propose a robust multi-view spectral clustering (RMSC) method, which extracts a consensus transition probability matrix with a low-rank constraint and a sparse constraint on each residual error matrix. Hong et al. [13] integrate different Low Rank Representation (LRR) affinity matrices to form hypergraph Laplacian matrix. However, existing robust clustering methods regard the information from each view indiscriminately, and the redundant information between views degrades the clustering performance of these methods. The problems arising from the neglect of correlation between different views are listed as follows: (i) Information from similar views would be redundant and conquer much attention in clustering. (ii) Information from a special view would be suppressed because of its low occupancy rate.

To address the correlation between different views, some researchers [4], [16] have shown that the independence of different views helps in multi-view learning. Previous studies propose that a high independence translates to a high diversity of two variables [20], [21]. All types of dependence measures have the same intrinsic factor that expresses the similarity between views. Cao et al. [2] extend the existing subspace clustering into the multi-view domain and incorporates the Hilbert Schmidt Independence Criterion (HSIC) as a diversity term to explore the complementarity of multi-view representations. Liu et al. [19] propose a multiple kernel k-means clustering method with a matrix-induced regularization that reduces redundancy and enhances diversity. Kumar and Daumé [15] apply a co-training approach for multi-view spectral clustering to explore intra-cluster information. Zhou et al. [32] utilize Kullback–Leibler (KL) divergence to guide clustering ensemble and achieve outstanding performance. However, these multi-view methods have a finite ability to resist noises.

Considering that diversity increases clustering performance, we construct a matrix-induced regularization term for our methods. To recap the powerfulness of robust methods, we also provide a low-rank constraint to deal with noises. Therefore, a novel multi-view clustering method via low-rank and matrix-induced regularization (MCM-LRMIR) is proposed in this paper. The main contribution of our method is that we introduce matrix-induced regularization term into low-rank framework, which not only holds the advantages of low-rank framework in robustness, but also ensures the diversity between different views. First, we apply Gaussian kernels to define the similarity matrix. As Xia et al. [24] explicitly handle noises in transition probability matrices from different views. Hence, we apply low-rank and sparse decomposition constraints which are similar to the front method and introduce a matrix-induced regularization to guide the objective function. Finally, we propose an alternative optimization procedure based on the augmented Lagrangian multiplier (ALM) scheme to solve the objective function [17]. Experimental results show that this procedure converges in several iterations. In addition, experimental results on both synthetic and benchmark datasets of multi-view clustering demonstrate that the proposed method outperforms several state-of-the-art ones in the literature.

Section snippets

Related work

Multi-view clustering has been a hot topic in recent years. Related algorithms can be roughly grouped into two main categories: (i) Construct a common feature representation for all views before or during clustering. (ii) Integrate the clustering results from each view to obtain a final one.

Existing methods belonging to the first category have diverse ways of obtaining a common feature representation. With a priori hypothesis that the optimal kernel can be obtained by the linear combination of

Single view clustering

In single-view clustering, MKSC is similar to spectral clustering via Markov chains [30]. Let X={x1,x2,,xn}Rd×n be a set of n data points. Define a similarity matrix SRn×n to denote the similarity between samples. In general, Gaussian kernels are used to define similarity matrix: Sij=exp(xixj22δ2) where ·2 denotes the ℓ2 norm and δ2 denotes average Euclidean distance over all pairs of data points. Let G=(V,E,S) be a weighted graph with vertex set V, edge set E, and the similarity

The proposed algorithm

Existing algorithms have shown promising clustering performance. However, most of them have not considered the diversity between information from different views adequately [5], [7], [14], [24], [26]. For example, Eq. (3) in RMSC combines all views indiscriminately. The optimization procedure would render the error matrices from similar views small and sparse because similar views lead to similar error matrices and show advantages in numbers. In this aspect, similar views are dominant and over

Optimization

The optimization in Eq. (6) is still difficult because of the trace norm and the ℓ1 norm. In this section, we apply the ALM scheme [17] to solve it.

First, we replace X^ with J in the function and add their equality as a constraint. minJ,X^,E(i),ωJ+λi=1mωiE(i)1+βωHωs.t.X(i)=X^+E(i),X^0,X^1=1,ω1=1,J=X^.Then, we obtain the corresponding augmented Lagrange function: L(J,X^,E(i),ω)=J+λi=1mωiE(i)1+i=1mY(i),X^+E(i)X(i)+μ2i=1mX^+E(i)X(i)F2+Z,X^J+μ2X^JF2+ωHωs.t.X^0,X^1=1,ω

Experiments

In this section, we evaluate our algorithm in both synthetic and benchmark data sets. Our proposed algorithm shows better performance compared with other methods. We use several real-world and synthetic data sets to show the performance of our proposed algorithm. The number of clusters is set to the true number of classes for all the data sets.

Conclusion

We propose a novel method (MCM-LRMIR) that not only retains the benefit of consensus matrix recovery methods but also considers the diversity of multi-views to guide clustering. Our algorithm is easily complemented with an alternative optimization in which most subproblems can be solved using off-the-shelf methods. Experiments on both synthetic and benchmark data sets show that our algorithm outperforms state-of-the-art ones. Parameter analysis shows that our proposed method is insensitive to

Acknowledgments

This work was supported by the National Natural Science Foundation of China (Project no. 61403405).

Yang Zhao received the B.S. degree in computer science from the National University of Defense Technology, Changsha, China, in 2014. He is currently working toward the M.S. degree at the National University of Defense Technology. His research interests include machine learning and computer vision.

References (33)

  • Chaoqun Hong et al.

    Multi-view hypergraph learning by patch alignment framework

    Neurocomputing

    (2013)
  • K. Zeng et al.

    Image clustering by hyper-graph regularized non-negative matrix factorization

    Neurocomputing

    (2014)
  • Sihang Zhou et al.

    Random Fourier extreme learning machine with 2,1-norm regularization

    Neurocomputing

    (2016)
  • Jian-Feng Cai et al.

    A singular value thresholding algorithm for matrix completion

    SIAM J. Optim.

    (2010)
  • Xiaochun Cao, Changqing Zhang, Huazhu Fu, Si Liu, Hua Zhang, Diversity-induced multi-view subspace clustering,...
  • Xiaochun Cao, Changqing Zhang, Huazhu Fu, Si Liu, Hua Zhang, Diversity-induced multi-view subspace clustering, in:...
  • Kamalika Chaudhuri, Sham M. Kakade, Karen Livescu, Karthik Sridharan, Multi-view clustering via canonical correlation...
  • Liang Du, Peng Zhou, Lei Shi, Hanmo Wang, Mingyu Fan, Wenjian Wang, Yi-Dong Shen, Robust multiple kernel k-means...
  • John Duchi, Shai Shalev-Shwartz, Yoram Singer, Tushar Chandra, Efficient projections onto the l 1-ball for learning in...
  • Mehmet Gönen, Adam A. Margolin, Localized data fusion for kernel k-means clustering with application to cancer biology,...
  • Derek Greene, Pádraig Cunningham, A matrix factorization approach for integrating multiple data views, in:Machine...
  • Arthur Gretton, Olivier Bousquet, Alex Smola, Bernhard Schölkopf, Measuring statistical dependence with Hilbert–Schmidt...
  • Arthur Gretton, Karsten M. Borgwardt, Malte J. Rasch, Bernhard Schölkopf, Alexander J. Smola, A kernel method for the...
  • Chaoqun Hong et al.

    Image-based three-dimensional human pose recovery by multiview locality-sensitive sparse retrieval

    IEEE Trans. Ind. Electron.

    (2015)
  • Chaoqun Hong et al.

    Multimodal deep autoencoder for human pose recovery

    IEEE Trans. Image Process.

    (2015)
  • Hsin-Chien Huang et al.

    Multiple kernel fuzzy clustering

    IEEE Trans. Fuzzy Syst.

    (2012)
  • Cited by (12)

    • An active three-way clustering method via low-rank matrices for multi-view data

      2020, Information Sciences
      Citation Excerpt :

      In contrast, the strategy in the ATCLM is to find the informative objects from the fringe regions by taking the advantage of three-way representation. To show the performance of dealing with multi-view data, we test the proposed method UCLM, the BSV, the FeatCon, the AWC, the CRSC [10], the WCFS [32] and the LRMIR [44] on the seven datasets. For fair comparison, the proposed UCLM is executed in this experiment because the compared algorithms are unsupervised methods.

    • Auto-weighted multi-view clustering via kernelized graph learning

      2019, Pattern Recognition
      Citation Excerpt :

      In the past decade, several methods have been proposed to solve the multi-view clustering problem. In general, existing multi-view clustering methods can be roughly classified into two categories including subspace approaches which aim to explore the hidden common subspace shared by all views [14–19], and graph-based methods which were designed based on the traditional spectral learning models [20–24]. However, most existing multi-view clustering methods still suffer from the following drawbacks: (1) These methods are sensitive to outliers and noisy data, thus impair the final clustering performance greatly; (2) Nonlinear relationships exist in real-world dataset have not been effectively searched by most existing methods. (

    • Diversity-induced fuzzy clustering

      2019, International Journal of Approximate Reasoning
    • Multi-view metric learning based on KL-divergence for similarity measurement

      2017, Neurocomputing
      Citation Excerpt :

      This section will provide authors a brief comprehension of multi-view learning and distance metric learning. Meanwhile, researchers have proposed several related works [30–38] in this field. Zhai et al. proposed a multi-view metric learning [30] method which aims to reveal the shared latent feature space of the multiview observations by embodying global consistency constraints and preserving local geometric structures.

    View all citing articles on Scopus

    Yang Zhao received the B.S. degree in computer science from the National University of Defense Technology, Changsha, China, in 2014. He is currently working toward the M.S. degree at the National University of Defense Technology. His research interests include machine learning and computer vision.

    Yong Dou received his B.S., M.S., and Ph.D. degrees from National University of Defense Technology in 1989, 1992 and 1995. His research interests include high performance computer architecture, high performance embedded microprocessor, reconfigurable computing, machine learning, and bioinformatics. He is a member of the IEEE and the ACM.

    Xinwang Liu received the M.S. and Ph.D. degree from National University of Defense Technology, China in 2008 and 2013, respectively. From January 2014, he works as a research assistant at National Laboratory for Parallel and Distributed Processing, National University of Defense Technology, Changsha, China. His research interests focus on designing algorithms on kernel learning, feature selection and multi-view clustering.

    Teng Li received the B.S. degree and the M.S. degree in computer science from the National University of Defense Technology, Changsha, China, in 2013 and 2015, respectively. He is currently working toward the Ph.D. degree at the National University of Defense Technology. His research interests include machine learning and computer vision.

    View full text