Elsevier

Neurocomputing

Volume 273, 17 January 2018, Pages 78-88
Neurocomputing

Manifold NMF with L21 norm for clustering

https://doi.org/10.1016/j.neucom.2017.08.025Get rights and content

Abstract

Nonnegative matrix factorization has been widely used in data mining and machine learning fields as a clustering algorithm. The standard nonnegative matrix factorization algorithm utilizes the sum of squared error to measure the quality of factorization, however, the noise and outliers in the dataset will reduce the performance of algorithm significantly. This paper proposes a robust manifold nonnegative matrix factorization algorithm based on L21 norm, and the projected gradient method is utilized to obtain the updating rules. The proposed algorithm utilizes the L21 norm to measure the quality of factorization, which is insensitive to the noise and outliers, also it utilizes the geometrical structure of the dataset and considers the local invariance. The experimental results on several data sets and the comparison with other clustering algorithms demonstrate the effectiveness of the proposed algorithm.

Introduction

The “semantic gap” between image visual features and high-level semantic is the bottleneck of the development of image retrieval. The core problem of improving the utilization efficiency of image content and reducing the semantic gap is how to represent the visual features of the image effectively, that is, how to find a reasonable method to show the data [1], [2]. In order to deal with the data information in the matrix, the matrix is usually decomposed [3]. After the decomposition of the matrix, not only the dimension of the original matrix can be greatly reduced, but also the data stored in the original matrix can be summarized and compressed. The matrix decomposition technology has attracted more and more attention. The typical matrix decomposition methods mainly includes: principal component analysis (PCA) [4], linear discriminant analysis (LDA) [5], independent component analysis (ICA) [6], singular value decomposition (SVD) [7], etc. These methods usually make decomposition or linear transformation on the original data matrix under certain restrictions. They are common in allowing negative values in the decomposition results, which is correct from the view of calculation, but in practice, negative values are meaningless [8].

Recently, non-negative matrix factorization (NMF) as an effective matrix decomposition method has attracted wide attention. In 1999, Lee and Seung [9] first propose the concept of non-negative matrix factorization on ”Nature” journal. Different from the above-mentioned typical matrix decomposition methods, NMF method is unique in that all elements of the matrix decomposition process are non-negative. NMF method decomposes the original matrix into two non-negative matrix: the left matrix is called basis matrix, the right matrix is called coefficient matrix. Therefore, the column vector in the original matrix can be interpreted as the weighted sum of all the columns in the basis matrix, and the weight coefficients are contained in the coefficient matrix. Such a vector combination has an intuitive semantic interpretation, in line with the concept of ”local constitute a whole” in human thinking [10], [11]. The solution methods of NMF include multiplicative update rule, least square substitution method, and projected gradient method [12]. By adding non-negative constraints, NMF can not only ensure the interpretation of the decomposition results, but also has the advantages of easy implementation, small occupation of storage space, and so on. Therefore, it has a certain practical significance to explore the non-negative matrix factorization.

NMF has received considerable attention since it was presented. In order to improve the efficiency and recognition rate of NMF, many scholars have made a deep research on the NMF algorithm, and obtained a lot of research results. Akashi and Okatani [13] introduce sparse encoding theory into NMF. They apply sparseness constraints on both the basis matrix and coefficient matrix, and propose the sparse non-negative matrix factorization algorithm (SNMF). Liu et al. [14] design the constrained non-negative matrix factorization algorithm (CNMF), which uses the known label information to guide the matrix factorization process. Later, Shu and Zhao [15] improve the CNMF by adding sparse constraints, and propose the constrained non-negative matrix factorization with sparseness (CNMFS). Feng et al. [16] propose the weighted non-negative matrix factorization (WNMF), in which each training sample is accompanied by a non-negative weight. Then we may control the influence of the samples on the decomposition results by adjusting the sample weights. Bucak and Gunsel [17] combine incremental learning with non-negative matrix factorization, and propose the incremental non-negative matrix factorization algorithm (INMF), which reduces the computation scale using the idea of block matrix. Kong et al. [18] propose a robust non-negative matrix factorization algorithm based on L21 norm (L21NMF). It overcomes the drawbacks of standard NMF algorithm which is sensitive to the noise and outliers, and improves the robustness of the algorithm. In addition, NMF has been successfully applied to various fields, including text clustering [19], face recognition [20], visual tracking [21], image denoising [22], image retrieval [23], social network analysis [24], etc.

The manifold is a curled space with the properties of Euclidean space in local area. Manifold learning can achieve the dimension reduction by finding the low dimensional manifold embedded in the high-dimensional space. Manifold learning algorithms includes isometric mapping (Isomap), locally linear embedding (LLE), Laplacian eigenmap (LE). They use the local invariant features that two close sample points in high-dimensional space also have short distance in the low dimensional manifold [25]. Cai et al. [26] propose the graph regularized non-negative matrix factorization algorithm (GNMF), which combines manifold learning and NMF. GNMF algorithm considers the geometric structure contained in the original data and will preserve the local geometric information of data sets during the decomposition process. Jiang et al. [27] propose the graph regularized non-negative matrix factorization algorithm with sparse constraints (GNMFSC), which not only takes the geometrical information of data into account, but also impose sparse constraints on the coefficient matrix, so the face images after decomposition have higher recognition rate. But in GNMF algorithm, the quality of matrix decomposition is measured by the Frobenius norm, which is easily affected by the noise and abnormal values in the data. In this case, large errors of the objective function may exist and the performance of the algorithm therefore decrease. In order to get more robust clustering results, this paper presents a manifold non-negative matrix factorization algorithm based on L21 norm (MNMFL21), in which the update rules are calculated by projection gradient method. MNMFL21 algorithm uses the L21 norm to measure the quality of matrix decomposition, so it is not sensitive to the data noise. Also it combines the manifold learning with the robust NMF, and uses the local invariance property to detect the geometric structure of data. The main contribution of this paper are: (1) we use L21 norm to improve the GNMF algorithm and propose MNMFL21 algorithm; (3) we give the objective function of MNMFL21 algorithm and deduce its equivalent form to facilitate the calculation; (4) we present the detailed update rules of the non-negative matrix in MNMFL21 algorithm with the projected gradient method.

The rest parts of the article are as follows: Section 2 introduces the standard NMF algorithm, L21NMF algorithm and GNMF algorithm; Section 3 gives the definition of MNMFL21 algorithm and its equivalence objective function; Section 4 uses the gradient projection method to solve the update rules of MNMFL21 algorithm; Section 5 shows the experimental results of the MNMFL21 algorithm on several data sets and illustrates the validity of MNMFL21 algorithm by comparing it with other clustering algorithms; finally, we summarize the work in this paper.

Section snippets

Related works

In this section, we briefly review the standard NMF algorithm, L21NMF algorithm and GNMF algorithm. The input matrix is defined as X={x1,x2,,xn}, where each p dimensional column vector xi represents a non-negative data sample.

The proposed MNMFL21 algorithm

In this section, we will give the definition of the proposed MNMFL21 algorithm. For convenience of the calculation of update rules, we derive the equivalent objective function of MNMFL21 algorithm.

Update rules for MNMFL21 algorithm

In this section, we first use the projection gradient method to solve the general constrained optimization problem, and then use the projected gradient method to calculate the update rules of F and G in MNMFL21 algorithm.

Contrast algorithms

We compare the MNMFL21 algorithm with the following algorithms:

  • (1)

    k-means algorithm [28]: the classical partition clustering algorithm that obtains the best clustering results by minimizing the distance between the data points and their cluster centers.

  • (2)

    PCA algorithm [29]: using PCA to reduce the dimension of the original data set, and then clustering on the low dimensional data with k-means algorithm.

  • (3)

    SNMF algorithm [13]: representing the data set by sparse coding method, and using F norm to

Conclusion

This paper proposes the MNMFL21 algorithm, which is a robust manifold NMF clustering algorithm based on L21 norm. This algorithm inherits the advantages of L21NMF and GNMF algorithms. It uses the L21 norm to measure the quality of matrix decomposition, and considers the manifold structure and local invariance of the data. Through introducing the Laplacian matrix, the low dimensional space G obtained by the MNMFL21 algorithm can accurately reflect the geometric structure of the data in high

Acknowledgments

This project is supported by the Fundamental Research Funds for the Central Universities (2014ZDPY23) and the Priority Academic Program Development of Jiangsu Higher Education Institutions (PAPD).

Baolei Wu received the M.S. degree in Computer Application Technology from China University of Mining and Technology, Xuzhou, China, in 2010. His current research interests include saliency detection and image processing.

References (31)

  • S.S. Bucak et al.

    Incremental subspace learning via non-negative matrix factorization

    Pattern Recognit.

    (2009)
  • LiuQ. et al.

    Projective nonnegative matrix factorization for social image retrieval

    Neurocomputing

    (2016)
  • R. Hettiarachchi et al.

    Multi-manifold LLE learning in pattern recognition

    Pattern Recognit.

    (2015)
  • ZhaoJ. et al.

    Recovering seabed topography from sonar image with constraint of sounding data

    J. China Univ. Min. Technol.

    (2017)
  • HuZ. et al.

    Sparse principal component analysis via rotation and truncation

    IEEE Trans. Neural Netw. Learn. Syst.

    (2016)
  • Cited by (30)

    • Robust multi-view subspace enhanced representation based on collaborative constraints and HSIC induction

      2023, Engineering Applications of Artificial Intelligence
      Citation Excerpt :

      Furthermore, it not only extracts useful identification information in the data but also effectively identifies the optimal solution. Existing studies have shown that the manifold information in data plays an important role in the performance improvement of clustering methods (Cai et al., 2011; Lu et al., 2020; Wu et al., 2018; Wen et al., 2018; Zheng et al., 2010). Further, useful information from real-world data often remains hidden in the low ranks of the data (Li et al., 2017; Vidal and Favaro, 2014).

    View all citing articles on Scopus

    Baolei Wu received the M.S. degree in Computer Application Technology from China University of Mining and Technology, Xuzhou, China, in 2010. His current research interests include saliency detection and image processing.

    Enyuan Wang is a professor in the College of Safety Engineering at China University of Mining and Technoloty, Xuzhou, China. He received his Ph.D. degree from the same university in 1997. And he worked as a postdoctor in college of electrical engineering at China University of Mining and Technology from December 1997 to November 1999. His current research interests are rock mechanics and coal mining induced disasters monitoring and forcasting.

    Zhen Zhen is a graduate student in School of Safety Engineering at China University of Mining and Technology, Xuzhou, China. His current research are detection technology and automatic equipment.

    Wei Chen received the Ph.D degree in communications and information systems from China University of Mining and Technology, Beijing, China, in 2008. In 2008, he joined the School of Computer Science and Technology, China University of Mining and Technology at Xuzhou, where he is currently a Professor. His research interests include machine learning, image processing, and wireless communications.

    Pengcheng Xiao is an assistant professor in the Department of Mathematics at University of Evansville. He received his Ph.D. degree from the University of Texas at Arlington in 2015. His current research interests are computational neuroscience and biomathematics.

    View full text