Elsevier

Information Sciences

Volume 578, November 2021, Pages 574-591
Information Sciences

Iterative rank-one matrix completion via singular value decomposition and nuclear norm regularization

https://doi.org/10.1016/j.ins.2021.07.035Get rights and content

Abstract

Matrix completion is widely used in many fields. In existing matrix completion methods, such as rank minimization and matrix factorization, the hyperparameters must be learned. However, hyperparameter tuning is a time-consuming and tedious process. In this paper, we propose a novel matrix completion method called IMC, i.e., iterative rank-one matrix completion via singular value decomposition (SVD) and nuclear norm regularization. First, we construct a rank-one matrix completion model using nuclear norm regularization. Then, the variables to be optimized in the model are divided into several blocks. Finally, the blocks are iteratively optimized one by one until convergence. For the optimization of each block, we propose an efficient solution scheme based on SVD, in which only the maximum singular value and leading singular vectors of a sparse matrix must be calculated. Further, a nonparametric singular value penalty function is designed to ensure a low-rank completion matrix. In addition, the optimization of each block uses only the values inside the observed entries; hence, no errors are accumulated. Test results shows that the proposed method converges rapidly and outperforms some state-of-the-art methods when applied to grayscale image restoration, recommendation systems, and vote networks.

Introduction

Matrix completion is based on partial observations of a matrix to recover missing entries. It has been widely applied to image processing [1], [2], [3], recommendation systems [4], [5], [6], relational networks [7], [8], compressed sensing [9], sensor networks [10], classification [11], and clustering [12], [13]. Most existing matrix completion methods assume that the target matrix has a low-rank structure, as most real-world data (e.g., images and user ratings) are low rank or approximately low rank [14]. In general, these methods can be broadly classified into three categories: rank minimization (RM), nuclear norm minimization (NNM), and matrix factorization.

Consider an incomplete observation matrix ARm×l. The set Ω denotes the positions of the observed (known) entries in matrix A, and entries outside Ω are often considered missing. The RM method aims to recover a low-rank matrix, X, that exactly matches A over Ω, and describes matrix completion as the problem of minimizing rank(X) subject to PΩ(X)=PΩ(A), where rank(X) is the rank of X and PΩ· denotes a projection operator for extracting the entries of the positions in Ω (see the “Notation” part at the end of this section). However, considering the noise in real-world data, many studies have attempted to find a low-rank matrix that is sufficiently close to the original incomplete matrix in terms of the observed entries, and have reformulated the above-mentioned RM problem as follows:minX12PΩ(X)-PΩ(A)F2s.t.rank(X)τ,where τ is a rank constraint and ·F is the Frobenius norm (see the “Notation” part at the end of this section). Some studies have proposed solutions for this model. For example, Wang et al. [15] extended the orthogonal matching pursuit method from the vector case to the matrix case and completed the matrix by tuning the rank. Vandereycken [16] proposed using low-rank nonlinear conjugated gradients in Euclidean space to perform optimization on manifolds (LRGeomCG); however, he fixed the minimum rank of the matrix. Chen et al. [17] proposed a first-order fast optimization algorithm (FOA) and further transformed the fixed rank into an unfixed rank. To increase the robustness of the algorithm, Le Pendu et al. [18] added an equality constraint to Model (1), defined an augmented Lagrangian function, and then used the alternating direction method of multipliers [19] to update several variables iteratively. Guo et al. [20] filled the missing data in a tensor using the Frank–Wolfe algorithm under the condition of rank constraint. In these methods, the rank constraint setting is critical and can strongly affect the results. However, adjusting the parameters to obtain the optimal rank constraint is a tedious and time-consuming process. Furthermore, in practical applications, the matrix to be recovered is only approximately low rank, not exactly low rank [21]. Moreover, owing to the discrete nature of the rank function, Model (1) is usually an NP-hard problem [22].

Candès and Recht [23] proved that the RM problem can be surrogated by NNM, a convex relaxation of RM. In NNM, the matrix completion substitutes the nuclear norm regularization for the rank constraint and thus becomes the following optimization problem:minX12PΩ(X)-PΩ(A)F2+γX,where · denotes the nuclear norm (see the “Notation” part at the end of this section), γX is a regularization term, and γ is a regularization parameter. For Model (2), the soft-impute method [24] uses data-specific “sparse plus low-rank” structures to process large-scale data. Yao and Kwok [25] proposed an accelerated soft-impute method based on the power method [26]. As the nuclear norm treats each singular value equally and may cause large singular values to prevail, some researchers have used the nonconvex nuclear norm [21], [27] or the weighted nuclear norm [28] in Model (2). Zhang et al. [29] believed that the nuclear norm does not necessarily approximate the matrix rank well in matrix completion; hence, they proposed the concept of the truncated nuclear norm. The truncated nuclear norm was also used by Hu et al. [30] and Hu et al. [31]. However, adjusting the regularization parameters in these methods is a time-consuming process, and the best way to select the appropriate norm according to the application scenario is a complex problem.

Matrix factorization is a classical method of matrix completion. A matrix, X, can be completed via X=Um×rVr×l [32], [33], where r is the rank of X. In some studies, X has been decomposed via singular value decomposition (SVD): X = Um×rΣr×rVl×rT, where Σ is a diagonal matrix comprising singular values and the column vectors of U and V are orthogonal [34], [35]. Because the value of rank r must be adjusted, these methods also suffer from the same shortcomings as previous RM methods. Rank-one matrix completion (R1MC) uses the sum of several rank-one matrices to approximate the observation matrix, i.e., i=1NαiuiviT. Owing to its simple form, R1MC has been widely used in matrix completion [14], [36], [37]. As SVD naturally has the form of R1MC, SVD-based matrix completion is a special case of R1MC. R1MC methods usually adopt an iterative method to ensure completion accuracy, and penalize singular values, including the quantity limit [14] and threshold operation [24], to ensure a low-rank completion matrix. However, existing R1MC methods have the following shortcomings. (i) Partial SVD is required for each iteration; hence, the computational overhead involved in the completion of large-scale matrices is too high. (ii) Each iteration depends on values outside the observed entries calculated by the previous iterations, which may lead to noise accumulation. (iii) Finally, the rank or regularization parameter needs to be adjusted. R1MC is discussed in detail in Section 2.1.

To address the above-mentioned challenges, we propose a novel matrix completion method called IMC, i.e., iterative R1MC via SVD and nuclear norm regularization. This method constructs an R1MC model using nuclear norm regularization and designs an iterative SVD-based solution algorithm according to the model characteristics. The variables to be optimized in the model are divided into several blocks, which are iteratively optimized one by one until convergence. To optimize each block, we propose an efficient solution based on SVD and penalize singular values to ensure a low-rank completion matrix. The main contributions of this study are summarized as follows:

  • We propose a new R1MC method that converts the complex optimization model into multistep subproblems with iterative solutions by dividing the variables to be optimized into blocks. Each subproblem uses only the values inside the observed entries without accumulating errors. More importantly, we analyze the convergence of such iterative solutions.

  • To optimize several variables in each block, many studies have adopted an iterative approach. By analyzing the characteristics of Lagrangian partial derivatives, we propose a one-time optimization scheme based on SVD without iterative calculation. Moreover, our SVD approach needs to calculate only the maximum singular value and leading singular vectors of a sparse matrix, and the calculation cost is relatively low.

  • Penalizing singular values, including the quantity limit and threshold operation, is a common method for ensuring a low-rank completion matrix; however, existing methods require the rank or regularization parameter to be set. Based on the idea of a singular value penalty, we design a nonparametric nonconvex penalty function that can be used in various applications. Some related studies have used other forms of norms as the regularization term, which treat different singular values differently; our penalty function also achieves this effect.

  • The proposed method was applied to grayscale image restoration, recommendation systems, and vote networks. The experimental results verified the robustness and good performance of our method, compared to some state-of-the-art methods.

The remainder of this paper is organized as follows. Section 2 introduces the related work and preliminaries. Section 3 describes the proposed method and discusses the theoretical analysis and refinement. Section 4 presents detailed experimental results. Finally, Section 5 concludes the paper and explores the scope for future work.

Notation: In this paper, vectors and matrices are denoted by lowercase and uppercase bold italic letters, respectively (e.g., vector am and matrix Am×l). Scalars are represented by lowercase italic letters (e.g., a). The relevant symbols and formulas include A=iσi, which is the nuclear norm of matrix A, where σi is the singular value of A. The formula A,B=i1=1mi2=1lAi1i2Bi1i2 denotes the inner product of two matrices A and B of the same dimensions, where Ai1i2 and Bi1i2 are the elements of A and B, respectively. In addition, AF=A,A is the Frobenius norm of matrix A. The set Ω denotes the positions of the observed (known) entries in an observation matrix A, and PΩ· denotes a projection operator (i.e., PΩAi1i2=Ai1i2 if the position (i1,i2) is included in Ω and 0 otherwise). The range outside Ω is represented by Ω-. Additional notation is introduced as it occurs.

Section snippets

Related work and preliminaries

This section first summarizes the relevant R1MC methods by focusing on how they guarantee a low-rank completion matrix. Then, it briefly introduces the block coordinate descent (BCD) method, which is the preliminary basis of our method.

R1MC via SVD and nuclear norm regularization

In this section, we present IMC. First, we describe the rank-one matrix completion model and design an iteration-based solution framework. Then, we analyze the convergence of the iteration. Finally, we present design and refinement schemes for some specific links in the solution framework.

Experiments

This section describes the performance evaluation of IMC. First, we tested its convergence speed and parameter sensitivity. Then, we compared its performance with that of other matrix completion methods in several applications. Finally, we demonstrated the performance of our nonconvex penalty function. All the experiments were performed on a computer with an Intel i7 CPU, 16 GB RAM, and the Windows 10 operating system. The software environment was MATLAB R2017a. In addition, each experiment was

Conclusions

We proposed a novel matrix completion method called IMC, i.e., iterative R1MC via SVD and nuclear norm regularization. We first constructed an R1MC model using nuclear norm regularization and then converted the optimization model into multistep subproblems with iterative solutions by dividing the variables to be optimized into blocks. Further, we analyzed the convergence of such iterative solutions, where each iteration uses only the values inside the observed entries without accumulating

CRediT authorship contribution statement

Kai Xu: Conceptualization, Writing - original draft, Experimentation, Writing - review. Ying Zhang: Formal analysis, Experimentation, Writing - review.. Zhi Xiong: Conceptualization, Formal analysis, Writing - review & editing, Verification proof.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Acknowledgment

Funding: This work was supported in part by the Natural Science Foundation of Guangdong Province [grant numbers 2018A030313438 and 2021A1515012527] and the Scientific Research Project of Colleges and Universities in Guangdong Province [grant number 2020ZDZX3073]. We thank Elsevier Language Editing Services for English language editing.

Kai Xu received the M.S. degree in Computer Technology from Shantou University, China, in 2020. He is an outstanding graduate of Shantou University. His research interests include convex optimization, recommendation system, data mining, and natural language processing.

References (50)

  • S. Osher, Z. Shi, W. Zhu, Low dimensional manifold model for image processing, S.I.A.M. J. Imaging Sci. 10(4) (2017)...
  • Y. Koren et al.

    Matrix factorization techniques for recommender systems

    Computer

    (2009)
  • X. Guan et al.

    Matrix factorization with rating completion: An enhanced SVD model for collaborative filtering recommender systems

    I.E.E.E. Access

    (2017)
  • F. Shang, Y. Liu, J. Cheng, D. Yan, Fuzzy double trace norm minimization for recommendation systems, I.E.E.E. Trans....
  • Z. Zhao, L. Zhang, X. He, W. Ng, Expert finding for question answering via graph regularized matrix completion,...
  • T. Zhang et al.

    Social recommendation algorithm based on stochastic gradient matrix decomposition in social network

    J. Ambient Intell. Human. Comput.

    (2020)
  • K. Xie et al.

    Recover corrupted data in sensor networks: A matrix completion solution

    IEEE Trans. on Mobile Comput.

    (2017)
  • M. Liu, Y. Luo, D. Tao, et al., Low-rank multi-view learning in matrix completion for multi-label image classification....
  • X. Li, M. Chen, Q. Wang, Discrimination-aware projected matrix factorization, I.E.E.E. Trans. Knowl. Data Eng. 32(4)...
  • Q. Shi, H. Lu, Y.M. Cheung, Rank-one matrix completion with automatic rank estimation via L1-norm regularization,...
  • Z. Wang, M.J. Lai, Z. Lu, et al., Orthogonal rank-one matrix pursuit for low rank matrix completion, S.I.A.M. J. Sci....
  • B. Vandereycken, Low-rank matrix completion by Riemannian optimization, S.I.A.M. J. Optim. 23(2) (2013) 1214–1236....
  • M. Le Pendu, X. Jiang, C. Guillemot, Light field inpainting propagation via low rank matrix completion, I.E.E.E. Trans....
  • S. Boyd et al.

    Distributed optimization and statistical learning via the alternating direction method of multipliers

    Found. Trends Mach. Learn.

    (2011)
  • X. Guo, Q. Yao, J.T. Kwok, Efficient sparse low-rank tensor completion using the Frank-Wolfe algorithm. in: Proceedings...
  • Cited by (0)

    Kai Xu received the M.S. degree in Computer Technology from Shantou University, China, in 2020. He is an outstanding graduate of Shantou University. His research interests include convex optimization, recommendation system, data mining, and natural language processing.

    Ying Zhang received the bachelor’s degree in Information Management and Information System from South China Normal University, China, in 2009. She is currently a Senior Engineer with the Information Department, the Second Affiliated Hospital of Shantou University Medical College, China. Her research interests include information security, machine learning, and data mining.

    Zhi Xiong received the Ph.D. degree in Communication and Information System from Wuhan University, China, in 2006. He is currently a Professor with the Department of Computer Science and Technology, Shantou University, China. His research interests include big data, cloud computing, and information security.

    View full text