Alternate algorithms to most referenced techniques of numerical optimization to solve the symmetric rank-R approximation problem of symmetric tensors

https://doi.org/10.1016/j.cam.2022.114792Get rights and content

Abstract

The tensor low-rank approximation and tensor CANDECOMP/PARAFAC (CP) decomposition are useful in various fields such as machine learning, dimension reduction, tensor completion, data visualization etc. A symmetric tensor is a higher order generalization of a symmetric matrix. Comon et al. (2008) show that every symmetric tensor has a symmetric CP-decomposition. In this paper, we study numerical methods of real-valued symmetric CP-decomposition of symmetric tensors. We present an alternate gradient descent method, an alternate BFGS method and an alternate Levenberg–Marquardt (L–M) method for real-valued symmetric rank-R approximation of symmetric tensors. Moreover, we prove the convergence and effectiveness of the algorithms. Numerical examples show that the alternate gradient descent method costs more computing time than the other two methods and the latter two methods have high success rate and good stability.

Introduction

Notations: In the paper, we denote F= as the complex field or R as the real field, respectively. Let s be a positive integer, denote [s]={1,,s} as an integer set from 1 to s. High-order tensors are denoted by calligraphic letters as A,B,C,, matrices are denoted by capital letters as A,B,C,, and vectors are denoted by letters as a,b,c,. Let n=(n1,,nm) be a positive integer vector, T[m]F[n] denotes the set of order-m dimension-(n1,,nm) tensors. For positive integers m and n, T[m]F[n] and S[m]F[n] denote the set of order-m dimension-n square tensors and the set of order-m dimension-n symmetric tensors, respectively.

Given matrix URn×R, we consider the following products: Um=UUUm,Um=UUUm,Um=UUUm,Um=UUUm, where , , , denote Matrix Kronecker product, Matrix Khatri–Rao product, Matrix Hadamard product and vector outer product, respectively.

A tensor is usually denoted as T=(Ti1,,im)T[m]F[n] and represents a multi-array of entries Ti1,,imF, where ij=1,,nj, j=1,,m. F= or R. When n1==nm=n, T is called an order-m dimension-n square tensor. For any order-m dimension-n square tensor S=(Si1,,im)T[m]F[n], if its entries are invariant under any permutation of its indices, then S is called a symmetric tensor [1]. The CANDECOMP/PARAFAC(CP) decomposition can be considered as higher-order generalizations of the matrix singular value decomposition(SVD) and principal component analysis(PCA) [2]. The CP-decomposition factorizes a tensor into a sum of component rank-one tensors. For example, given a tensor TT[m]R[n], we wish to write it as T=k=1Ruk(1)uk(2)uk(m),where R is a positive integer and uk(j)Rnj for k=1,,R, j=1,,m. ‘’ denotes the tensor product or the vector outer product.

Let A represent an order-m dimension-n symmetric tensor. Given a real-valued vector u of length n, we let um denote the order-m dimension-n outer product rank-one symmetric tensor such that (um)i1,,im=ui1uim. Comon et al. [3] show that any real-valued symmetric tensor A can be decomposed as A=k=1Rλkukm,with λkR and ukRn [4].

The tensor low-rank approximation and tensor CP-decomposition are useful in various fields such as machine learning [5], dimension reduction [6], tensor completion [7], data visualization [8]. etc. The alternating least squares (ALS) algorithm is a common numerical algorithm solving the tensor decomposition problem. There are also some other algorithms such as the gradient-based optimization algorithm [9], the conjugate gradient algorithm for nonnegative 3-way tensor factorization [10], the damped Gauss–Newton algorithm for factorization of low-rank real- and complex-valued tensors by deriving a fast inverse for the approximate Hessian [11], randomized algorithms for the low multilinear rank approximations of tensors [12] and the second-order algorithm for fitting the canonical polyadic decomposition with non-least-squares cost [13].

In 2015, Kolda studies the problem of symmetric tensor real-valued decomposition with low-rank structure focusing on both unconstrained and nonnegative by computing the gradients [4]. In 2022, Liu proposes the alternate gradient descent method to solve the symmetric tensor decomposition problem for given orthogonal symmetric tensors or the orthogonal symmetric tensors with a small perturbations [14].

In this paper, we focus on the problem of general symmetric tensor real-valued decomposition. We first introduce an alternate algorithm framework for rank-R symmetric approximation of symmetric tensors, and then we design three numerical algorithms, which are alternate gradient descent algorithm, alternate BFGS algorithm and alternate Levenberg–Marquardt (L–M) algorithm. We also prove the convergence and effectiveness of these algorithms. Numerical experiments show the effectiveness of these algorithms.

The paper is structured as follows. In Section 2, we introduce the basic knowledge of tensor CP-decomposition. In Section 3, we deduce optimization formulation for symmetric rank-R approximation problem of symmetric tensors. In Section 4, we propose an alternate algorithm framework as well as an alternate gradient descent method, an alternate BFGS method and an alternate L–M method. Some numerical examples and experimental results are given in Section 5.

Section snippets

Tensor and symmetric tensor

Definition 2.1

An order-m dimension-n tensor T is an array over the field F indexed by integer tuples i1,,im, i.e., T=Ti1,,imFn1×n2××nm,with ij[nj], j[m]. Denote T[m]F[n] as the space of all such tensors over a field F. If n1=n2==nm=n, then T is called a square tensor.

Definition 2.2

[15]

A tensor ST[m]F[n] is called symmetric if Si1,,im=Siσ(1),iσ(2),,iσ(m),for every permutation σS, where S is the set of all permutations of 1,2,,m. Denote S[m]F[n] as the space of all symmetric tensors over the field F.

Let A,BT[m]R[n]

Symmetric rank-R approximation

In this section, we study the symmetric rank-R approximation problem of symmetric tensors. For a given symmetric tensor AS[m]R[n] and a positive integer number R, the symmetric rank-R approximation problem of the symmetric tensor A is the following optimization problem min12Ak=1RλkukmF2,s.t.λkR,uk2=1,k[R]. Denote Aˆ=k=1Rλkukm,U=u1u2uRRn×R,λ=(λ1,λ2,,λR)RR.Then Aˆ is a symmetric tensor, U is called a factor matrix of Aˆ and Aˆ can also be written as the following Aˆ=[[λ;U,,Um]].

Alternate numerical algorithms

In this section, we introduce three alternate methods for solving the symmetric rank-R approximation problem. In each method, we compute the factor matrix U and coefficients λ alternately. An algorithm framework of alternate methods is as in Algorithm 1.

Numerical examples

In this section, we take some numerical experiments with the alternate gradient descent method, the alternate BFGS method and the alternate L–M method to solve the symmetric rank-R approximation problem of symmetric tensors. In numerical examples, we mainly use tensor-toolbox to program and solve the problem (3.4). The computations are implemented in MATLAB 2019a on a Microsoft Win10 laptop with 16 GB memory and AMD 5pro 4650U CPU. The relative error is defined as related error=Ak=1RλkukmF

Conclusion

In this paper, we discuss the symmetric rank-R approximation problem and the symmetric CP-decomposition problem of real symmetric tensors. We establish the optimization model of rank-R approximation problem of symmetric tensors, and present the gradients of objective function in this model. We propose an alternate gradient descent algorithm, an alternate BFGS algorithm and an alternate L–M algorithm. Numerical experiments show our algorithms are efficient for dealing with real symmetric

Acknowledgments

The authors would like to thank the Principal Editor, Prof. Andre A. Keller, and two anonymous referees for their valuable suggestions, which helped them to improve this manuscript. This work is supported by the National Natural Science Foundation of China (No. 11871472).

References (21)

  • CheM. et al.

    Randomized algorithms for the low multilinear rank approximations of tensors

    J. Comput. Appl. Math.

    (2021)
  • QiL. et al.

    Tensor Eigenvalues and their Applications, Advances in Mechanics and Mathematics, Vol. 39

    (2018)
  • KoldaT.G. et al.

    Tensor decompositions and applications

    SIAM Rev.

    (2009)
  • ComonP. et al.

    Symmetric tensors and symmetric tensor rank

    SIAM J. Matrix Anal. Appl.

    (2008)
  • KoldaT.G.

    Numerical optimization for symmetric tensor decomposition

    Math. Program.

    (2015)
  • BeylkinG. et al.

    Multivariate regression and machine learning with sums of separable functions

    SIAM J. Sci. Comput.

    (2009)
  • RendleS.

    Factorization machines with libFM

    ACM T. Intel. Syst. Tech.

    (2012)
  • ChenY. et al.

    New ALS methods with extrapolating search directions and optimal step size for complex-valued tensor decompositions

    IEEE T. Signal Proces.

    (2011)
  • HongD. et al.

    Generalized canonical polyadic tensor decomposition

    SIAM Rev.

    (2020)
  • AcarE. et al.

    A scalable optimization approach for fitting canonical tensor decompositions

    J. Chemometr.

    (2011)
There are more references available in the full text version of this article.

Cited by (0)

View full text