Skip to main content
Log in

Adaptive graph regularized non-negative Tucker decomposition for multiway dimensionality reduction

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

Non-negative Tucker decomposition (NTD) is a powerful tool for data representation to capture rich internal structure information from non-negative high-dimensional tensor data. Arguing that NTD methods often give global-like information, graph constraint has been introduced to capture the important local nonlinear structure of data. However, existing methods generally use fixed graphs and lack the ability to adaptively learn the optimal graph that best benefits the learning task at hand. In this paper, we propose an Adaptive Graph Regularized Non-negative Tucker Decomposition (AGRNTD) model. Not only is the new model able to capture the global multilinear structure of tensor data, but also it adaptively learns the optimal graph to capture local manifold information. An updating rule is designed to optimize the new model with the guarantee of local convergence. By mapping and visualizing the features, our method exhibits better feature extraction compared with other algorithms. The clustering results on five real-world datasets demonstrate the effectiveness and robustness of our method.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

Availability of data and materials

Data sharing not applicable to this article as no datasets were generated or analyzed during the current study.

References

  1. Belkin M, Niyogi P (2003) Laplacian eigenmaps for dimensionality reduction and data representation. Neural Comput 15:1373–1396

    Article  Google Scholar 

  2. Boyd SP, Vandenberghe L (2006) Convex optimization. IEEE Trans Autom Control 51:1859–1859

    Article  Google Scholar 

  3. Cai D, He X, Han J et al (2011) Graph regularized nonnegative matrix factorization for data representation. IEEE Trans Pattern Anal Mach Intell 33(8):1548–60

    Article  Google Scholar 

  4. Calvi GG, Moniri A, Mahfouz M, et al (2019) Tucker tensor layer in fully connected neural networks. ArXiv abs/1903.06133

  5. Donoho D, Grimes C (2003) Hessian eigenmaps: locally linear embedding techniques for high-dimensional data. Proc Natl Acad Sci USA 100:5591–5596

    Article  MathSciNet  Google Scholar 

  6. Fang X, Han N, Wong W et al (2019) Flexible affinity matrix learning for unsupervised and semisupervised classification. IEEE Trans Neural Netw Learn Syst 30:1133–1149

  7. Feng Y, Zhou G (2020) Orthogonal random projection for tensor completion. IET Comput Vis 14:233–240

    Article  Google Scholar 

  8. Huang H, Ma Z, Zhang G (2022) Dimensionality reduction of tensors based on manifold-regularized tucker decomposition and its iterative solution. Int J Mach Learn Cybern 13:509–522

    Article  Google Scholar 

  9. Jiang B, Ding C, Tang J et al (2019) Image representation and learning with graph-laplacian tucker tensor decomposition. IEEE Trans Cybern 49:1417–1426

    Article  Google Scholar 

  10. Khan G, Hu J, Li T et al (2022) Multi-view data clustering via non-negative matrix factorization with manifold regularization. Int J Mach Learn Cybern 13:677–689

    Article  Google Scholar 

  11. Kim Y, Choi S (2007) Nonnegative tucker decomposition. 2007 IEEE Conference on Computer Vision and Pattern Recognition 1–8

  12. Kolda T, Bader B (2009) Tensor decompositions and applications. SIAM Rev 51:455–500

    Article  MathSciNet  Google Scholar 

  13. Kossaifi J, Lipton ZC, Khanna A, et al (2020) Tensor regression networks. ArXiv abs/1707.08308

  14. Kuhn H (1955) The hungarian method for the assignment problem. Naval Research Logistics Quarterly 2:83–97

    Article  MathSciNet  Google Scholar 

  15. Lebedev V, Ganin Y, Rakhuba M, et al (2015) Speeding-up convolutional neural networks using fine-tuned cp-decomposition. CoRR abs/1412.6553

  16. Lee D, Seung HS (2000) Algorithms for non-negative matrix factorization. In: NIPS

  17. Li X, Ng M, Cong G et al (2017) Mr-ntd: manifold regularization nonnegative tucker decomposition for tensor data dimension reduction and representation. IEEE Trans Neural Netw Learn Syst 28:1787–1800

    Article  MathSciNet  Google Scholar 

  18. Martínez A, Kak A (2001) Pca versus lda. IEEE Trans Pattern Anal Mach Intell 23:228–233

    Article  Google Scholar 

  19. Nie F, Wang X, Huang H (2014) Clustering and projected clustering with adaptive neighbors. Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining

  20. Nie F, Wang X, Jordan MI, et al (2016) The constrained laplacian rank algorithm for graph-based clustering. In: AAAI

  21. Oseledets I (2011) Tensor-train decomposition. SIAM. J Sci Comput 33:2295–2317

    MathSciNet  Google Scholar 

  22. Paatero P (1999) The multilinear engine–a table-driven, least squares program for solving multilinear problems, including the n-way parallel factor analysis model. J Comput Graph Stat 8:854–888

    MathSciNet  Google Scholar 

  23. Peng Y, Long Y, Qin F, et al (2019) Flexible non-negative matrix factorization with adaptively learned graph regularization. ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) pp 3107–3111

  24. Qiu Y, Zhou G, Wang Y, et al (2020) A generalized graph regularized non-negative tucker decomposition framework for tensor data representation. IEEE Transactions on Cybernetics

  25. Roweis S, Saul L (2000) Nonlinear dimensionality reduction by locally linear embedding. Science 290(5500):2323–6

    Article  Google Scholar 

  26. Shi J, Malik J (2000) Normalized cuts and image segmentation. IEEE Trans Pattern Anal Mach Intell 22:888–905

    Article  Google Scholar 

  27. Tenenbaum J, de Silva V, Langford J (2000) A global geometric framework for nonlinear dimensionality reduction. Science 290(5500):2319–23

    Article  Google Scholar 

  28. Wang C, He X, Bu J et al (2011) Image representation using laplacian regularized nonnegative tensor factorization. Pattern Recognit 44:2516–2526

    Article  Google Scholar 

  29. Wu J, Xie X, Nie L, et al (2020) Unified graph and low-rank tensor learning for multi-view clustering. In: AAAI

  30. Yi Y, Wang J, Zhou W et al (2019) Joint graph optimization and projection learning for dimensionality reduction. Pattern Recognit 92:258–273

    Article  Google Scholar 

  31. Yin M, Gao J, Lin Z (2016) Laplacian regularized low-rank representation and its applications. IEEE Trans Pattern Anal Mach Intell 38:504–517

    Article  Google Scholar 

  32. Yu J, Li C, Zhao Q, et al (2019) Tensor-ring nuclear norm minimization and application for visual: Data completion. ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) pp 3142–3146

  33. Yu Y, Zhou G, Zheng N, et al (2022) Graph regularized nonnegative tensor ring decomposition for multiway representation learning. IEEE Transactions on Cybernetics PP

  34. Zhai H, Zhang H, Zhang L et al (2019) Laplacian-regularized low-rank subspace clustering for hyperspectral image band selection. IEEE Trans Geosci Remote Sens 57:1723–1740

    Article  Google Scholar 

  35. Zhao Q, Zhou G, Xie S, et al (2016) Tensor ring decomposition. ArXiv abs/1606.05535

  36. Zhao YY, Jiao CN, Wang ML et al (2021) Htrpca: hypergraph regularized tensor robust principal component analysis for sample clustering in tumor omics data. Interdisciplinary Sciences: Computational Life Sciences 14:22–33

    Google Scholar 

  37. Zhou G, Cichocki A, Zhao Q et al (2014) Nonnegative matrix and tensor factorizations: an algorithmic perspective. IEEE Signal Process Mag 31:54–65

    Article  Google Scholar 

Download references

Funding

This work is supported in part by the National Defense Basic Scientific Research Project (JCKY2020903B002), the Natural Science Foundation of China under Grant 62073087, Grant 62071132, and Grant 62203124, and in part by the Guangdong Natural Science Foundation under Grant 2023A1515012916.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Guoxu Zhou.

Ethics declarations

Competing interests

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Chen, D., Zhou, G., Qiu, Y. et al. Adaptive graph regularized non-negative Tucker decomposition for multiway dimensionality reduction. Multimed Tools Appl 83, 9647–9668 (2024). https://doi.org/10.1007/s11042-023-15622-4

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-023-15622-4

Keywords

Navigation