Skip to main content
Log in

Component preserving laplacian eigenmaps for data reconstruction and dimensionality reduction

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

Laplacian Eigenmaps (LE) is a widely used dimensionality reduction and data reconstruction method. When the data has multiple connected components, the LE method has two obvious deficiencies. First, it might reconstruct each component as a single point, resulting in loss of information within the component. Second, it only focuses on local features but ignores the location information between components, which might cause the reconstructed components to overlap or to completely change their relative positions. To solve these two problems, this article first modifies the optimization objective of the LE method, by characterizing the relative positions between components of data with the similarity between high-density core points, and then solves the optimization problem by using a gradient descent method to avoid the over-compression of data points in the same connected component. A series of experiments on synthetic data and real-world data verify the effectiveness of the proposed method.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Algorithm 1
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14

Similar content being viewed by others

Data availability and access

The datasets generated and analysed during the current study are available from the corresponding author on reasonable request.

References

  1. Nie F, Wang Z, Wang R, Li X (2022) Adaptive local embedding learning for semi-supervised dimensionality reduction. IEEE Trans Knowl Data Eng 34(10):4609–4621

    Article  Google Scholar 

  2. Das S, Pal NR (2022) Nonlinear dimensionality reduction for data visualization: An unsupervised fuzzy rule-based approach. IEEE Trans Fuzzy Syst 30(7):2157–2169

    Article  Google Scholar 

  3. Wang R, Nie F, Hong R, Chang X, Yang X, Yu W (2017) Fast and orthogonal locality preserving projections for dimensionality reduction. IEEE Trans Image Process 26(10):5019–5030

    Article  MathSciNet  MATH  Google Scholar 

  4. Leem S, Park T (2017) An empirical fuzzy multifactor dimensionality reduction method for detecting gene-gene interactions. BMC Genomics 18(2):1–12

    Google Scholar 

  5. Abdi H, Williams LJ (2010) Principal component analysis. Wiley Interdisciplinary Rev: Comput Stat 2(4):433–459

    Article  Google Scholar 

  6. Balakrishnama S, Ganapathiraju A (1998) Linear discriminant analysis-a brief tutorial. Inst Signal Inf Process 18(1998):1–8

    Google Scholar 

  7. Wang S, Lu J, Gu X, Du H, Yang J (2016) Semi-supervised linear discriminant analysis for dimension reduction and classification. Pattern Recognit 57:179–189

    Article  MATH  Google Scholar 

  8. Schölkopf B, Smola A, Müller K-R (1997) Kernel principal component analysis. In: International conference on artificial neural networks, pp 583–588

  9. Mika S, Ratsch G, Weston J, Scholkopf B, Mullers K-R (1999) Fisher discriminant analysis with kernels. In: IEEE Workshop on neural networks for signal processing, pp 41–48

  10. Cox MA, Cox TF (2008) Multidimensional Scaling. Springer, New York, USA vol, p 1

    MATH  Google Scholar 

  11. Van der Maaten L, Hinton G (2008) Visualizing data using t-SNE. J Mach Learn Res 9(11)

  12. Belkin M, Niyogi P (2003) Laplacian eigenmaps for dimensionality reduction and data representation. Neural Comput 15(6):1373–1396

    Article  MATH  Google Scholar 

  13. McInnes L, Healy J, Melville J (2018) Umap: Uniform manifold approximation and projection for dimension reduction. arXiv:1802.03426

  14. Roweis ST, Saul LK (2000) Nonlinear dimensionality reduction by locally linear embedding. Science 290(5500):2323-2326

  15. Tenenbaum JB, Silva Vd, Langford JC (2000) A global geometric framework for nonlinear dimensionality reduction. Science 290(5500):2319–2323

  16. Tan C, Chen S, Geng X, Ji G (2023) A novel label enhancement algorithm based on manifold learning. Pattern Recognit 135:109189

    Article  Google Scholar 

  17. Valem LP, Pedronette DCG, Latecki LJ (2023) Rank flow embedding for unsupervised and semi-supervised manifold learning. IEEE Trans Image Process 32:2811–2826

    Article  Google Scholar 

  18. Gao Z, Wu Y, Fan X, Harandi M, Jia Y (2023) Learning to optimize on riemannian manifolds. IEEE Trans Pattern Anal Mach Intell 45(5):5935–5952

    Google Scholar 

  19. Liu C, JaJa J, Pessoa L (2018) LEICA: Laplacian eigenmaps for group ICA decomposition of fMRI data. NeuroImage 169:363–373

    Article  Google Scholar 

  20. Ye X, Li H, Imakura A, Sakurai T (2020) An oversampling framework for imbalanced classification based on Laplacian eigenmaps. Neurocomputing 399:107–116

    Article  Google Scholar 

  21. Li B, Li Y-R, Zhang X-L (2019) A survey on Laplacian eigenmaps based manifold learning methods. Neurocomputing 335:336–351

    Article  Google Scholar 

  22. He X, Niyogi P (2003) Locality preserving projections. Adv Neural Inf Process Syst 16

  23. Yu W, Wang R, Nie F, Wang F, Yu Q, Yang X (2018) An improved locality preserving projection with \(l_1\)-norm minimization for dimensionality reduction. Neurocomputing 316:322–331

    Article  Google Scholar 

  24. Wang A, Zhao S, Liu J, Yang J, Liu L, Chen G (2020) Locality adaptive preserving projections for linear dimensionality reduction. Expert Syst Appl 151:113352

    Article  Google Scholar 

  25. Nie F, Zhu W, Li X (2022) Unsupervised large graph embedding based on balanced and hierarchical k-means. IEEE Trans Knowl Data Eng 34(4):2008–2019

    Google Scholar 

  26. Lu X, Long J, Wen J, Fei L, Zhang B, Xu Y (2022) Locality preserving projection with symmetric graph embedding for unsupervised dimensionality reduction. Pattern Recognit 131:108844

    Article  Google Scholar 

  27. Tai M, Kudo M, Tanaka A, Imai H, Kimura K (2022) Kernelized supervised Laplacian Eigenmap for visualization and classification of multi-label data. Pattern Recognit 123:108399

    Article  Google Scholar 

  28. Zhu H, Sun K, Koniusz P (2021) Contrastive Laplacian Eigenmaps. In: Advances in neural information processing systems, pp 5682–5695

  29. Donoho DL, Grimes C (2003) Hessian eigenmaps: Locally linear embedding techniques for high-dimensional data. Proc National Academy Sci 100(10):5591–5596

    Article  MathSciNet  MATH  Google Scholar 

  30. Chen B, Gao Y, Wu S, Pan J, Liu J, Fan Y (2022) Soft adaptive loss based laplacian eigenmaps. Appl Intell 52(1):321–338

    Article  Google Scholar 

  31. Zhang H, Ding Y, Meng H, Ma S, Long Z (2022) Component preserving and adaptive laplacian eigenmaps for data reconstruction and dimensionality reduction. In: International conference on intelligent systems and knowledge engineering, pp 642–649

  32. Von Luxburg U (2007) A tutorial on spectral clustering. Stat Comput 17(4):395–416

    Article  MathSciNet  Google Scholar 

  33. Rodriguez A, Laio A (2014) Clustering by fast search and find of density peaks. Science 344(6191):1492–1496

    Article  Google Scholar 

  34. Long Z, Gao Y, Meng H, Yao Y, Li T (2022) Clustering based on local density peaks and graph cut. Inf Sci 600:263–286

  35. Gao B, Liu X, Yuan Y (2019) Parallelizable algorithms for optimization problems with orthogonality constraints. SIAM J Sci Comput 41(3):1949–1983

    Article  MathSciNet  MATH  Google Scholar 

  36. Xu W, Liu X, Gong Y (2003) Document clustering based on non-negative matrix factorization. In: Annual international ACM SIGIR conference on research and development in information retrieval, pp 267–273

  37. Yang Y, Xu D, Nie F, Yan S, Zhuang Y (2010) Image clustering using local discriminant models and global integration. IEEE Trans Image Process 19(10):2761–2773

    Article  MathSciNet  MATH  Google Scholar 

  38. Steinley D (2004) Properties of the Hubert-Arable Adjusted Rand Index. Psychol Methods 9(3):386–396

    Article  Google Scholar 

Download references

Acknowledgements

This work was supported by the National Natural Science Foundation of China [grant numbers 11871353, 12231007, 61806170], the National Key Research and Development Program of China [grant number 2019YFB1706104], and the Fundamental Research Funds for the Central Universities [grant numbers 2682022ZTPY082, 2682023ZTPY027]. We would also like to thank the anonymous reviewers for their helpful comments and suggestions to improve the article.

Author information

Authors and Affiliations

Authors

Contributions

Hua Meng: Conceptualization, Methodology, Supervision, Writing - Reviewing and Editing. Hanlin Zhang: Software, Visualization, Data curation, Investigation, Validation, Writing-Reviewing and Editing. Yu Ding: Visualization, Data curation. Shuxia Ma: Supervision, Project administration, Writing-Reviewing and Editing. Zhiguo Long: Conceptualization, Formal analysis, Supervision, Writing-Original Draft, Funding acquisition.

Corresponding author

Correspondence to Zhiguo Long.

Ethics declarations

Competing interest

The authors declare that they have no competing interest to this work.

Ethical and informed consent for data used

The data used in this article are either from public datasets or generated by the authors. The research in this article does not involve humans or animals.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Meng, H., Zhang, H., Ding, Y. et al. Component preserving laplacian eigenmaps for data reconstruction and dimensionality reduction. Appl Intell 53, 28570–28591 (2023). https://doi.org/10.1007/s10489-023-05012-6

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10489-023-05012-6

Keywords

Navigation