Abstract
Cognitive computing needs to handle large amounts of data and information. Spectral clustering is a powerful data mining tool based on algebraic graph theory. Because of the solid theoretical foundation and good clustering performance, spectral clustering has aroused extensive attention of academia in recent years. Spectral clustering transforms the data clustering problem into the graph partitioning problem. Cheeger cut is an optimized graph partitioning criterion. To minimize the objective function of Cheeger cut, the eigen-decomposition of p-Laplacian matrix is required. However, the clustering results are sensitive to the selection of similarity measurement and the parameter p of p-Laplacian matrix. Therefore, we propose a self-tuning p-spectral clustering algorithm based on shared nearest neighbors (SNN-PSC). This algorithm uses shared nearest neighbors to measure the similarities of data couples and then applies fruit fly optimization algorithm to find the optimal parameters p of p-Laplacian matrix that leads to better data classification. Experiments show that SNN-PSC algorithm can produce more balanced clusters and has strong adaptability and robustness compared to traditional spectral clustering algorithms.





Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Byun SS, Balashingham I, Vasilakos AV, et al. Computation of an equilibrium in spectrum markets for cognitive radio networks. IEEE Trans Comput. 2014;63(2):304–16.
Jiu MY, Wolf C, Garcia C, et al. Supervised learning and codebook optimization for bag-of-words models. Cogn Comput. 2012;4(4):409–19.
Huang XX, Huang HX, Liao BS, et al. An ontology-based approach to metaphor cognitive computation. Mind Mach. 2013;23(1):105–21.
Bian XY, Zhang TX, Zhang XL, et al. Clustering-based extraction of near border data samples for remote sensing image classification. Cogn Comput. 2013;5(1):19–31.
Zeng S, Huang R, Kang Z, et al. Image segmentation using spectral clustering of Gaussian mixture models. Neurocomputing. 2014;144:346–56.
Mital PK, Smith TJ, Hill RL, et al. Clustering of gaze during dynamic scene viewing is predicted by motion. Cogn Comput. 2011;3(1):5–24.
Gao XB, Deng C, Li XL, Tao DC. Local feature based geometric-resistant image information hiding. Cogn Comput. 2010;2(2):68–77.
Li XW. A new text clustering algorithm based on improved K-means. Journal of Software. 2012;7(1):95–101.
Jia HJ, Ding SF, Xu XZ, Nie R. The latest research progress on spectral clustering. Neural Comput Appl. 2014;24(7–8):1477–86.
Blekas K, Lagaris IE. A spectral clustering approach based on Newton’s equations of motion. Int J Intell Syst. 2013;28(4):394–410.
Liu AH, Poon LKM, Liu TF, et al. Latent tree models for rounding in spectral clustering. Neurocomputing. 2014;144:448–62.
Jia HJ, Ding SF, Meng LH, et al. A density-adaptive affinity propagation clustering algorithm based on spectral dimension reduction. Neural Comput Appl. 2014;25(7–8):1557–67.
Amghibech S. Eigenvalues of the discrete p-Laplacian for graphs. Ars Comb. 2003;67:283–302.
Bühler T, Hein M. Spectral clustering based on the graph p-Laplacian. Proceedings of the 26th international conference on machine learning (ICML 2009), 2009; p. 81–88.
Fiedler M. Algebraic connectivity of graphs. Czechoslov Math J. 1973;23(98):298–305.
MacDonald JK. Successive approximations by the Rayleigh-Ritz variation method. Phys Rev. 1933;43(10):830–3.
Hagen L, Kahng AB. New spectral methods for radio cut partitioning and clustering. IEEE Trans Comput Aided Des Integr Circuits Syst. 1992;11(9):1074–85.
Shi J, Malik J. Normalized cuts and image segmentation. IEEE Trans Pattern Anal Mach Intell. 2000;22(8):888–905.
Amghibech S. Bounds for the largest p-Laplacian eigenvalue for graphs. Discrete Math. 2006;306(21):2762–71.
Hein M, Audibert JY, Von Luxburg U. Graph Laplacians and their convergence on random neighborhood graphs. J Mach Learn Res. 2007;8(12):1325–68.
Arthur Szlam, Xavier Bresson. Total variation and Cheeger cuts. Proceedings of the 27th international conference on machine learning, 2010; p. 233–240.
Nguyen T, Khosravi A, Creighton D, et al. Spike sorting using locality preserving projection with gap statistics and landmark-based spectral clustering. J Neurosci Methods. 2014;238:43–53.
Ding SF, Jia HJ, Zhang LW, Jin FX. Research of semi-supervised spectral clustering algorithm based on pairwise constraints. Neural Comput Appl. 2014;24(1):211–9.
Pan WT. A new fruit fly optimization algorithm: taking the financial distress model as an example. Knowl Based Syst. 2012;26:69–74.
Gosciniak I. A new approach to particle swarm optimization algorithm. Expert Syst Appl. 2015;42(2):844–54.
Ng AY, Jordan MI, Weiss Y. On spectral clustering: analysis and an algorithm. Adv Neural Inf Process Syst. 2002;14:849–56.
Ding SF, Jia HJ, Shi ZZ. Spectral clustering algorithm based on adaptive Nyström sampling for big data analysis. J Softw. 2014;25(9):2037–49.
Acknowledgments
This work is supported by the National Natural Science Foundation of China (No. 61379101), and the National Key Basic Research Program of China (No. 2013CB329502).
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Jia, H., Ding, S. & Du, M. Self-Tuning p-Spectral Clustering Based on Shared Nearest Neighbors. Cogn Comput 7, 622–632 (2015). https://doi.org/10.1007/s12559-015-9331-2
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s12559-015-9331-2