skip to main content
10.1145/3616901.3616922acmotherconferencesArticle/Chapter ViewAbstractPublication PagesfaimlConference Proceedingsconference-collections
research-article

K-means Optimization Method Based On Adaptive Parallel Hierarchical Clustering

Published:05 March 2024Publication History

ABSTRACT

The two key steps of the K-means algorithm are the selection of the clustering number and the selection of the initial clustering center, which will seriously affect the classification accuracy and efficiency of K-means, and need further optimization. Aiming at the selection of the number of clusters, a K-means optimization method based on adaptive parallel hierarchical clustering is proposed. In the merging process of hierarchical clustering, the optimal number of clusters is selected adaptively by improving the clustering effect evaluation function, and the Parallel computing method is used instead of the serial computing method to improve the computing speed. Aiming at selecting cluster centers more accurately, an optimized data density model is proposed to make full use of potentially related information between samples, which improves the classification accuracy of the algorithm. More importantly, it overcomes the problem of the strong subjectivity of super-parameter selection. The improved algorithm was tested with the ablation experiment method and compared to other traditional algorithms on iris and seed data sets. The results showed that the optimization algorithm could accelerate the calculation speed and improve the classification accuracy.

References

  1. Li Peng. Research on Hierarchical K-means based clustering algorithm [D]. Harbin: Harbin Engineering University, 2015.Google ScholarGoogle Scholar
  2. SHI Xiaoyu, TANG Xiaoyu, WANG Xiaoli, SUN Yaming, QI Zixuan, ZHANG Yanxin. Cluster Analysis of Dairy Consumption Preference in Hebei Province Based on K-means Clustering [J]. Journal of Hebei Agricultural Sciences, 2021, 25 (2): 29-33.Google ScholarGoogle Scholar
  3. Pelleg, Dan & Moore, Andrew. (2002). X-means: Extending K-means with Efficient Estimation of the Number of Clusters. Machine Learning, p.Google ScholarGoogle Scholar
  4. Redmond S J, Heneghan C.A Method for Initialising the K − means Clustering Algorithm Using Kd − trees[J]. Pattern Recognition Letters, 2007, 28( 8) : 965 − 973.Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Jia Ruiyu, Song Jianlin. K-means Optimal Clustering Number Determination MethodBased on Clustering Center Optimization. MICROELECTRONICS &COMPUTER, 2016, 33(5): 62-66.Google ScholarGoogle Scholar
  6. WANG S,LIU C,XING S J. Review on K-means clustering algorithm[J]. Journal of East China Jiaotong University, 2022, 39(5): 119-126.Google ScholarGoogle Scholar
  7. LI Y S, YANG S L, MA X J, HU X X, CHEN Z M. Optimization Study on K Value of Spatial Clustering[J]. Journal of System Simulation, 2006, 18(3): 573-576.Google ScholarGoogle Scholar
  8. He Xuansen, He Fan, Xu Li, Fan Yueping. Determination of the Optimal Number of Clusters in K-Means Algorithm[J]. Journal of University of Electronic Science and Technology of China, 2022,51(6): 904 – 912.Google ScholarGoogle Scholar
  9. Li Chunfang, Zhang Ruifeng, Jia Lu, Wang Fang, Guo Fei. A new electricity stealing identification model and simulation based on improved k-means algorithm and big data analysis [J].Electronic Design Engineering, 2022, 30(22) : 84-88.Google ScholarGoogle Scholar
  10. WANG Zhong, LIU Gui-Quan, CHEN En-Hong. A K-means Algorithm Based on Optimized Initial Center Points. PR&AI, 2009, 22(2): 299−303.Google ScholarGoogle Scholar
  11. Jones D R,Beltramo M A.Solving Partitioning Problems with Genetic Algorithms[C]. In: Proceedings of the 4th International Conference Genetic Algorithms, San Diego,CA,USA. 1991: 442 − 494.Google ScholarGoogle Scholar
  12. Lai Yuxia, Liu Jianping, Yang Guoxing. K-Means Clustering Analysis Based on Genetic Algorithm[J]. Computer Engineering, 2008, 34(20):200-202.Google ScholarGoogle Scholar
  13. Zhang Chao. K-means Clustering Center Selection [J]. Journal of Jilin University, 2019, 37(4):437-441Google ScholarGoogle Scholar
  14. Tao Yonghui, Wang Yong. Improved K-means algorithm based on the selection of initial clustering center [J].theories and methods, 2022,41(9):54 – 59.Google ScholarGoogle Scholar
  15. Sun Lin, Liu Menghan, Xu Jiucheng .K-means Clustering Algorithm Using Optimal Initial Clustering Center and Contour Coefficie. Fuzzy Systems and Mathematics, 2022, 36(1):47-64.Google ScholarGoogle Scholar
  16. Novoselsky, Alexander & Kagan, Eugene. (2021). An introduction to cluster analysis. 10.13140/RG.2.2.25993.57448/1.Google ScholarGoogle Scholar
  17. HAN Ling-bo,WANG Qiang,JIANG Zheng-feng,et al.Improved k-means initial clustering center selection algorithm. Computer Engineering and Applications,2010,46(17):150-152.Google ScholarGoogle Scholar
  18. X. Wu, Z. Chen, S. Yuan, J. Wei and X. Wang, "An improved k-means algorithm based on density normalization," 2021 IEEE 2nd International Conference on Information Technology, Big Data and Artificial Intelligence (ICIBA), Chongqing, China, 2021, pp. 1141-1146, doi: 10.1109/ICIBA52610.2021.9687899.Google ScholarGoogle ScholarCross RefCross Ref
  19. Mitra P, Murthy C A, Pal S K. Density-based multiscale data condensation[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2002, 24(6): 734-747.Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. CAI Yuhao, LIANG Yongquan, FAN Jiancong, LI Xuan, LIU Wenhua . Optimizing Initial Cluster Centroids by Weighted Local Variance in K-means Algorithm[J]. Journal of Frontiers of Computer Science and Technology, 2016, 10(5): 732-741.Google ScholarGoogle Scholar
  21. Rezaee M R, Lelieveldt B P, Reiber J H. A New Cluster Validity Index for the Fuzzy C-Means[J].Pattern Recognition Letters,1998, 19( 3 − 4) : 237 − 246.Google ScholarGoogle Scholar
  22. Chen Yin, He Zhongshi . The study on improved K-means algorithm[J]. Manufacturing automation, 2012, 34(4):19-22.Google ScholarGoogle Scholar
  23. Huang He, Xiong Wu, Wu Kun, Wang Huifeng. K-means Hybrid Iterative Clustering Basedon Memory Transfer Sail fish Optimization.JOURNAL OF SHANGHAIJIAOTONG UNIVERSITY, 2022, 56(12) :1638-1648.Google ScholarGoogle Scholar

Index Terms

  1. K-means Optimization Method Based On Adaptive Parallel Hierarchical Clustering
        Index terms have been assigned to the content through auto-classification.

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Other conferences
          FAIML '23: Proceedings of the 2023 International Conference on Frontiers of Artificial Intelligence and Machine Learning
          April 2023
          296 pages
          ISBN:9798400707544
          DOI:10.1145/3616901

          Copyright © 2023 ACM

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 5 March 2024

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • research-article
          • Research
          • Refereed limited
        • Article Metrics

          • Downloads (Last 12 months)15
          • Downloads (Last 6 weeks)0

          Other Metrics

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader

        HTML Format

        View this article in HTML Format .

        View HTML Format