Abstract
Coke image segmentation is a crucial step in coke particle size control of the sintering process. However, due to the complexity of model architecture and the dense distribution of coke particles in the images, existing segmentation methods fail to satisfy the efficiency and accuracy requirements for coke image segmentation in industrial scenarios. To address these challenges, this paper proposes a two-stage distillation-aware adaptive segment anything model to balance efficiency and accuracy in coke image particle size segmentation, referred to as TsDa-ASAM. In the first stage, knowledge distillation methods are employed to distill the Segment Anything Model (SAM) into a lightweight model, explicitly focusing on enhancing segmentation efficiency. In the second stage, a domain knowledge injection strategy is formulated, which incorporates domain knowledge into the distillation model to effectively enhance the accuracy. Moreover, an adaptive prompt point selection algorithm is introduced to address the redundancy issue of prompt points in SAM, improving the efficiency of TsDa-ASAM. The effectiveness of TsDa-ASAM is validated through extensive experiments on the publicly available dataset SA-1B and the coke image dataset from industrial sites. After distillation and fine-tuning, the segmentation accuracy of the proposed model improved by 10%, and the segmentation efficiency of TsDa-ASAM was enhanced by 2 to 3 times with the integration of the adaptive prompt point selection algorithm. The experimental results have effectively demonstrated the potential of the proposed model in balancing accuracy and efficiency.













Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Xiong L et al (2018) Combustion behavior of granulated coke breeze in iron ore sintering. Powder Technol 340:131–138
Zhang Z et al (2020) Multi-information online detection of coal quality based on machine vision. Powder Technol 374:250–262
Wang X, Guo Y, Yu Y (2023) An appropriate approach to recognize coke size distribution in a blast furnace. Processes 11(1):187
Minaee S, Boykov Y, Porikli F, Plaza A, Kehtarnavaz N, Terzopoulos D (2021) Image segmentation using deep learning: A survey. IEEE Trans Pattern Anal Mach Intell 44(7):3523–3542
Hafiz AM, Bhat GM (2020) A survey on instance segmentation: state of the art. Int J Multimedia Inf Retrieval 9(3):171–189
Kirillov A, Mintun E, Ravi N, Mao H, Rolland C, Gustafson L, Xiao T, Whitehead S, Berg AC, Lo WY, Dollár P, Girshick R (2023) Segment anything. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV). pp 4015–4026
Chen F, Chen L, Han H, Zhang S, Zhang D, Liao H (2023) The ability of Segmenting Anything Model (SAM) to segment ultrasound images. BioSci Trends 17(3):211–218
Fazekas B, Morano J, Lachinov D, Aresta G, Bogunović H (2023) Adapting segment anything model (SAM) for retinal OCT. International Workshop on Ophthalmic Medical Image Analysis. Springer, pp 92–101
Zhang C et al (2023) Faster segment anything: towards lightweight SAM for mobile applications. arXiv preprint arXiv:2306.14289
Zhang L, Ma K (2020) Improve object detection with feature-based knowledge distillation: towards accurate and efficient detectors. In International Conference on Learning Representations
Choudhary T, Mishra V, Goswami A, Sarangapani J (2020) A comprehensive survey on model compression and acceleration. Artif Intell Rev 53:5113–5155
Liu X, Peng H, Zheng N, Yang Y, Hu H, Yuan Y (2023) EfficientViT: Memory efficient vision transformer with cascaded group attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp 14420–14430
Gou J, Yu B, Maybank SJ, Tao D (2021) Knowledge distillation: A survey. Int J Comput Vision 129:1789–1819
Wang L, Yoon K-J (2021) Knowledge distillation and student-teacher learning for visual intelligence: A review and new outlooks. IEEE Trans Pattern Anal Mach Intell 44(6):3048–3068
Tung F, Mori G (2019) Similarity-preserving knowledge distillation. In Proceedings of the IEEE/CVF international conference on computer vision. pp 1365–1374
Zhao B, Cui Q, Song Y, Qiu Y, Liang J (2022) Decoupled knowledge distillation. In Proceedings of the IEEE/CVF Conference on computer vision and pattern recognition. pp 11953–11962
Hinton G, Vinyals O, Dean J (2015) Distilling the knowledge in a neural network. Comput Sci 14(7):38–39
Shu C, Liu Y, Gao J, Yan Z, Shen C (2021) Channel-wise knowledge distillation for dense prediction. In Proceedings of the IEEE/CVF International Conference on Computer Vision. pp 5311–5320
Kirillov A, He K, Girshick R, Rother C, Dollár P (2019) Panoptic segmentation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. pp 9404–9413
Yang Z, Ye X, Xu H, Zhao P (2023) Emotional polar predictions based on matching prompts. In 2023 8th International Conference on Intelligent Computing and Signal Processing (ICSP). IEEE, pp 253–257
Jia M et al (2022) Visual prompt tuning. European Conference on Computer Vision. Springer, pp 709–727
Hu J, Li Y, Lin L, Chen YW (2023) Integrating spatial prior adapter for enhancing SAM performance in medical image segmentation. 2023 IEEE 12th Global Conference on Consumer Electronics (GCCE). pp 20–23
Liu X. A SAM-based method for large-scale crop field boundary delineation. In 2023 20th Annual IEEE International Conference on Sensing, Communication, and Networking (SECON). IEEE, pp 1–6
Lan X et al (2023) FoodSAM: any food segmentation. IEEE Transactions on Multimedia
Chen T et al (2023) Sam-adapter: adapting segment anything in underperformed scenes. In Proceedings of the IEEE/CVF International Conference on Computer Vision. pp 3367–3375
Khan S, Naseer M, Hayat M, Zamir SW, Khan FS, Shah M (2022) Transformers in vision: a survey. ACM Comput Surv (CSUR) 54(10s):1–41
Mazurowski MA, Dong H, Gu H, Yang J, Konz N, Zhang Y (2023) Segment anything model for medical image analysis: an experimental study. Med Image Anal 89:102918
Huang Y et al (2023) Segment anything model for medical images? Medical Image Analysis. pp 103061
Li Y, Wang D, Yuan C, Li H, Hu J (2023) Enhancing agricultural image segmentation with an agricultural segment anything model adapter. Sensors 23(18):7884
Zhao X et al (2023) Fast segment anything. arXiv preprint arXiv:2306.12156
He J, Li P, Geng Y, Xie X (2023) FastInst: A simple query-based model for real-time instance segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp 23663–23672
Sun S, Cheng Y, Gan Z, Liu J (2019) Patient knowledge distillation for BERT model compression. Hong Kong, China: Association for Computational Linguistics, in Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp 4323–4332. https://doi.org/10.18653/v1/D19-1441
Wang W, Wei F, Dong L, Bao H, Yang N, Zhou M (2020) Minilm: Deep self-attention distillation for task-agnostic compression of pre-trained transformers. Adv Neural Inf Process Syst 33:5776–5788
Jiao X et al (2020) TinyBERT: Distilling BERT for Natural Language Understanding. Online, November 2020: Association for Computational Linguistics, in Findings of the Association for Computational Linguistics: EMNLP, pp 4163-4174. https://doi.org/10.18653/v1/2020.findings-emnlp.372
Wu K et al (2022) Tinyvit: Fast pretraining distillation for small vision transformers. European Conference on Computer Vision. Springer, pp 68–85
Bodla N, Singh B, Chellappa R, Davis LS (2017) Soft-NMS--improving object detection with one line of code. In Proceedings of the IEEE international conference on computer vision. pp 5561–5569
He K, Chen X, Xie S, Li Y, Dollár O, Girshick R (2022) Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. pp 16000–16009
Ronneberger O, Fischer P, Brox T (2015) U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference. Munich, Germany, October 5-9, 2015, Proceedings, Part III 18, 2015: Springer, pp 234-241
Kornilov AS, Safonov IV (2018) An overview of watershed algorithm implementations in open source libraries. J Imaging 4(10):123
Funding
This work was supported in part by the National Natural Science Foundation of China under Grant 92267205.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Wang, Y., Peng, Y., Tan, X. et al. TsDa-ASAM: Balancing efficiency and accuracy in coke image particle size segmentation via two-stage distillation-aware adaptive segment anything model. Appl Intell 55, 529 (2025). https://doi.org/10.1007/s10489-025-06427-z
Accepted:
Published:
DOI: https://doi.org/10.1007/s10489-025-06427-z