Skip to main content
Log in

TsDa-ASAM: Balancing efficiency and accuracy in coke image particle size segmentation via two-stage distillation-aware adaptive segment anything model

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

Coke image segmentation is a crucial step in coke particle size control of the sintering process. However, due to the complexity of model architecture and the dense distribution of coke particles in the images, existing segmentation methods fail to satisfy the efficiency and accuracy requirements for coke image segmentation in industrial scenarios. To address these challenges, this paper proposes a two-stage distillation-aware adaptive segment anything model to balance efficiency and accuracy in coke image particle size segmentation, referred to as TsDa-ASAM. In the first stage, knowledge distillation methods are employed to distill the Segment Anything Model (SAM) into a lightweight model, explicitly focusing on enhancing segmentation efficiency. In the second stage, a domain knowledge injection strategy is formulated, which incorporates domain knowledge into the distillation model to effectively enhance the accuracy. Moreover, an adaptive prompt point selection algorithm is introduced to address the redundancy issue of prompt points in SAM, improving the efficiency of TsDa-ASAM. The effectiveness of TsDa-ASAM is validated through extensive experiments on the publicly available dataset SA-1B and the coke image dataset from industrial sites. After distillation and fine-tuning, the segmentation accuracy of the proposed model improved by 10%, and the segmentation efficiency of TsDa-ASAM was enhanced by 2 to 3 times with the integration of the adaptive prompt point selection algorithm. The experimental results have effectively demonstrated the potential of the proposed model in balancing accuracy and efficiency.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

References

  1. Xiong L et al (2018) Combustion behavior of granulated coke breeze in iron ore sintering. Powder Technol 340:131–138

    Article  MATH  Google Scholar 

  2. Zhang Z et al (2020) Multi-information online detection of coal quality based on machine vision. Powder Technol 374:250–262

    Article  MATH  Google Scholar 

  3. Wang X, Guo Y, Yu Y (2023) An appropriate approach to recognize coke size distribution in a blast furnace. Processes 11(1):187

    Article  MATH  Google Scholar 

  4. Minaee S, Boykov Y, Porikli F, Plaza A, Kehtarnavaz N, Terzopoulos D (2021) Image segmentation using deep learning: A survey. IEEE Trans Pattern Anal Mach Intell 44(7):3523–3542

    Google Scholar 

  5. Hafiz AM, Bhat GM (2020) A survey on instance segmentation: state of the art. Int J Multimedia Inf Retrieval 9(3):171–189

    Article  MATH  Google Scholar 

  6. Kirillov A, Mintun E, Ravi N, Mao H, Rolland C, Gustafson L, Xiao T, Whitehead S, Berg AC, Lo WY, Dollár P, Girshick R (2023) Segment anything. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV). pp 4015–4026

  7. Chen F, Chen L, Han H, Zhang S, Zhang D, Liao H (2023) The ability of Segmenting Anything Model (SAM) to segment ultrasound images. BioSci Trends 17(3):211–218

    Article  MATH  Google Scholar 

  8. Fazekas B, Morano J, Lachinov D, Aresta G, Bogunović H (2023) Adapting segment anything model (SAM) for retinal OCT. International Workshop on Ophthalmic Medical Image Analysis. Springer, pp 92–101

    Chapter  Google Scholar 

  9. Zhang C et al (2023) Faster segment anything: towards lightweight SAM for mobile applications. arXiv preprint arXiv:2306.14289

  10. Zhang L, Ma K (2020) Improve object detection with feature-based knowledge distillation: towards accurate and efficient detectors. In International Conference on Learning Representations

  11. Choudhary T, Mishra V, Goswami A, Sarangapani J (2020) A comprehensive survey on model compression and acceleration. Artif Intell Rev 53:5113–5155

    Article  MATH  Google Scholar 

  12. Liu X, Peng H, Zheng N, Yang Y, Hu H, Yuan Y (2023) EfficientViT: Memory efficient vision transformer with cascaded group attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp 14420–14430

  13. Gou J, Yu B, Maybank SJ, Tao D (2021) Knowledge distillation: A survey. Int J Comput Vision 129:1789–1819

    Article  MATH  Google Scholar 

  14. Wang L, Yoon K-J (2021) Knowledge distillation and student-teacher learning for visual intelligence: A review and new outlooks. IEEE Trans Pattern Anal Mach Intell 44(6):3048–3068

    Article  MATH  Google Scholar 

  15. Tung F, Mori G (2019) Similarity-preserving knowledge distillation. In Proceedings of the IEEE/CVF international conference on computer vision. pp 1365–1374

  16. Zhao B, Cui Q, Song Y, Qiu Y, Liang J (2022) Decoupled knowledge distillation. In Proceedings of the IEEE/CVF Conference on computer vision and pattern recognition. pp 11953–11962

  17. Hinton G, Vinyals O, Dean J (2015) Distilling the knowledge in a neural network. Comput Sci 14(7):38–39

    MATH  Google Scholar 

  18. Shu C, Liu Y, Gao J, Yan Z, Shen C (2021) Channel-wise knowledge distillation for dense prediction. In Proceedings of the IEEE/CVF International Conference on Computer Vision. pp 5311–5320

  19. Kirillov A, He K, Girshick R, Rother C, Dollár P (2019) Panoptic segmentation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. pp 9404–9413

  20. Yang Z, Ye X, Xu H, Zhao P (2023) Emotional polar predictions based on matching prompts. In 2023 8th International Conference on Intelligent Computing and Signal Processing (ICSP). IEEE, pp 253–257

  21. Jia M et al (2022) Visual prompt tuning. European Conference on Computer Vision. Springer, pp 709–727

    MATH  Google Scholar 

  22. Hu J, Li Y, Lin L, Chen YW (2023) Integrating spatial prior adapter for enhancing SAM performance in medical image segmentation. 2023 IEEE 12th Global Conference on Consumer Electronics (GCCE). pp 20–23

  23. Liu X. A SAM-based method for large-scale crop field boundary delineation. In 2023 20th Annual IEEE International Conference on Sensing, Communication, and Networking (SECON). IEEE, pp 1–6

  24. Lan X et al (2023) FoodSAM: any food segmentation. IEEE Transactions on Multimedia

  25. Chen T et al (2023) Sam-adapter: adapting segment anything in underperformed scenes. In Proceedings of the IEEE/CVF International Conference on Computer Vision. pp 3367–3375

  26. Khan S, Naseer M, Hayat M, Zamir SW, Khan FS, Shah M (2022) Transformers in vision: a survey. ACM Comput Surv (CSUR) 54(10s):1–41

    Article  MATH  Google Scholar 

  27. Mazurowski MA, Dong H, Gu H, Yang J, Konz N, Zhang Y (2023) Segment anything model for medical image analysis: an experimental study. Med Image Anal 89:102918

    Article  Google Scholar 

  28. Huang Y et al (2023) Segment anything model for medical images? Medical Image Analysis. pp 103061

  29. Li Y, Wang D, Yuan C, Li H, Hu J (2023) Enhancing agricultural image segmentation with an agricultural segment anything model adapter. Sensors 23(18):7884

    Article  MATH  Google Scholar 

  30. Zhao X et al (2023) Fast segment anything. arXiv preprint arXiv:2306.12156

  31. He J, Li P, Geng Y, Xie X (2023) FastInst: A simple query-based model for real-time instance segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp 23663–23672

  32. Sun S, Cheng Y, Gan Z, Liu J (2019) Patient knowledge distillation for BERT model compression. Hong Kong, China: Association for Computational Linguistics, in Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp 4323–4332. https://doi.org/10.18653/v1/D19-1441

  33. Wang W, Wei F, Dong L, Bao H, Yang N, Zhou M (2020) Minilm: Deep self-attention distillation for task-agnostic compression of pre-trained transformers. Adv Neural Inf Process Syst 33:5776–5788

    MATH  Google Scholar 

  34. Jiao X et al (2020) TinyBERT: Distilling BERT for Natural Language Understanding. Online, November 2020: Association for Computational Linguistics, in Findings of the Association for Computational Linguistics: EMNLP, pp 4163-4174. https://doi.org/10.18653/v1/2020.findings-emnlp.372

  35. Wu K et al (2022) Tinyvit: Fast pretraining distillation for small vision transformers. European Conference on Computer Vision. Springer, pp 68–85

    MATH  Google Scholar 

  36. Bodla N, Singh B, Chellappa R, Davis LS (2017) Soft-NMS--improving object detection with one line of code. In Proceedings of the IEEE international conference on computer vision. pp 5561–5569

  37. He K, Chen X, Xie S, Li Y, Dollár O, Girshick R (2022) Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. pp 16000–16009

  38. Ronneberger O, Fischer P, Brox T (2015) U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference. Munich, Germany, October 5-9, 2015, Proceedings, Part III 18, 2015: Springer, pp 234-241

  39. Kornilov AS, Safonov IV (2018) An overview of watershed algorithm implementations in open source libraries. J Imaging 4(10):123

    Article  MATH  Google Scholar 

Download references

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 92267205.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Xujie Tan.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Wang, Y., Peng, Y., Tan, X. et al. TsDa-ASAM: Balancing efficiency and accuracy in coke image particle size segmentation via two-stage distillation-aware adaptive segment anything model. Appl Intell 55, 529 (2025). https://doi.org/10.1007/s10489-025-06427-z

Download citation

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s10489-025-06427-z

Keywords