Skip to main content

Advertisement

Log in

Automatic mandible segmentation from CT image using 3D fully convolutional neural network based on DenseASPP and attention gates

  • Original Article
  • Published:
International Journal of Computer Assisted Radiology and Surgery Aims and scope Submit manuscript

Abstract

Purpose

In cranio-maxillofacial surgery, it is of great clinical significance to segment mandible accurately and automatically from CT images. However, the connected region and blurred boundary in teeth and condyles make the process challenging. At present, the mandible is commonly segmented by experienced doctors using manually or semi-automatic methods, which is time-consuming and has poor segmentation consistency. In addition, existing automatic segmentation methods still have problems such as region misjudgment, low accuracy, and time-consuming.

Methods

For these issues, an automatic mandibular segmentation method using 3d fully convolutional neural network based on densely connected atrous spatial pyramid pooling (DenseASPP) and attention gates (AG) was proposed in this paper. Firstly, the DenseASPP module was added to the network for extracting dense features at multiple scales. Thereafter, the AG module was applied in each skip connection to diminish irrelevant background information and make the network focus on segmentation regions. Finally, a loss function combining dice coefficient and focal loss was used to solve the imbalance among sample categories.

Results

Test results showed that the proposed network obtained a relatively good segmentation result, with a Dice score of 97.588 ± 0.425%, Intersection over Union of 95.293 ± 0.812%, sensitivity of 96.252 ± 1.106%, average surface distance of 0.065 ± 0.020 mm and 95% Hausdorff distance of 0.491 ± 0.021 mm in segmentation accuracy. The comparison with other segmentation networks showed that our network not only had a relatively high segmentation accuracy but also effectively reduced the network's misjudgment. Meantime, the surface distance error also showed that our segmentation results were relatively close to the ground truth.

Conclusion

The proposed network has better segmentation performance and realizes accurate and automatic segmentation of the mandible. Furthermore, its segmentation time is 50.43 s for one CT scan, which greatly improves the doctor's work efficiency. It will have practical significance in cranio-maxillofacial surgery in the future.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5

Similar content being viewed by others

References

  1. Maschaa F, Winterb K, Pietzkaa S, Heufeldera M, Schramma A, Wilde F (2017) Accuracy of computer-assisted mandibular reconstructions using patient-specific implants in combination with CAD/CAM fabricated transfer keys. J Craniomaxillofac Surg 45(11):1884–1897

    Article  Google Scholar 

  2. Cristoforettia A, Stavolac LD, Fincatoc A, Mase M, Ravellib F, Nolloa G, Tessaroloa F (2019) Assessing the accuracy of computer-planned osteotomy guided by stereolithographic template: a methodological framework applied to the mandibular bone harvesting. Comput Biol Med 114:103435

    Article  Google Scholar 

  3. Chen X, Xu L, Sun Y, Politis C (2016) A review of computer-aided oral and maxillofacial surgery: planning, simulation and navigation. Expert Rev Med Devices 13(11):1043–1051

    Article  CAS  Google Scholar 

  4. Tang H, Chen X, Liu Y, Lu Z, You J, Yang M, Yao S, Zhao G, Xu Y, Chen T, Liu Y, Xie X (2019) Clinically applicable deep learning framework for organs at risk delineation in CT images. Nat Mach Intell 1:480–491

    Article  Google Scholar 

  5. Chuang Y, Doherty BM, Adluru N, Chung MK, Vorperian HK (2018) A novel registration-based semiautomatic mandible segmentation pipeline using computed tomography images to study mandibular development. J Comput Assist Tomogr 42:306–316

    Article  Google Scholar 

  6. Wallner J, Hochegger K, Chen X, Mischak I, Reinbacher K, Pau M, Zrnc T, Schwenzer-Zimmerer K, Zemann W, Schmalstieg D, Egger J (2018) Clinical evaluation of semi-automatic open-source algorithmic software segmentation of the mandibular bone: practical feasibility and assessment of a new course of action. PLoS ONE 13(5):e0196378

    Article  Google Scholar 

  7. Gass T, Szekely G, Goksel O (2014) Simultaneous segmentation and multiresolution nonrigid atlas registration. IEEE Trans Image Process 23(7):2931–2943

    Article  Google Scholar 

  8. Pei Y, Yi Y, Ma G, Kim T, Guo Y, Xu T, Zha H (2018) Spatially consistent supervoxel correspondences of cone-beam computed tomography images. IEEE Trans Med Imag 37(10):2310–2321

    Article  Google Scholar 

  9. Fan Y, Beare R, Matthews H, Schneider P, Kilpatrick N, Clement J, Claes P, Penington A, Adamson C (2019) Marker-based watershed transform method for fully automatic mandibular segmentation from CBCT images. Dentomaxillofac Radiol 48(2):20180261

    Article  Google Scholar 

  10. Taghizadeh E, Terrier A, Becce F, Farron A, Büchler P (2019) Automated CT bone segmentation using statistical shape modelling and local template matching. Comput Methods Biomech Biomed Engin 22(16):1303–1310

    Article  Google Scholar 

  11. Xu J, Jing M, Wang S, Yang C, Chen X (2019) A review of medical image detection for cancers in digestive system based on artificial intelligence. Expert Rev Med Dev 16(10):877–889

    Article  CAS  Google Scholar 

  12. Ibragimov B, Xing L (2017) Segmentation of organs-at-risks in head and neck CT images using convolutional neural networks. Med Phys 44(2):547–557

    Article  CAS  Google Scholar 

  13. Yan M, Guo J, Tian W, Yi Z (2018) Symmetric convolutional neural network for mandible segmentation. Knowl Based Syst 159:63–71

    Article  Google Scholar 

  14. Qiu B, Guo J, Kraeima J, Glas HH, Borra RJH, Witjes MJH, Ooijen PMAV (2019) Automatic segmentation of the mandible from computed tomography scans for 3D virtual surgical planning using the convolutional neural network. Phys Med Biol 64(17):175020

    Article  Google Scholar 

  15. Gou S, Tong N, Qi S, Yang S, Chin R, Sheng K (2020) Self-channel-and-spatial-attention neural network for automated multi-organ segmentation on head and neck CT images. Phys Med Biol 65:245034

    Article  Google Scholar 

  16. Yong M, Yu K, Zhang C, Li Z, Yang K (2018) Denseaspp for semantic segmentation in street scenes. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), pp 3684–3692

  17. Oktay, O, Schlemper J, Folgoc LL, Lee M, Heinrich M, Misawa K, Mori K, McDonagh S, Hammerla NY, Kainz B, Glocker B, Rueckert D (2018) Attention U-Net: learning where to look for the pancreas. arXiv preprint, http://arxiv.org/abs/1804.03999

  18. Lin T, Goyal P, Girshick R, He K, Dollar P (2020) Focal loss for dense object detection. IEEE Trans Pattern Anal Mach Intell 42(2):318–327

    Article  Google Scholar 

  19. Wu Y, He K (2020) Group normalization. Int J Comput Vision 128(3):742–755

    Article  Google Scholar 

  20. Yu F, Koltun V (2015) Multi-scale context aggregation by dilated convolutions. arXiv preprint http://arxiv.org/abs/1511.07122

  21. Shelhamer E, Long J, Darrell T (2015) Fully convolutional networks for semantic segmentation. IEEE Trans Pattern Anal Mach Intell 39(4):640–651

    Article  Google Scholar 

  22. Ronneberger O, Fischer P, Brox T (2015) U-Net: convolutional networks for biomedical image segmentation. In: Proceedings of the international conference on medical image computing and computer-assisted intervention, Springer, Cham, pp 234–241

  23. Milletari F, Navab N, Ahmadi SA (2016) V-Net: fully convolutional neural networks for volumetric medical image segmentation. In: Proceedings of the IEEE fourth international conference on 3D vision, pp 565–571

  24. Raith S, Wolff S, Steiner T, Modabber A, Weber M, Hölzle F, Fischer H (2017) Planning of mandibular reconstructions based on statistical shape models. Int J Comput Assist Radiol Surg 15:1457–1465

    Google Scholar 

  25. Zhu X, Han J, Zhou Z, Min X, Zhang Q, Liu J, Zhai G (2019) Improving accuracy of jaw reconstruction based on gender differences. J Oral Maxillofac Surg 78:662.e1-662.e13

    Article  Google Scholar 

Download references

Acknowledgements

This work was supported by grants from the National Key Research and Development Program of China (No. 2017YFB1302900), National Natural Science Foundation of China (81971709; M-0019; 82011530141), the Foundation of Science and Technology Commission of Shanghai Municipality (19510712200, 20490740700), Shanghai Jiao Tong University Foundation on Medical and Technological Joint Science Research (ZH2018ZDA15; YG2019ZDA06; ZH2018QNA23), and 2020 Key Research project of Xiamen Municipal Government (3502Z20201030).

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to Chenping Zhang or Xiaojun Chen.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Ethical approval

All procedures performed in studies involving human participants were in accordance with the ethical standards of the institutional and/or national research committee and with the 1964 Helsinki Declaration and its later amendments or comparable ethical standards.

Informed consent

There was no informed consent required for the work reported in this manuscript.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Xu, J., Liu, J., Zhang, D. et al. Automatic mandible segmentation from CT image using 3D fully convolutional neural network based on DenseASPP and attention gates. Int J CARS 16, 1785–1794 (2021). https://doi.org/10.1007/s11548-021-02447-5

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11548-021-02447-5

Keywords

Navigation