Deep learning network for medical volume data segmentation based on multi axial plane fusion

https://doi.org/10.1016/j.cmpb.2021.106480Get rights and content

Highlights

  • A network segmentation is presented that can extract the features of different pinacoidal planes based on the medical volume data.

  • The network is used for the multi-input semantic convolution neural network for fusion and segmentation.

  • The global feature maps of different pinacoidal planes could bring further compensation for losses in the segmentation process, so as to obtain a better result.

Abstract

Background and Objective: High-dimensional data generally contains more accurate information for medical image, e.g., computerized tomography (CT) data can depict the three dimensional structure of organs more precisely. However, the data in high-dimension often needs enormous computation and has high memory requirements in the deep learning convolution networks, while dimensional reduction usually leads to performance degradation. Methods: In this paper, a two-dimensional deep learning segmentation network was proposed for medical volume data based on multi-pinacoidal plane fusion to cover more information under the control of computation.This approach has conducive compatibility while using the model proposed to extract the global information between different inputs layers. Results: Our approach has worked in different backbone network. Using the approach, DeepUnet’s Dice coefficient (Dice) and Positive Predictive Value (PPV) are 0.883 and 0.982 showing the satisfied progress. Various backbones can enjoy the profit of the method. Conclusions: Through the comparison of different backbones, it can be found that the proposed network with multi-pinacoidal plane fusion can achieve better results both quantitively and qualitatively.

Introduction

In the resection of gastric cancer, the direction of celiac trunk is particularly important for the surgical plans. Therefore, the segmentation of celiac trunk is of great significance for the diagnosis of gastric surgery. Medical equipment often generate three-dimensional or even four-dimensional data, but due to the limitation of preservation, this data usually need to be formed into two-dimensional or three-dimensional structure. Hence, how to segment more accurate results from the transformed medical data [1], [2]–3] are still one of the most popular topics in the medical image segmentation convolution network [4,5] field.

Due to the complexity of the shape of abdominal organ [6], [7]–8], two dimensional data generally can not display the real shape of the organ in an all round way. In computerized tomography(CT) imaging, the data are expressed as three-dimensional within a broader range thus it can retain more values, and present the exact shape of complex organs more comprehensively. In the research of convolution network segmentation, the use of three-dimensional data will bring tremendous amount of calculation so many researchers tend to transform the data into two-dimension and use other ways to supplement the missing information of the formatting transition. Definitely, this method will bring serious loss of dimensional reduction to the data, and the final segmentation result will naturally be much worse than that of 3D input/output method. Even though many convolutional neural networks can improve the segmentation accuracy to some extent by using feature reusing [9], [10]–11] and other network layer modified algorithms from one pinacoidal of the data [12], [13]–14], they can not supply the original and nondestructive data. Therefore, it is not a good solution to directly use the two-dimensional data of one pinacoidal. It remains a huge progress to be made.

In this paper, we proposed a novel network which can extract the features of different pinacoidal planes based on the medical volume data. The new network is different from the previous methods of convolutional neural network segmentation of one pinacoidal of the data. And then inputing them into the multi-input semantic convolution neural network for fusion and segmentation. The global feature maps of different pinacoidal planes could bring further compensation for losses in the segmentation process. The aim is to obtain a better result.

In general, our contributions are summarized as below:

  • 1.

    We proposed a segmentation structure which could be adapted into many different convolutional network. Moreover, it has hybrid architecture which is composed of multiple branches so as to extract the global information of different pinacoidal planes to make the segmentation of the target more accurate.

  • 2.

    Furthermore, We added X-module part to the network in order to extract the global information between different inputs layers. In the experiment, the feature maps generated by X-module can be merged into the corresponding layer of decoder part in parallel. So it can obtain more information.

Section snippets

Related work

In semantic convolutional segmentation neural network, The design of the loss function and the feature extractor can directly affect the overall performance of the network. Reusing of different level of features, superposition of feature maps and various extraction methods are still very universal in the current research. Besides, adding some widgets [12,15] or combining with traditional feature extraction method [16] are also popular in this field. Jonathan Long et al. [17] proposed fully

Theory

This paper proposed a novel convolutional segmentation network which used three kinds of inputs generated from different pinacoidal planes of CT volume data for fusion. It has three parallel branches, each of them inputs distinct pinacoidal plane slices image. In the encoder part, the X-module is used to extract the global information [35] of the different inputs layer by layer, and integrates the feature maps into the subsequent feature extraction [36] steps. Without increasing a large amount

Dataset and training setups

This paper used the Upper-Abdomen Routine Enhanced Scan CT volume data from Shanghai Changzheng Hospital and resized them into 256*256, and the size of the original data are 512 * 512 * N, the range of N is [35],[70]. Here, the appropriate sections will be selected according to the area of the stomach, so that these data are fixed in the same size. The training set containg 1600 pictures of 50 patients marked by experts. The test set includes 10 patients which are 320 pieces. The training set

Discussion

In the current segmentation research, it’s still a challenging task to segment the targets with more details and edges. For the normal tasks, the segmentation accuracy of some state-of-the-art can basically meet the requirements, but for the medical image with strict precision requirements the current method is far from satisfactory. In addition, medical data usually bear more informations such as volume data, so how to use this type of data to further improve the segmentation is also a very

Conclusion

In this paper, we proposed the multi-input network to fuse the three pinacoidal plane data by intercepting the different pinacoidal plane data of CT, and extract the global information of different pinacoidal planes layer by layer through the X-module so as to get better segmentation results. It can be found that the result of Dice and sensitivity is significantly improved compared with the original network. Therefore, we can draw a conclusion that the multi-axis plane fusion network can

Authors Contributions

Xianhua Tang and Ziran Wei contributed equally to this work and should be considered co-first authors.

Declaration of competing Interst

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Acknowledgements

This work is supported by the Key Projects of Shanghai Science and Technology Commission, cooperation project with Shanghai Changzheng Hospital Grant 18411952800, Grant 0232-E2-6202-19-022 and the National Natural Science Foundation of China (No. 61802251), and National Natural Science Foundation of China (No. 82072228) and National Key R&D Program of China (No. 2020YFC2008700).

References (43)

  • S. Chatterjee et al.

    Integration of morphological preprocessing and fractal based feature extraction with recursive feature elimination for skin lesion types classification

    Comput. Methods Programs Biomed.

    (2019)
  • Y. Kim et al.

    Automatic myocardial segmentation in dynamic contrast enhanced perfusion MRI using monte carlo dropout in an encoder-decoder convolutional neural network

    Comput. Methods Programs Biomed.

    (2020)
  • H. Wu et al.

    Automated comprehensive adolescent idiopathic scoliosis assessment using mvc-net

    Medical Image Anal.

    (2018)
  • P. Tang et al.

    Efficient skin lesion segmentation using separable-unet with stochastic weight averaging

    Comput Methods Programs Biomed

    (2019)
  • E. Gibson et al.

    Automatic multi-organ segmentation on abdominal CT with dense v-networks

    IEEE Trans. Medical Imaging

    (2018)
  • K. He et al.

    Deep residual learning for image recognition

    2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27–30, 2016

    (2016)
  • G. Huang et al.

    Densely connected convolutional networks

    2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, July 21–26, 2017

    (2017)
  • W. Zaremba et al.

    Recurrent neural network regularization

    CoRR

    (2014)
  • Z. Gu et al.

    Ce-net: context encoder network for 2d medical image segmentation

    IEEE Trans. Medical Imaging

    (2019)
  • J.J. Bouza et al.

    Mvc-net: a convolutional neural network architecture for manifold-valued images with applications

    CoRR

    (2020)
  • Y. Shen et al.

    Coronary arteries segmentation based on 3d FCN with attention gate and level set function

    IEEE Access

    (2019)
  • Cited by (5)

    • Multi-Dimensional Cascaded Net with Uncertain Probability Reduction for Abdominal Multi-Organ Segmentation in CT Sequences

      2022, Computer Methods and Programs in Biomedicine
      Citation Excerpt :

      Moreover, some researchers try to use 2.5D networks for multi-organ segmentation. Huang et al. [6] applied a 2.5D-based Dense-Unet to perform multi-class segmentation. Zhou et al. [7] also used a 2.5D FCN to segment multi-organ in CT image by voting approach.

    • Multimodality medical image fusion analysis with multi-plane features of PET and MRI images using ONSCT

      2024, Computer Methods in Biomechanics and Biomedical Engineering: Imaging and Visualization
    View full text