3D dense connectivity network with atrous convolutional feature pyramid for brain tumor segmentation in magnetic resonance imaging of human heads

https://doi.org/10.1016/j.compbiomed.2020.103766Get rights and content

Highlights

  • A densely connected convolutional neural network is developed as the backbone to learn dense features of brain tumors via MR images.

  • A 3D atrous convolution feature pyramid is connected with the end of the backbone to learn multiscale features of brain tumors.

  • A 3D deep supervision mechanism is further proposed to guide the optimization of the gradient flow during training.

Abstract

The existing deep convolutional neural networks (DCNNs) based methods have achieved significant progress regarding automatic glioma segmentation in magnetic resonance imaging (MRI) data. However, there are two main problems affecting the performance of traditional DCNNs constructed by simply stacking convolutional layers, namely, exploding/vanishing gradients and limitations to the feature computations. To address these challenges, we propose a novel framework to automatically segment brain tumors. First, a three-dimensional (3D) dense connectivity architecture is used to build the backbone for feature reuse. Second, we design a new feature pyramid module using 3D atrous convolutional layers and add this module to the end of the backbone to fuse multiscale contexts. Finally, a 3D deep supervision mechanism is equipped with the network to promote training. On the multimodal brain tumor image segmentation benchmark (BRATS) datasets, our method achieves Dice similarity coefficient values of 0.87, 0.72, and 0.70 on the BRATS 2013 Challenge, 0.84, 0.70, and 0.61 on the BRATS 2013 LeaderBoard, 0.83, 0.70, and 0.62 on the BRATS 2015 Testing, 0.8642, 0.7738, and 0.7525 on the BRATS 2018 Validation in terms of whole tumors, tumor cores, and enhancing cores, respectively. Compared to the published state-of-the-art methods, the proposed method achieves promising accuracy and fast processing, demonstrating good potential for clinical medicine.

Introduction

Brain tumors are a common clinical cancer of the nervous system that occurs in the cranial cavity and can be classified into two categories from their initial origin, i.e., primary brain tumors and metastatic brain tumors [1]. Brain tumors in the first category originate from brain lesions, while those in the other category originate from cancerous tissues in any other part of the body. Brain tumors are severely dangerous diseases that divide and grow uncontrollably, gradually damaging the nervous system and ultimately leading to death. To protect healthy tissues but damage tumor cells during therapy, segmenting the tumor becomes crucial before applying any therapy. Therefore, it is crucial to detect brain tumors early for timely treatment and prolonging the life of the patient. Since the advantage of MRI is its high capability to present extensive information about the soft tissue of the brain with both noninvasive and radiation-free characteristics, MRI is a popular technique to diagnose brain tumors. However, the tumor boundary is usually unclear, discontinuous, and irregular. Moreover, the intensities or other variations in MR images usually vary since both the imaging devices and environments seriously differ. These conditions make manually annotating a large number of multimodal MRI subjects impractical. Therefore, developing robust methods for automatic or semiautomatic brain tumor segmentation has become an interesting and challenging research topic.

With the rapid development of artificial intelligence (AI), especially the breakthroughs in deep learning techniques and the rise of medical big data, many researchers have started to focus on how to efficiently study medical images via AI [2]. The methods can be classified into none DCNNs-based and DCNN-based methods. The none DCNNs-based methods usually employ hand-crafted features with specific classifiers for processing medical images. Soltaninejad et al. [3] used a support vector machine (SVM) on statistical features for brain tumor grading. Superpixels/supervoxels were generated from a multimodal MRI dataset to extract statistical features, and then random forests were implemented to classify each superpixel/supervoxel [4], [5]. Yang et al. [6] proposed a 3D morphological analysis for tumor segmentation. Raschke et al. [7] proposed a retrospective analysis of MRI from patients with a histological diagnosis of glioma that used Bayesian inference to achieve the whole brain tumor tissue-type maps and automated segmentation. For classifying the tumor type, Timothy et al. [8] proposed whole-brain diffusion tensor imaging segmentation to delineate the tumor volumes of interest. Similar to tensor imaging, Yang et al. [9] presented a computerized decision support framework that discriminates between high-grade glioblastoma multiform and solitary metastasis using MRI. Due to the state-of-the-art performance achieved by DCNNs in the field of computer vision, DCNNs-based methods have been gradually applied in medical image segmentation tasks. In contrast to the previous none DCNNs-based methods, DCNNs-based techniques are a category of data-driven methods that build complex models via representation learning mechanisms in an end-to-end manner. The existing DCNNs-based methods applied to medical images can be classified into 2D DCNNs and 3D DCNNs. The first category of networks [10], [11], [12], [13] usually uses a multistream (pathway) architecture. These architectures are also referred to as cascaded networks. In these networks, the input images are sampled from the same original subject and then input separately for feature learning and dense prediction. In contrast to multistream architecture, another U-Net-based method [14], whose lower layers are responsible for transforming the input into low-resolution representations whereas deeper layers produce pixelwise predictions using upsampling operations, has received more attention for brain tumor segmentation [15], [16], [17], [18], [19], [20], [21]. Although the 2D DCNNs-based methods have achieved promising performance, they only take into account the local dependencies of the ground truth but ignore the appearance and spatial consistency. The use of common 2D DCNNs-based methods on 3D MR images fails to detect some important spatial information. To address this issue, 3D DCNNs-based models [22], [23], [24], [25], [26], whose convolutional layers consist of 3D filters, were developed. Similar to 2D DCNNs-based methods, the general architectures of 3D DCNNs are classified into multistream and U-Net. Different from 2D DCNNs-based methods, these 3D DCNNs-based approaches directly extract features from MRI data without fusing the final results. Hence, the processing pipeline of 3D DCNNs is easier to implement than that of 2D DCNNs.

Although the aforementioned networks have greatly promoted the development of medical image segmentation, some challenges still exist in brain tumor segmentation. On the one hand, most existing DCNNs-based approaches strengthen the representative ability by stacking more layers. However, simply increasing layers results in exploding/vanishing gradients, which has a negative influence on training. Additionally, the networks usually feed the feature maps produced by the last hidden layer to the final classifier. Under this stacked method, the features of previous layers are indirectly input to the end of the network, and thus, the gradient flow may be impeded during training. On the other hand, brain tumors appear in a variety of sizes and shapes in MR images. This requires DCNNs-based models to be able to analyze and handle tumors of various scales. Therefore, single-scale-based networks have limitations in feature computation to address this challenge. In this manuscript, we propose a 3D dense connectivity convolutional neural network equipped with an atrous convolutional feature pyramid and 3D deep supervision mechanism to comprehensively segment brain tumors. The proposed network is referred to as DenseAFPNet. The contributions of this study are summarized as follows:

  • To alleviate the difficulties in training deep networks and improve the efficiency of the gradient flow, a densely connected convolutional neural network is developed as the backbone for learning dense features for brain tumors via MR images.

  • A 3D hierarchical feature pyramid using 3D atrous convolutional layers is connected with the end of the backbone to learn multiscale features of brain tumors. The proposed pyramid provides the network with different receptive fields and thus fuses the contextual information with the lesions to improve the segmentation performance.

  • Through the introduction of an auxiliary loss function for each pyramidal hierarchy, a 3D deep supervision mechanism is further proposed. Such a mechanism guides the optimization of both the lower and upper hierarchies to reinforce the propagation of the gradient flow during training, and thus this mechanism makes the network learn more discriminating and powerful features. The mechanism is not only a strong regularization technique as a prompter to improve the convergence of the network but also provides an effective strategy for addressing the challenge of overfitting.

Section snippets

Deep architecture for representation learning

A deeper network with reasonable size is widely considered to be able to achieve better discriminating performance than a shallower network [27]. Thus, because of this observation, very deep architectures have been developed for most neural networks for effective feature learning in recent years [28], [29]. However, the opposite effect occurs from simply stacking convolutional layers to deepen the network because too deep of networks always result in the problem of exploding/vanishing gradients 

Overview

Fig. 1 illustrates the proposed DenseAFPNet architecture, which is a fully convolutional neural network and consists of the following key modules: backbone network, 3D atrous convolutional feature pyramid, and classification layers. Inspired by the conclusion proposed by DeepMedic [23] that the input size of MRI scans of 25 × 25 × 25 was suitable for training a the 3D convolutional network in a patch-by-patch manner, we follow this viewpoint and use 25 × 25 × 25 as the input size to feed the

Datasets

We take advantage of public datasets, namely, the BRATS 2013 [42], 2015 [43], and 2018 [44] datasets, to train and evaluate the proposed model. The clinical images from BRATS 2013 were acquired at four different centers, i.e., Heidelberg University, Debrecen University, Bern University, and Massachusetts General Hospital [43]. The clinical instances of both BRATS 2015 and 2018 came from contributions from the Cancer Imaging Archive (TCIA) repository, Heidelberg University, and the Center for

Discussion

Due to the factors of the irregular appearances of tumors, fuzzy boundaries caused by invasive growth, and varying intensities produced by different configurations of MRI machines, brain tumor segmentation is a challenging task. To address this challenge, most existing deep learning based methods improve the discriminating ability of the models by simply stacking more layers. The training within this type of network not only causes gradients to explode or vanish but also has impeded feature

Conclusion

We developed a 3D dense connectivity fully convolutional neural network as a backbone for end-to-end volumetric brain tumor segmentation of MR images. To fuse multiscale contextual features tor improve the segmentation performance, a 3D atrous convolutional feature pyramid is integrated with the backbone. Through the control of the atrous rate of the convolutions in each hierarchy of the feature pyramid, the network can adjust distinct receptive fields to capture multiscale contexts.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Acknowledgments

This research is funded by Chongqing Research Program of Basic Research and Frontier Technology (NO. cstc2018jcyjAX0287), the graduate research and innovation foundation of Chongqing (Grant No. CYS17023).

Zexun Zhou received his B.Sc. degree in computer science and technology from the Department of Computer Science, Hanshan Normal University, China, in 2009. He received his M.S. degree in computer software and theory from the College of Computer Science and Engineering, Northwest Normal University, China, in 2013. He is currently pursuing the Ph.D. degree in College of Computer Science at Chongqing University, China. His research interests include machine learning, computer vision and deep

References (65)

  • GeethuM. et al.

    MRI based medical image analysis: Survey on brain tumor grade classification

    Biomed. Signal Process. Control

    (2018)
  • M. Soltaninejad, X. Ye, G. Yang, N. Allinson, T. Lambrou, et al. Brain tumour grading in different MRI protocols using...
  • SoltaninejadM. et al.

    Automated brain tumour detection and segmentation using superpixel-based extremely randomized trees in FLAIR MRI

    Int. J. Comput. Assist. Radiol. Surg.

    (2017)
  • YangG. et al.

    Morphometric model for discrimination between glioblastoma multiforme and solitary metastasis using three-dimensional shape analysis

    Magn. Reson. Med.

    (2016)
  • FelixR. et al.

    Tissue-type mapping of gliomas

    NeuroImage: Clinical

    (2019)
  • JonesT.L. et al.

    Brain tumor classification using the diffusion tensor image segmentation (D-SEG) technique

    Neuro-Oncology

    (2014)
  • YangG. et al.

    Discrimination between glioblastoma multiforme and solitary metastasis using morphological features derived from the p:q tensor decomposition of diffusion tensor imaging

    NMR Biomedicine

    (2014)
  • MohammadrezaS. et al.

    MRI Brain tumor segmentation using random forests and fully convolutional networks

  • O. Ronneberger, P. Fischer, T. Brox, U-net: Convolutional networks for biomedical image segmentation, in: International...
  • ChenH. et al.

    Brain tumor segmentation with deep convolutional symmetric neural network

    Neurocomputing

    (2019)
  • P. Bilinski, V. Prisacariu, Dense decoder shortcut connections for single-pass semantic segmentation, in: Proceedings...
  • A.V. Dalca, J. Guttag, M.R. Sabuncu, Anatomical priors in convolutional networks for unsupervised biomedical...
  • JiangH. et al.

    Multi-class multimodal semantic segmentation with an improved 3D fully convolutional networks

    Neurocomputing

    (2019)
  • DongH. et al.

    Automatic brain tumor detection and segmentation using u-net based fully convolutional networks

  • G. Urban, M. Bendszus, F.A. Hamprecht, J. Kleesiek, Multi-modal brain tumor segmentation using deep convolutional...
  • C. Ozgun, A. Abdulkadir, S.S. Lienkamp, T. Brox, O. Ronneberger, 3D U-Net: Learning Dense Volumetric Segmentation from...
  • F. Milletari, N. Navab, S.A. Ahmadi, V-net: Fully convolutional neural networks for volumetric medical image...
  • K. Tseng, Y. Lin, W. Hsu, C. Huang, Joint Sequence Learning and Cross-Modality Convolution for 3D Biomedical...
  • Yoshua BengioO.D.

    Shallow vs. deep sum-product networks

  • SimonyanK. et al.

    Very deep convolutional networks for large-scale image recognition

    (2014)
  • SzegedyC. et al.

    Going deeper with convolutions

    (2014)
  • X. Glorot, Y. Bengio, Understanding the difficulty of training deep feedforward neural networks, in: Proceedings of the...
  • Cited by (40)

    • MBANet: A 3D convolutional neural network with multi-branch attention for brain tumor segmentation from MRI images

      2023, Biomedical Signal Processing and Control
      Citation Excerpt :

      Zhou et al. [42] used atrous convolution instead of pooling to designed a 3D atrous convolution feature pyramid, and a 3D fully-connected conditional random field (CRF) is used for post-processing of the network. Zhou et al. [43] used 3D atrous convolution to design a new feature pyramid module and using 3D dense connections to build feature reuse, which can fuse more multi-scale contextual information. In the patient's brain MRI images, the tissue area of the tumor region is small, and the boundaries between the various modalities are not clear.

    View all citing articles on Scopus

    Zexun Zhou received his B.Sc. degree in computer science and technology from the Department of Computer Science, Hanshan Normal University, China, in 2009. He received his M.S. degree in computer software and theory from the College of Computer Science and Engineering, Northwest Normal University, China, in 2013. He is currently pursuing the Ph.D. degree in College of Computer Science at Chongqing University, China. His research interests include machine learning, computer vision and deep learning techniques.

    Zhongshi He received the B.Sc. degree in applied mathematics and the Ph.D. degree in computer science from Chongqing University, Chongqing, China, in 1987 and in 1996, respectively. He was a Postdoctoral Fellow in the School of Computer Science at the University of Witwatersrand in South Africa from September 1999 to August 2001. He is currently a Full Professor, Ph.D. Supervisor, and the Vice-Dean of the College of Computer Science and Technology at the Chongqing University. He is a Member of AIPR Professional Committee of China Federation of Computer, a candidate of the 322-key talent project of Chongqing, and a Science and Technology Leader in Chongqing. His research interests include machine learning and data mining, natural language computing, and image processing.

    Meifeng Shi is the lecturer of Chongqing University of Technology since 2018. She received her B.S. degree(2010) and Ph.D. degree(2017) in Computer Science at Chongqing University. Her research interests include Intelligence Computation, Multi-objective Optimization, Image Processing ,Multi-agent systems and Machine learning.

    Jinglong Du received the M.Sc. degrees from College of Computer Science, Chongqing University, Chongqing, China, in 2016. He is currently pursuing the Ph.D. degree at the College of Computer Science, Chongqing University. His research interests include machine learning and medical image processing and analysis, with a focus on MRI super-resolution reconstruction and deep learning.

    Dingding Chen received his B.Sc. degree in Computer Science from Faculty of Information Engineering and Automation of Kunming University of Science and Technology in 2015. He is currently pursing the Ph.D. degree in Computer Science at Chongqing University. His research interests include multi-agent systems, Distributed Constraint Optimization Problems and machine learning.

    View full text