ISCA Archive Interspeech 2018
ISCA Archive Interspeech 2018

Air-Tissue Boundary Segmentation in Real-Time Magnetic Resonance Imaging Video Using Semantic Segmentation with Fully Convolutional Networks

Valliappan CA, Renuka Mannem, Prasanta Kumar Ghosh

In this paper, we propose a new technique for the segmentation of the Air-Tissue Boundaries (ATBs) in the vocal tract from the real-time magnetic resonance imaging (rtMRI) videos of the upper airway in the midsagittal plane. The proposed technique uses the approach of semantic segmentation using the Deep learning architecture called Fully Convolutional Networks (FCN). The architecture takes an input image and produces images of the same size with air and tissue class labels at each pixel. These output images are post processed using morphological filling and image smoothing to predict realistic ATBs. The performance of the predicted contours is evaluated using Dynamic Time Warping (DTW) distance between the manually annotated ground truth contours and the predicted contours. Four fold experiments with four subjects from USC-TIMIT corpus (with ~2900 training images in every fold) demonstrate that the proposed FCN based approach has 8.87% and 9.65% lesser average error than the baseline Maeda Grid based scheme, for the lower and upper ATBs respectively. In addition, the proposed FCN based rtMRI segmentation achieves an average pixel classification accuracy of 99.05% across all subjects.


doi: 10.21437/Interspeech.2018-1939

Cite as: CA, V., Mannem, R., Ghosh, P.K. (2018) Air-Tissue Boundary Segmentation in Real-Time Magnetic Resonance Imaging Video Using Semantic Segmentation with Fully Convolutional Networks. Proc. Interspeech 2018, 3132-3136, doi: 10.21437/Interspeech.2018-1939

@inproceedings{ca18_interspeech,
  author={Valliappan CA and Renuka Mannem and Prasanta Kumar Ghosh},
  title={{Air-Tissue Boundary Segmentation in Real-Time Magnetic Resonance Imaging Video Using Semantic Segmentation with Fully Convolutional Networks}},
  year=2018,
  booktitle={Proc. Interspeech 2018},
  pages={3132--3136},
  doi={10.21437/Interspeech.2018-1939}
}