Abstract
3-D visualisation of cellular structures within the placenta is important for advancing research into the factors determining fetal growth, which are linked to chronic disease risk and quality of lifelong health. 2-D analysis can be challenging, and spatial interaction between structures can be easily missed, but obtaining 3-D structural images is extremely labour-intensive due to the high level of rigorous manual processing required. Deep neural networks are used to automate this previously manual process to construct fast and accurate 3-D structural images, which can be used for 3-D image analysis. The deep networks described in this chapter are trained to label both single cell, a fibroblast and a pericyte, and multicellular, endothelial, structures from within serial block-face scanning electron microscopy placental imaging. Automated labels are equivalent, pixel-to-pixel, to manual labels by over 98% on average over all cell structures and network architectures, and are able to successfully label unseen regions and stacks.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Mackay, B., et al.: Automated 3-D labelling of fibroblasts and endothelial cells in SEM-imaged placenta using deep learning. In: Proceedings of the 13th International Joint Conference on Biomedical Engineering Systems and Technologies, BIOIMAGING 2020, INSTICC, vol. 2, pp. 46–53. SciTePress, Malta (2020)
Palaiologou, E., et al.: Human placental villi contain stromal macrovesicles associated with networks of stellate cells. J Anat. 236(1), 132–141 (2019)
Lewis, R.M., Cleal, J.K., Hanson, M.A.: Review: placenta, evolution and lifelong health. Placenta 33, S28–S32 (2012)
Lewis, R.M., Pearson-Farr, J.E.: Multiscale three-dimensional imaging of the placenta. Placenta. https://doi.org/10.1016/j.placenta.2020.01.016. (Article in press)
Burton, G.J.: Scanning electron microscopy of intervillous connections in the mature human placenta. J. Anat. 147, 245–254 (1986)
Mayhew, T.M.: Morphomics: an integral part of systems biology of the human placenta. Placenta 36(4), 329–340 (2015)
Wang, Y., Zhao, S.: Vascular Biology of the Placenta. Morgan & Claypool, San Rafael (2010)
Cahill, L.S., et al.: Feto- and utero-placental vascular adaptions to chronic maternal hypoxia in the mouse. J. Physiol 596(15), 3285–3297 (2018)
Kherlopian, A.R., et al.: A review of imaging techniques for systems biology. BMC Syst Biol. 2, 74 (2008)
Denk, W., Horstmann, H.: Serial block-face scanning electron microscopy to reconstruct three-dimensional tissue nanostructure. PLoS Biol. 2(100), e329 (2004)
Kazemian, A., Hooshmandabbasi, R., Schraner, E.M., Boos, A., Klisch, K.: Evolutionary implications of fetal and maternal microvillous surfaces in epitheliochorial placentae. J. Morphol. 280(4), 615–622 (2019)
Palaiologou, E., et al.: Serial block-face scanning electron microscopy reveals novel intercellular connections in human term placental microvasculature. J. Anat. 237, 1–9 (2020)
Deerinck, T.J., Bushong, E.A., Lev-Ram, V., Shu, X., Tsien, R.Y., Ellisman, M.H.: Enhancing serial block-face scanning electron microscopy to enable high resolution 3-D nanohistology of cells and tissues. Microsc. Microanal. 16(2), 1138–1139 (2010)
Zachow, S., Zilske, M., Hege, H.: 3-D reconstruction of individual anatomy from medical image data: segmentation and geometry processing. Konrad-Zuse-Zentrum für Informationstechnik Berlin, Berlin (2007)
Pugin, E., Zhiznyakov, A.: Histogram method of image binarization based on fuzzy pixel representation. In: Dynamics of Systems, Mechanisms and Machines 2017 (Dynamics), p. 17467698. IEEE, Omsk (2017)
Yoo, K.D., et al.: A machine learning approach using survival statistics to predict graft survival in kidney transplant recipients: a multicenter cohort study. Sci. Rep. 7, 8904 (2017)
Jamshidi, A., Pelletier, J., Martel-Pelletier, J.: Machine-learning-based patient-specific prediction models for knee osteoarthritis. Nat. Rev. Rheumatol. 15, 49–60 (2019)
Google Cloud: Vision AI| Derive Image Insights via ML. Google. https://cloud.google.com/vision/. Accessed 14 Nov 2019
Grant-Jacob, J.A., et al.: Real-time particle pollution sensing using machine learning. Opt. Express 26(21), 27237–27246 (2018)
Krizhevsky, A., Sutskever, L., Hinton, G.: ImageNet classification with deep convolutional neural networks. In: Proceedings of the Advances in Neural Information Processing Systems 25, pp. 1090–1098 (2012)
Szegedy, C., et al.: Intriguing properties of neural networks. arXiv:1312.6199v4 (2014)
Schmidhuber, J.: Deep learning in neural networks: an overview. Neural Netw. 61, 85–117 (2015)
Simonyan, K., Zisserman, A.: Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv:1409.1556v6 (2015)
Kermany, D.S., Goldbaum, M., Cai, W., Lewis, M.A., Xia, H., Zhang, K.: Identifying medical diagnoses and treatable diseases by image-based deep learning. Cell 172(5), 1122–1131 (2018)
Litjens, G., et al.: A survey on deep learning in medical image analysis. Med. Image Anal. 42, 60–88 (2017)
Esteva, A., et al.: Dermatologist-level classification of skin cancer with deep neural networks. Nature 542, 115–118 (2017)
Haenssle, H.A., et al.: Man against machine: diagnostic performance of a deep learning convolutional neural network for dermoscopic melanoma recognition in comparison to 58 dermatologists. Ann. Oncol. 29, 1836–1842 (2018)
Le Cun, Y., Huang, F. J., Bottou, L.: Learning methods for generic object recognition with invariance to pose and lighting. In: Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, vol. 2, p. II-97 (2004)
Lo, S.B., Chan, H., Freedman, M.T., Min, S.K.: Artificial convolution neural network for medical image pattern recognition. Neural Netw. 8(7–8), 1201–1214 (1995)
Le Cun, Y., Bengio, Y.: Convolutional Networks for Images Speech and Time-Series. The Handbook of Brain Theory and Neural Networks. MIT Press, Cambridge (1995)
Sermanet, P., Eigen, D., Zhang, X., Mathieu, M., Fergus, R., Le Cun, Y.: OverFeat: Integrated Recognition, Localization and Detection using Convolutional Networks. arXiv:1312.6229v4 (2013)
Donahue, J., et al.: 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2625–2634 (2015)
Zhang, X., Le Cun, Y.: Text Understanding from Scratch. arXiv:1502.01710v5 (2015)
Dong, A., He, K., Tang, X.: Image super-resolution using deep convolutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 38(2), 295–307 (2016)
Simard, D., Steinkraus, P.Y., Platt, J.C.: Best practices for convolutional neural networks. In: Proceedings of the Document Analysis and Recognition, pp. 958–963 (2003)
Vaillant, R., Monrocq, C., Le Cun, Y.: Original approach for the localisation of objects in images. Proc. Vis. Image Sig. Process. 141, 245–250 (1994)
Nowlan, S., Platt, J.: A convolutional neural network hand tracker. In: Advances in Neural Information Processing Systems, pp. 901–908 (1995)
Le Cun, Y., et al.: Handwritten digit recognition with a back-propagation network. In: Advances in Neural Information Processing Systems (1990)23
Hubel, D.H., Wiesel, T.N.: Receptive fields, binocular interaction, and functional architecture in the cat’s visual cortex. J. Physiol. 106, 160 (1962)
LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521, 436–444 (2015)
Fukushima, K., Miyake, S., Ito, T.: Neocognitron: a neural network model for a mechanism of visual pattern recognition. IEEE Trans. Syst. Man Cybern. 13(5), 826–834 (1983)
Felleman, D.J., Essen, D.C.V.: Distributed hierarchical processing in the primate cerebral cortex. Cereb. Cortex 1, 1–47 (1991)
Mirza, M., Osindero, S.: Conditional Generative Adversarial Nets. arXiv:1411.1784v1 (2014)
Pathak, D., Krahenbuhl, P., Donahue, J., Darrell, T., Efros, A. A.: Context Encoders: Feature Learning by Inpainting. arXiv:1604.07379v2 (2016)
Isola, P., Zhue, J., Zhou T., Efros, A.A.: Image-to-Image Translation with Conditional Adversarial Networks. arXiv:1611.07004v3 (2018)
Esteva, A., et al.: A guide to deep learning in healthcare. Nat. Med. 25, 24–29 (2019)
Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234–241. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-24574-4_28
Long, J., Shelhamer, E., Darrell, T.: Fully Convolutional Networks for Semantic Segmentation. arXiv:1411.4038 (2014)
Falk, T., et al.: U-Net: deep learning for cell counting, detection, and morphometry. Nat. Methods 16, 67–70 (2019)
Barreto, R.S.N., Romagnolli, P., Cereta, A.D., Coimbra-Campos, L.M.C., Birbrair, A., Miglino, M.A.: Pericytes in the placenta: role in placental development and homeostasis. In: Birbrair, A. (ed.) Pericyte Biology in Different Organs. AEMB, vol. 1122, pp. 125–151. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-11093-2_8
Bergers, G., Song, S.: The role of pericytes in blood-vessel formation and maintenance. Neuro Oncol. 7(4), 452–464 (2005)
Grant-Jacob, J.A., et al.: A neural lens for super-resolution biological imaging. J. Phys. Commun. 3(6), 065004 (2019)
Santosh, K.C., Roy, P.P.: Arrow detection in biomedical images using sequential classifier. Int. J. Mach. Learn. Cybern. 9(6), 993–1006 (2016). https://doi.org/10.1007/s13042-016-0623-y
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Mackay, B.S., Grant-Jacob, J.A., Eason, R.W., Lewis, R., Mills, B. (2021). Deep Learning for the Automated Feature Labelling of 3-Dimensional Imaged Placenta. In: Ye, X., et al. Biomedical Engineering Systems and Technologies. BIOSTEC 2020. Communications in Computer and Information Science, vol 1400. Springer, Cham. https://doi.org/10.1007/978-3-030-72379-8_6
Download citation
DOI: https://doi.org/10.1007/978-3-030-72379-8_6
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-72378-1
Online ISBN: 978-3-030-72379-8
eBook Packages: Computer ScienceComputer Science (R0)