Skip to main content

Decomposition-and-Fusion Network for HE-Stained Pathological Image Classification

  • Conference paper
  • First Online:
Intelligent Computing Theories and Application (ICIC 2021)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 12838))

Included in the following conference series:

  • 1448 Accesses

Abstract

Building upon the clinical evidence supporting that decomposing a pathological image into different components can improve diagnostic value, in this paper we propose a Decomposition-and-Fusion Network (DFNet) for HE-stained pathological image classification. The medical goal of using HE-stained pathological images is to distinguish between nucleus, cytoplasm and extracellular matrix, thereby displaying the overall layouts of cells and tissues. We embed this most basic medical knowledge into a deep learning framework that decomposes a pathological image into cell nuclei and the remaining structures (that is, cytoplasm and extracellular matrix). With such decomposed pathological images, DFNet first extracts independent features using three independent CNN branches, and then gradually merges these features together for final classification. In this way, DFNet is able to learn more representative features with respect to different structures and hence improve the classification performance. Experimental results on two different datasets with various cancer types show that the DFNet achieves competitive performance.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Deng, S., Zhang, X., Yan, W., et al.: Deep learning in digital pathology image analysis: a survey. Front. Med. 1–18 (2020)

    Google Scholar 

  2. Zhou, S.K., Greenspan, H., Davatzikos, C., et al.: A review of deep learning in medical imaging: image traits, technology trends, case studies with progress highlights, and future promises. arXiv preprint arXiv:2008.09104 (2020)

  3. Lin, W.-A., Liao, H., Peng, C., et al.: DudoNet: dual domain network for CT metal artifact reduction. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 10512–10521 (2019)

    Google Scholar 

  4. Li, Z., Li, H., Han, H., Shi, G., Wang, J., Zhou, S.: Encoding CT anatomy knowledge for unpaired chest x-ray image decomposition. In: Shen, D., et al. (eds.) MICCAI 2019. LNCS, vol. 11769, pp. 275–283. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-32226-7_31

    Chapter  Google Scholar 

  5. Gozes, O., Greenspan, H.: Lung structures enhancement in chest radiographs via CT based FCNN training. In: Stoyanov, D., et al. (eds.) RAMBO/BIA/TIA -2018. LNCS, vol. 11040, pp. 147–158. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-00946-5_16

    Chapter  Google Scholar 

  6. Hayakawa, T., Prasath, V.S., Kawanaka, H., et al.: Computational nuclei segmentation methods in digital pathology: a survey. Arch. Comput. Methods Eng. 1–13 (2019)

    Google Scholar 

  7. Litjens, G., Kooi, T., Bejnordi, B.E., et al.: A survey on deep learning in medical image analysis. Med. Image Anal. 42(9), 60–88 (2017)

    Article  Google Scholar 

  8. Chen, L.-C., Zhu, Y., Papandreou, G., Schroff, F., Adam, H.: Encoder-decoder with atrous separable convolution for semantic image segmentation. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11211, pp. 833–851. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01234-2_49

    Chapter  Google Scholar 

  9. Chollet, F.: Xception: deep learning with depthwise separable convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1251–1258 (2017)

    Google Scholar 

  10. Szegedy, C., Vanhoucke, V., Ioffe, S., et al.: Rethinking the inception architecture for computer vision, 2818–2826 (2015)

    Google Scholar 

  11. Lin, M., Chen, Q., Yan, S.: Network in network. arXiv preprint arXiv:1312.4400 (2013)

  12. Yan, R., Ren, F., Wang, Z., et al.: Breast cancer histopathological image classification using a hybrid deep neural network. Methods 173, 52–60 (2020)

    Article  Google Scholar 

  13. Awan, R., Sirinukunwattana, K., Epstein, D., et al.: Glandular morphometrics for objective grading of colorectal adenocarcinoma histology images. Sci. Rep. 7(1), 1–12 (2017)

    Article  Google Scholar 

  14. Kumar, N., Verma, R., Sharma, S., et al.: A dataset and a technique for generalized nuclear segmentation for computational pathology. IEEE Trans. Med. Imaging 36(7), 1550–1560 (2017)

    Article  Google Scholar 

  15. Shaban, M., Awan, R., Fraz, M.M., et al.: Context-aware convolutional neural network for grading of colorectal cancer histology images. IEEE Trans. Med. Imaging 2395–2405 (2020)

    Google Scholar 

  16. Vang, Y.S., Chen, Z., Xie, X.: Deep learning framework for multi-class breast cancer histology image classification. In: International Conference Image Analysis and Recognition, pp. 914–922 (2018)

    Google Scholar 

  17. Golatkar, A., Anand, D., Sethi, A.: Classification of breast cancer histology using deep learning. In: Campilho, A., Karray, F., ter Haar Romeny, B. (eds.) ICIAR 2018. LNCS, vol. 10882, pp. 837–844. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-93000-8_95

    Chapter  Google Scholar 

  18. Sirinukunwattana, K., Alham, N.K., Verrill, C., Rittscher, J.: Improving whole slide segmentation through visual context - a systematic study. In: Frangi, A.F., Schnabel, J.A., Davatzikos, C., Alberola-López, C., Fichtinger, G. (eds.) MICCAI 2018. LNCS, vol. 11071, pp. 192–200. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-00934-2_22

    Chapter  Google Scholar 

  19. He, K., Zhang, X., Ren, S., et al.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778 (2016)

    Google Scholar 

Download references

Acknowledgement

This paper is supported by National Key Research and Development Program of China (No. 2017YFE0103900 and 2017YFA0504702), the NSFC projects Grant (No. 61932018, 62072441 and 62072280), Beijing Municipal Natural Science Foundation Grant (No. L182053), Peking University International Hospital Research Grant (No. YN2018ZD05).

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to Fei Ren or Fa Zhang .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2021 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Yan, R. et al. (2021). Decomposition-and-Fusion Network for HE-Stained Pathological Image Classification. In: Huang, DS., Jo, KH., Li, J., Gribova, V., Premaratne, P. (eds) Intelligent Computing Theories and Application. ICIC 2021. Lecture Notes in Computer Science(), vol 12838. Springer, Cham. https://doi.org/10.1007/978-3-030-84532-2_18

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-84532-2_18

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-84531-5

  • Online ISBN: 978-3-030-84532-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics