Skip to main content

UAM-Net: An Attention-Based Multi-level Feature Fusion UNet for Remote Sensing Image Segmentation

  • Conference paper
  • First Online:
Pattern Recognition and Computer Vision (PRCV 2023)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 14428))

Included in the following conference series:

  • 358 Accesses

Abstract

Semantic segmentation of Remote Sensing Images (RSIs) is an essential application for precision agriculture, environmental protection, and economic assessment. While UNet-based networks have made significant progress, they still face challenges in capturing long-range dependencies and preserving fine-grained details. To address these limitations and improve segmentation accuracy, we propose an effective method, namely UAM-Net (UNet with Attention-based Multi-level feature fusion), to enhance global contextual understanding and maintain fine-grained information. To be specific, UAM-Net incorporates three key modules. Firstly, the Global Context Guidance Module (GCGM) integrates semantic information from the Pyramid Pooling Module (PPM) into each decoder stage. Secondly, the Triple Attention Module (TAM) effectively addresses feature discrepancies between the encoder and decoder. Finally, the computation-effective Linear Attention Module (LAM) seamlessly fuses coarse-level feature maps with multiple decoder stages. With the corporations of these modules, UAM-Net significantly outperforms the most state-of-the-art methods on two popular benchmarks.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 59.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 79.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Bai, H., Cheng, J., Huang, X., Liu, S., Deng, C.: HCANet: a hierarchical context aggregation network for semantic segmentation of high-resolution remote sensing images. IEEE Geosci. Remote Sens. Lett. 19, 1–5 (2021)

    Google Scholar 

  2. Chen, L.C., Papandreou, G., Schroff, F., Adam, H.: Rethinking atrous convolution for semantic image segmentation. arXiv preprint arXiv:1706.05587 (2017)

  3. Chen, L., et al.: SCA-CNN: spatial and channel-wise attention in convolutional networks for image captioning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5659–5667 (2017)

    Google Scholar 

  4. Diakogiannis, F.I., Waldner, F., Caccetta, P., Wu, C.: ResUNet-a: a deep learning framework for semantic segmentation of remotely sensed data. ISPRS J. Photogramm. Remote. Sens. 162, 94–114 (2020)

    Article  Google Scholar 

  5. Fu, J., et al.: Dual attention network for scene segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3146–3154 (2019)

    Google Scholar 

  6. Griffiths, P., Nendel, C., Hostert, P.: Intra-annual reflectance composites from Sentinel-2 and Landsat for national-scale crop and land cover mapping. Remote Sens. Environ. 220, 135–151 (2019)

    Google Scholar 

  7. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778 (2016)

    Google Scholar 

  8. Li, R., Zheng, S., Duan, C., Su, J., Zhang, C.: Multistage attention resU-Net for semantic segmentation of fine-resolution remote sensing images. IEEE Geosci. Remote Sens. Lett. 19, 1–5 (2021)

    Google Scholar 

  9. Li, R., et al.: Multi attention network for semantic segmentation of fine-resolution remote sensing images. IEEE Trans. Geosci. Remote Sens. 60, 1–13 (2021)

    Google Scholar 

  10. Li, R., Zheng, S., Zhang, C., Duan, C., Wang, L., Atkinson, P.M.: ABCNet: attentive bilateral contextual network for efficient semantic segmentation of fine-resolution remotely sensed imagery. ISPRS J. Photogramm. Remote. Sens. 181, 84–98 (2021)

    Article  Google Scholar 

  11. Lin, T.Y., Dollár, P., Girshick, R., He, K., Hariharan, B., Belongie, S.: Feature pyramid networks for object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2117–2125 (2017)

    Google Scholar 

  12. Misra, D., Nalamada, T., Arasanipalai, A.U., Hou, Q.: Rotate to attend: convolutional triplet attention module. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 3139–3148 (2021)

    Google Scholar 

  13. Oda, H., et al.: BESNet: boundary-enhanced segmentation of cells in histopathological images. In: Frangi, A.F., Schnabel, J.A., Davatzikos, C., Alberola-López, C., Fichtinger, G. (eds.) MICCAI 2018. LNCS, vol. 11071, pp. 228–236. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-00934-2_26

    Chapter  Google Scholar 

  14. Park, J., Woo, S., Lee, J., Kweon, I.S.: BAM: bottleneck attention module. In: British Machine Vision Conference 2018, BMVC 2018, Newcastle, UK, September 3–6, 2018. p. 147. BMVA Press (2018)

    Google Scholar 

  15. Picoli, M.C.A., et al.: Big earth observation time series analysis for monitoring Brazilian agriculture. ISPRS J. Photogramm. Remote. Sens. 145, 328–339 (2018)

    Article  Google Scholar 

  16. Ronneberger, Olaf, Fischer, Philipp, Brox, Thomas: U-Net: convolutional networks for biomedical image segmentation. In: Navab, Nassir, Hornegger, Joachim, Wells, William M.., Frangi, Alejandro F.. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234–241. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-24574-4_28

    Chapter  Google Scholar 

  17. Samie, A., et al.: Examining the impacts of future land use/land cover changes on climate in Punjab province, Pakistan: implications for environmental sustainability and economic growth. Environ. Sci. Pollut. Res. 27, 25415–25433 (2020)

    Article  Google Scholar 

  18. Tong, X.Y., et al.: Land-cover classification with high-resolution remote sensing images using transferable deep models. Remote Sens. Environ. 237, 111322 (2020)

    Article  Google Scholar 

  19. Wang, H., Wang, Y., Zhang, Q., Xiang, S., Pan, C.: Gated convolutional neural network for semantic segmentation in high-resolution images. Remote Sens. 9(5), 446 (2017)

    Article  Google Scholar 

  20. Wang, L., Li, R., Duan, C., Zhang, C., Meng, X., Fang, S.: A novel transformer based semantic segmentation scheme for fine-resolution remote sensing images. IEEE Geosci. Remote Sens. Lett. 19, 1–5 (2022)

    Google Scholar 

  21. Wang, L., Li, R., Wang, D., Duan, C., Wang, T., Meng, X.: Transformer meets convolution: a bilateral awareness network for semantic segmentation of very fine resolution urban scene images. Remote Sens. 13(16), 3065 (2021)

    Article  Google Scholar 

  22. Wang, L., et al.: UNetFormer: a UNet-like transformer for efficient semantic segmentation of remote sensing urban scene imagery. ISPRS J. Photogramm. Remote. Sens. 190, 196–214 (2022)

    Article  Google Scholar 

  23. Woo, Sanghyun, Park, Jongchan, Lee, Joon-Young., Kweon, In So.: CBAM: convolutional block attention module. In: Ferrari, Vittorio, Hebert, Martial, Sminchisescu, Cristian, Weiss, Yair (eds.) ECCV 2018. LNCS, vol. 11211, pp. 3–19. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01234-2_1

    Chapter  Google Scholar 

  24. Yin, H., Pflugmacher, D., Li, A., Li, Z., Hostert, P.: Land use and land cover change in inner Mongolia-understanding the effects of China’s re-vegetation programs. Remote Sens. Environ. 204, 918–930 (2018)

    Article  Google Scholar 

  25. Zhang, C., et al.: Joint deep learning for land cover and land use classification. Remote Sens. Environ. 221, 173–187 (2019)

    Article  Google Scholar 

  26. Zhang, Z., Liu, Q., Wang, Y.: Road extraction by deep residual U-Net. IEEE Geosci. Remote Sens. Lett. 15(5), 749–753 (2018)

    Article  Google Scholar 

  27. Zhao, H., Shi, J., Qi, X., Wang, X., Jia, J.: Pyramid scene parsing network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2881–2890 (2017)

    Google Scholar 

Download references

Acknowledgements

This work is supported by Industry-University Cooperation Project of Fujian Science and Technology Department (No. 2021H6035), the Science and Technology Planning Project of Fujian Province (No. 2021J011191), and Fujian Key Technological Innovation and Industrialization Projects (No. 2023XQ023), and Fu-Xia-Quan National Independent Innovation Demonstration Project (No. 2022FX4).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yun Wu .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Cao, Y., Jiang, N., Wang, DH., Wu, Y., Zhu, S. (2024). UAM-Net: An Attention-Based Multi-level Feature Fusion UNet for Remote Sensing Image Segmentation. In: Liu, Q., et al. Pattern Recognition and Computer Vision. PRCV 2023. Lecture Notes in Computer Science, vol 14428. Springer, Singapore. https://doi.org/10.1007/978-981-99-8462-6_22

Download citation

  • DOI: https://doi.org/10.1007/978-981-99-8462-6_22

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-99-8461-9

  • Online ISBN: 978-981-99-8462-6

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics