Skip to main content

GAN-Based Defogging and Multiscale Fusion Approach for UAV-Based Seagrass Bed Imagery Semantic Segmentation in Challenging Marine Environments

  • Conference paper
  • First Online:
Data Science (ICPCSEE 2024)

Abstract

Seagrass beds, as one of China’s “Three Major Marine Ecological Systems,” have consistently been a focal point of monitoring efforts. However, the quality of UAV-captured aerial images depicting seagrass beds is often compromised by atmospheric conditions, particularly during foggy weather, resulting in a reduction in overall feature expression. This situation poses challenges such as weak feature representation and low contrast, making accurate identification of seagrass beds in complex marine environments a significant challenge.

To address these challenges, this paper focuses on accurately identifying sea grass beds in complex marine environments. First, a DefoggingGAN network is proposed to mitigate the impact of foggy weather on the quality of seagrass bed images. This network, which is based on generative adversarial networks (GANs), is designed for image defogging. Second, for the precise identification of seagrass bed images captured by UAVs, a segmentation network for the Seagrass Bed Imagery Segmentation Network (SBISNet) is developed. This network incorporates an attention mechanism to capture context modules, focusing on high-resolution and low-resolution feature maps to obtain the global context. Additionally, a multiscale convolutional attention module is introduced to achieve the fusion of multiscale features. Furthermore, due to the limited availability of seagrass bed datasets in UAV scenarios, this paper utilizes UAVs to collect seagrass bed images and establishes a dataset named the Seagrass Bed Dataset. This research contributes to the broader exploration of cutting-edge technologies, particularly in the context of edge computing and the IoT, within the realm of UAV applications for environmental monitoring.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

Database Availability Statement

The authors will supply the relevant data in response to reasonable requests.

References

  1. Bollard, B., Doshi, A., Gilbert, N., et al.: Drone technology for monitoring protected areas in remote and fragile environments. Drones 6(2), 42 (2022)

    Article  Google Scholar 

  2. Kim, S., Lee, C.W., Park, H.J., et al.: Piloting an unmanned aerial vehicle to explore the floristic variations of inaccessible cliffs along Island coasts. Drones 7(2), 140 (2023)

    Article  Google Scholar 

  3. Yang, Y., Wang, C., Liu, R., et al.: Self-augmented unpaired image dehazing via density and depth decomposition. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 2037–2046 (2022)

    Google Scholar 

  4. Tahara, S., Sudo, K., Yamakita, T., et al.: Species level mapping of a seagrass bed using an unmanned aerial vehicle and deep learning technique. PeerJ 10, e14017 (2022)

    Article  Google Scholar 

  5. Goodfellow, I., Pouget-Abadie, J., Mirza, M., et al.: Generative adversarial networks. Commun. ACM 60(11), 139–144 (2020)

    Article  MathSciNet  Google Scholar 

  6. Dong, H., Pan, J., Xiang, L., et al.: Multiscale boosted dehazing network with dense feature fusion. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 2157–2167 (2020)

    Google Scholar 

  7. Woo, S., Park, J., Lee, J.Y., et al.: CBAM: convolutional block attention module. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 3–19 (2018)

    Google Scholar 

  8. Yu, C., Gao, C., Wang, J., et al.: Bisenet v2: bilateral network with guided aggrega - tion for real-timesemantic segmentation. Int. J. Comput. Vision 129, 3051–3068 (2021)

    Article  Google Scholar 

  9. Wang, J., Gou, C., Wu, Q., et al.: Rtformer: efficient design for real-time semantic segmentation with transformer. arXiv preprint arXiv:2210.07124 (2022)

  10. Guo, M.H., Lu, C.Z., Hou, Q., et al.: Segnext: rethinking convolutional attention design for semantic segmentation. arXiv preprint arXiv: 2209.08575 (2022)

    Google Scholar 

  11. Mirza, M., Osindero, S.: Conditional generative adversarial nets. arXiv preprint arXiv:1411.1784 (2014)

  12. Radford, A., Metz, L., Chintala, S.: Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434 (2015)

  13. Chen, X., Duan, Y., Houthooft, R., et al.: Infogan: interpretable representation learning by information maximizing generative adversarial nets. Adv. Neural Inform. Process. Syst. 29 (2016)

    Google Scholar 

  14. Su, J.: O-GAN: extremely concise approach for autoencoding generative adversarial networks. arXiv preprint arXiv:1903.01931 (2019)

  15. Isola, P., Zhu, J.Y., Zhou, T., et al.: Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1125–1134 (2017)

    Google Scholar 

  16. Wang, T., Liu, M., Zhu, J.: pix2pixhd: highresolution image synthesis and semantic manipulation with conditional GANs. In: IEEE CVF Conference on Computer Vision and Pattern Recognition (2018)

    Google Scholar 

  17. Zhu, J.Y., Park, T., Isola, P., et al.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2223–2232 (2017)

    Google Scholar 

  18. Ledig, C., Theis, L., Huszár, F., et al.: Photo-realistic single image superresolution using a generative adversarial network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4681–4690 (2017)

    Google Scholar 

  19. Choi, Y., Choi, M., Kim, M., et al.: StarGAN: unified generative adversarial networks for multidomain image-to-image translation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8789–8797 (2018)

    Google Scholar 

  20. Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4401–4410 (2019)

    Google Scholar 

  21. Long, J., Shelhamer, E., Darrell, T.: Fully convolutional networks for semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3431–3440 (2015)

    Google Scholar 

  22. Badrinarayanan, V., Kendall, A., Cipolla, R.: SegNet: a deep convolutional encoder - decoder architecture for image segmentation. IEEE Trans. Patt. Anal. Mach. Intell. [18], 39(12), 2481–2495 (2017)

    Google Scholar 

  23. Niu, Z., Liu, W., Zhao, J., et al.: Deeplab-based spatial feature extraction for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. [19], 16(2), 251–255 (2018)

    Google Scholar 

  24. Chen, L.C., Papandreou, G., Kokkinos, I., et al.: Deeplab: semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs. IEEE Trans. Patt. Anal. Mach. Intell. 40(4), 834–848 (2017)

    Google Scholar 

  25. Si, H., Shi, Z., Hu, X., et al.: Image semantic segmentation based on improved deeplab v3 model. Int. J. Model. Identific. Control 36(2), 116–125 (2020)

    Article  Google Scholar 

  26. Si, Y., Gong, D., Guo, Y., et al.: An advanced spectral–spatial classification framework for hyperspectral imagery based on deeplab v3+. Appl. Sci. [22], 11(12), 5703 (2021)

    Google Scholar 

  27. Zhao, H., Shi, J., Qi, X., et al.: Pyramid scene parsing network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, [23], pp. 2881–2890 (2017)

    Google Scholar 

  28. Yu, C., Wang, J., Peng, C., et al.: Learning a discriminative feature network for semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1857–1866 (2018)

    Google Scholar 

  29. Paszke, A., Chaurasia, A., Kim, S., et al.: Enet: a deep neural network architecture for real-time semantic segmentation. arXiv preprint arXiv:1606.02147 (2016)

  30. Chaurasia, A., Culurciello, E.: Linknet: exploiting encoder representations for efficient semantic segmentation. In: 2017 IEEE Visual Communications and Image Processing (VCIP). IEEE [26], pp. 1–4 (2017)

    Google Scholar 

  31. Yu, C., Wang, J., Peng, C., et al.: Bisenet: Bilateral segmentation network for real-time semantic segmentation. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 325–341 (2018)

    Google Scholar 

  32. Li, H., Xiong, P., Fan, H., et al.: Dfanet: deep feature aggregation for realtime semantic segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 9522–9531 (2019)

    Google Scholar 

  33. Yu, C., Gao, C., Wang, J., et al.: Bisenet v2: bilateral network with guided aggregation for real-time semantic segmentation. Int. J. Comput. Vision 129, 3051–3068 (2021)

    Article  Google Scholar 

  34. Fan, M., Lai, S., Huang, J., et al.: Rethinking bisenet for real-time semantic segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 9716–9725 (2021)

    Google Scholar 

  35. Lim, B., Son, S., Kim, H., et al.: Enhanced deep residual networks for single image superresolution. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 136–144 (2017)

    Google Scholar 

  36. Cai, B., Xu, X., Jia, K., et al.: Dehazenet: an end-to-end system for single image haze removal. IEEE Trans. Image Process. 25(11), 5187–5198 (2016)

    Article  MathSciNet  Google Scholar 

  37. Li, B., Peng, X., Wang, Z., et al.: Aod-net: all-in-one dehazing network. In: Proceedings of the IEEE International Conference on Computer Vision, [33], pp. 4770–4778 (2017)

    Google Scholar 

  38. Qin, X., Wang, Z., Bai, Y., et al.: Ffa-net: feature fusion attention network for single image dehazing. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34. pp. 11908–11915 (2020)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Xiaoli Song .

Editor information

Editors and Affiliations

Ethics declarations

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Qu, L., Song, X., Zhang, M., Wang, J., Wen, R., Wang, S. (2024). GAN-Based Defogging and Multiscale Fusion Approach for UAV-Based Seagrass Bed Imagery Semantic Segmentation in Challenging Marine Environments. In: Xu, C., et al. Data Science. ICPCSEE 2024. Communications in Computer and Information Science, vol 2213. Springer, Singapore. https://doi.org/10.1007/978-981-97-8743-2_5

Download citation

Publish with us

Policies and ethics