Abstract
Underwater images suffer from various degradation, which can significantly lower the visual quality and the accuracy of subsequent applications. Moreover, the artificial light source tends to invalidate many image restoration algorithms. In this paper, an underwater image restoration (UIR) method using a novel Convolutional Neural Network (CNN) architecture and a synthesized underwater dataset is proposed. We discuss the reason for the over enhancement that exists in current UIR methods and revise the underwater image formation model (IFM) to alleviate the problem. With the revised IFM, we proposed an underwater image synthesizing method that can create a realistic underwater dataset. In order to effectively conduct end-to-end supervised learning, we design a network based on the characteristics of image restoration tasks, namely FMSNet. Different from existing networks, the decomposition and fusion operation in FMSNet can process the feature maps more efficiently and improve the contrast more prominently. The UIR method built by FMSNet can directly recover the degraded underwater images without the need of any pre-processing and post-processing. The experimental results indicate that FMSNet performs favorably against the widely used network architectures and our UIR method can outperform the state-of-the-art methods on both qualitative and quantitative evaluations. Comparing with the original underwater images, the experiments carried out by subsequent mission shows that 285% more feature points can be detected in the restored images by using our method.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Lu, H., et al.: Underwater image enhancement method using weighted guided trigonometric filtering and artificial light correction. J. Vis. Commun. Image Represent. 38, 504–516 (2016). https://doi.org/10.1016/j.jvcir.2016.03.029
Chiang, J.Y., Chen, Y.-C.: Underwater image enhancement by wavelength compensation and dehazing. IEEE Trans. Image Process. 21(4), 1756–1769 (2012). https://doi.org/10.1109/TIP.2011.2179666
Schechner, Y.Y., Karpel, N.: Recovery of underwater visibility and structure by polarization analysis. IEEE J. Oceanic Eng. 30, 570–587 (2005). https://doi.org/10.1109/JOE.2005.850871
Song, W., Wang, Y., Huang, D., Tjondronegoro, D.: A rapid scene depth estimation model based on underwater light attenuation prior for underwater image restoration. In: Hong, R., Cheng, W.-H., Yamasaki, T., Wang, M., Ngo, C.-W. (eds.) PCM 2018. LNCS, vol. 11164, pp. 678–688. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-00776-8_62
Drews, Jr., P., Nascimento, E., Moraes, F., Botelho, S., Campos, M.: Transmission estimation in underwater single images. In: 2013 IEEE International Conference on Computer Vision Workshops, pp. 825–830 (2013)
He, K., Sun, J., Tang, X.: Single image haze removal using dark channel prior. IEEE Trans. Pattern Anal. Mach. Intell. 33, 2341–2353 (2011). https://doi.org/10.1109/TPAMI.2010.168
Li, C.-Y., Guo, J.-C., Cong, R.-M., Pang, Y.-W., Wang, B.: Underwater image enhancement by dehazing with minimum information loss and histogram distribution prior. IEEE Trans. on Image Process. 25, 5664–5677 (2016). https://doi.org/10.1109/TIP.2016.2612882
Fabbri, C., Islam, M.J., Sattar, J.: Enhancing underwater imagery using generative adversarial networks. In: 2018 IEEE International Conference on Robotics and Automation (ICRA), pp. 7159–7165. IEEE, Brisbane (2018)
Zhu, J.-Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp. 2242–2251 (2017)
Islam, M.J., Xia, Y., Sattar, J.: Fast underwater image enhancement for improved visual perception. IEEE Robot. Autom. Lett. 5, 3227–3234 (2020). https://doi.org/10.1109/LRA.2020.2974710
Li, C., Anwar, S., Porikli, F.: Underwater scene prior inspired deep underwater image and video enhancement. Pattern Recogn. 98, 107038 (2020). https://doi.org/10.1016/j.patcog.2019.107038
Wang, K., Hu, Y., Chen, J., Wu, X., Zhao, X., Li, Y.: Underwater image restoration based on a parallel convolutional neural network. Remote Sens. 11, 1591 (2019). https://doi.org/10.3390/rs11131591
Li, C., et al.: An underwater image enhancement benchmark dataset and beyond. arXiv:1901.05495 (2019)
Johnson, J., Alahi, A., Fei-Fei, L.: Perceptual losses for real-time style transfer and super-resolution. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9906, pp. 694–711. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46475-6_43
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770–778 (2016)
Zhang, Y., Tian, Y., Kong, Y., Zhong, B., Fu, Y.: Residual dense network for image super-resolution. arXiv:1802.08797 (2018)
Li, R., Cheong, L.-F., Tan, R.T.: Heavy rain image restoration: integrating physics model and conditional adversarial learning. In: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1633–1642. IEEE, Long Beach (2019)
Liu, X., Ma, Y., Shi, Z., Chen, J.: GridDehazeNet: attention-based multi-scale network for image dehazing. In: 2019 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 7313–7322. IEEE, Seoul (2019)
Zhao, H., Gallo, O., Frosio, I., Kautz, J.: Loss functions for neural networks for image processing. arXiv:1511.08861 (2018)
Li, B., et al.: Benchmarking single-image dehazing and beyond. IEEE Trans. Image Process. 28, 492–505 (2019). https://doi.org/10.1109/TIP.2018.2867951
Dai, C., Lin, M., Wang, Z., Zhang, D., Guan, Z.: Color compensation based on bright channel and fusion for underwater image enhancement. Acta Optica Sinica. 38, 1110003 (2018)
Acknowledgements
The work is supported by the study abroad program for graduate student of Guilin University of Electronic Technology.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Yin, X., Liu, X., Liu, H. (2021). FMSNet: Underwater Image Restoration by Learning from a Synthesized Dataset. In: Farkaš, I., Masulli, P., Otte, S., Wermter, S. (eds) Artificial Neural Networks and Machine Learning – ICANN 2021. ICANN 2021. Lecture Notes in Computer Science(), vol 12893. Springer, Cham. https://doi.org/10.1007/978-3-030-86365-4_34
Download citation
DOI: https://doi.org/10.1007/978-3-030-86365-4_34
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-86364-7
Online ISBN: 978-3-030-86365-4
eBook Packages: Computer ScienceComputer Science (R0)