Skip to main content
Log in

YUVDR: A residual network for image deblurring in YUV color space

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

Motion blur removal caused by camera shake and object motion in 3D space has long been a challenge in computer vision. Although RGB images are commonly used as input data for CNN-based image deblurring, their inherent issues of color overlap and high dimensionality can limit performance. To address these problems, we propose YUVDR, a residual network based on YUV color space, for image deblurring. By using YUV images, we mitigate the issues of color overlap and mutual influence. We introduce novel loss functions and conduct experiments on three datasets, namely GoPro, DVD and NFS, which offer a wide range of image quality levels, scene complexities, and types of motion blur. Our proposed method outperforms state-of-the-art algorithms, yielding a 3-5 dB improvement in the PSNR of test results. In addition, utilizing the YUV color space as the input data can greatly reduce the number of training parameters and model size, by approximately 15 times. This optimization of GPU memory not only improves training efficiency, but also reduces testing time for practical applications.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10

Similar content being viewed by others

Data availability statement

The datasets generated during and/or analysed during the current study are available, visit https://github.com/VITA-Group/DeblurGANv2#datasets. The links to download each dataset are provided below: GoPro https://drive.google.com/file/d/1KStHiZn5TNm2mo3OLZLjnRvd0vVFCI0W/view, DVD https://drive.google.com/file/d/1bpj9pCcZR_6-AHb5aNnev5lILQbH8GMZ/view, NFS https://drive.google.com/file/d/1Ut7qbQOrsTZCUJA_mJLptRMipD8sJzjy/view

References

  1. Amirkhani D, Bastanfard A (2019) Inpainted image quality evaluation based on saliency map features. In: 2019 5th Iranian Conference on Signal Processing and Intelligent Systems (ICSPIS), pp. 1–6. IEEE

  2. Ansari M, Singh DK, et al. (2022) Significance of color spaces and their selection for image processing: a survey. Recent Advances in Computer Science and Communications (Formerly: Recent Patents on Computer Science) 15(7):946–956

  3. Bahat Y, Efrat N, Irani M (2017) Non-uniform blind deblurring by reblurring. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 3286–3294

  4. Bastanfard A, Amirkhani D, Mohammadi M (2022) Toward image super-resolution based on local regression and nonlocal means. Multimed Tools Appl 81(16):23473–23492

    Article  Google Scholar 

  5. Bengio Y, Simard P, Frasconi P (1994) Learning long-term dependencies with gradient descent is difficult. IEEE transactions on neural networks 5(2):157–166

    Article  CAS  PubMed  Google Scholar 

  6. Bloomfield P, Steiger WL (1983) Least Absolute Deviations: Theory, Applications, and Algorithms. Springer, ???

  7. Bradley RA, Terry ME (1952) Rank analysis of incomplete block designs: I. the method of paired comparisons. Biometrika 39(3/4):324–345

  8. Cai C, Meng H, Zhu Q (2018) Blind deconvolution for image deblurring based on edge enhancement and noise suppression. IEEE Access 6:58710–58718

    Article  Google Scholar 

  9. Ghimire D, Kil D, Kim S-h (2022) A survey on efficient convolutional neural networks and hardware acceleration. Electronics 11(6):945

    Article  Google Scholar 

  10. Han K, Wang D (2014) Neural network based pitch tracking in very noisy speech. IEEE/ACM Trans Audio, Speech, Lang Process 22(12):2158–2168

    Article  Google Scholar 

  11. Hao R, Wang X, Du X, Zhang J, Liu J, Liu L (2022) End-to-end deep learning-based cells detection in microscopic leucorrhea images. Microsc Microanal 28(3):732–743

    Article  ADS  CAS  Google Scholar 

  12. He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778

  13. He K, Zhang X, Ren S, Sun J (2016) Identity mappings in deep residual networks. In: Computer Vision-ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11–14, 2016, Proceedings, Part IV 14, pp. 630–645. Springer

  14. Honghui Y, Junhao L, Meiping S (2022) Underwater acoustic target multi-attribute correlation perception method based on deep learning. Appl Acoustics 190:108644

    Article  Google Scholar 

  15. Hradiš M, Kotera J, Zemcı P, Šroubek F (2015) Convolutional neural networks for direct text deblurring. In: Proceedings of BMVC, vol. 10

  16. Instruments N (2013) Peak signal-to-noise ratio as an image quality metric

  17. Jiang W, Liu A (2022) Image motion deblurring based on deep residual shrinkage and generative adversarial networks. Computat Intell Neurosci 2022:1–15

    Google Scholar 

  18. Kiani Galoogahi H, Fagg A, Huang C, Ramanan D, Lucey S (2017) Need for speed: A benchmark for higher frame rate object tracking. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1125–1134

  19. Kupyn O, Budzan V, Mykhailych M, Mishkin D, Matas J (2018) Deblurgan: Blind motion deblurring using conditional adversarial networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8183–8192

  20. Kupyn O, Martyniuk T, Wu J, Wang Z (2019) Deblurgan-v2: Deblurring (orders-of-magnitude) faster and better. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 8878–8887

  21. Lai W-S, Huang J-B, Hu Z, Ahuja N, Yang M-H (2016) A comparative study for single image blind deblurring. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1701–1709

  22. LeCun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521(7553):436–444

    Article  ADS  CAS  PubMed  Google Scholar 

  23. Le N, Rathour VS, Yamazaki K, Luu K, Savvides M (2022) Deep reinforcement learning in computer vision: a comprehensive survey. Artif Intell Rev 1–87

  24. Li M, Yang J, Su Z-y (2010) Support vector regression based color image restoration in yuv color space. J Shanghai Jiaotong University (Science) 15:31–35

    Article  ADS  Google Scholar 

  25. Li B, Ren W, Fu D, Tao D, Feng D, Zeng W, Wang Z (2018) Benchmarking single-image dehazing and beyond. IEEE Trans Image Process 28(1):492–505

    Article  ADS  MathSciNet  Google Scholar 

  26. Li S, Araujo IB, Ren W, Wang Z, Tokuda EK, Junior RH, Cesar-Junior R, Zhang J, Guo X, Cao X (2019) Single image deraining: A comprehensive benchmark analysis. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3838–3847

  27. Li S, Guo H, Sun W, Sun X (2022) A low-illuminance image enhancement method in yuv color space. In: 2022 14th International Conference on Measuring Technology and Mechatronics Automation (ICMTMA), pp. 286–291.IEEE

  28. Li H, Lin Z, Shen X, Brandt J, Hua G (2015) A convolutional neural network cascade for face detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5325–5334

  29. Liu RW, Shi L, Yu SC, Wang D (2015) Box-constrained second-order total generalized variation minimization with a combined l 1, 2 data-fidelity term for image reconstruction. J Electron Imag 24(3):033026–033026

    Article  ADS  Google Scholar 

  30. Minoofam SAH, Bastanfard A, Keyvanpour MR (2021) Trcla: a transfer learning approach to reduce negative transfer for cellular learning automata. IEEE Trans Neural Netw Learn Syst

  31. Nah S, Hyun Kim T, Mu Lee K (2017) Deep multi-scale convolutional neural network for dynamic scene deblurring. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3883–3891

  32. Pan J, Sun D, Pfister H, Yang M-H (2016) Blind image deblurring using dark channel prior. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1628–1636

  33. Park K-B, Lee JY (2022) Swine-net: hybrid deep learning approach to novel polyp segmentation using convolutional neural network and swin transformer. J Computat Design Eng 9(2):616–632

    Article  Google Scholar 

  34. Park S, Shin Y-G (2022) Generative residual block for image generation. Appl Intell 1–10

  35. Pergoloni S, Biagi M, Colonnese S, Cusani R, Scarano G (2016) Camera communication deblurring: A semiblind spatial fractionally-spaced adaptive equalizer with flexible filter support design. In: 2016 Sixth International Conference on Image Processing Theory, Tools and Applications (IPTA), pp. 1–6. IEEE

  36. Podpora M, Korbas GP, Kawala-Janik A (2014) Yuv vs rgb-choosing a color space for human-machine interaction. In: FedCSIS (Position Papers), pp. 29–34 Citeseer

  37. Premachandra C, Ueda S, Suzuki Y (2019) Road intersection moving object detection by 360-degree view camera. In: 2019 IEEE 16th International Conference on Networking, Sensing and Control (ICNSC), pp. 369–372. IEEE

  38. Qu Z, Wang, J (2010) A color yuv image edge detection method based on histogram equalization transformation. In: 2010 Sixth International Conference on Natural Computation, 7:3546–3549. IEEE

  39. Satish P, Srikantaswamy M, Ramaswamy NK (2020) A comprehensive review of blind deconvolution techniques for image deblurring. Traitement du Signal 37(3)

  40. Schmidt M (2005) Least squares optimization with l1-norm regularization. CS542B Project Report 504:195–221

  41. Sehar U, Naseem ML (2022) How deep learning is empowering semantic segmentation: Traditional and deep learning techniques for semantic segmentation: A comparison. Multimed Tools Appl 81(21):30519–30544

    Article  PubMed  PubMed Central  Google Scholar 

  42. Su J, Xu B, Yin H (2022) A survey of deep learning approaches to image restoration. Neurocomputing 487:P46-65

    Article  Google Scholar 

  43. Subashini P, Krishnaveni M, Singh V (2011) Image deblurring using back propagation neural network. World Comp Sci Inf Technol J 1(6):277–282

    Google Scholar 

  44. Su S, Delbracio M, Wang J, Sapiro G, Heidrich W, Wang O (2017) Deep video deblurring for hand-held cameras. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1279–1288

  45. Taigman Y, Yang M, Ranzato M, Wolf L (2014) Deepface: Closing the gap to human-level performance in face verification. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1701–1708

  46. Tao X, Gao H, Shen X, Wang J, Jia J (2018) Scale-recurrent network for deep image deblurring. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8174–8182

  47. Wang Z, Bovik AC, Sheikh HR, Simoncelli EP (2004) Image quality assessment: from error visibility to structural similarity. IEEE Trans Image Process 13(4):600–612

    Article  ADS  PubMed  Google Scholar 

  48. Wang K-J, Rizqi DA, Nguyen H-P (2021) Skill transfer support model based on deep learning. J Intell Manufac 32:1129–1146

    Article  Google Scholar 

  49. Wang Z, Simoncelli EP, Bovik AC (2003) Multiscale structural similarity for image quality assessment. In: The Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, 2003, 2:1398–1402. Ieee

  50. Wu Y, Ling H, Yu J, Li F, Mei X, Cheng, E (2011) Blurred target tracking by blur-driven tracker. In: 2011 International Conference on Computer Vision, 1100–1107. IEEE

  51. Zhang K, Luo W, Zhong Y, Ma L, Stenger B, Liu W, Li H (2020) Deblurring by realistic blurring. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 2737–2746

  52. Zhao H, Gallo O, Frosio I, Kautz J (2015) Loss functions for neural networks for image processing. arXiv:1511.08861

  53. Zhao N, Wei Q, Basarab A, Kouamé D, Tourneret J-Y (2016) Blind deconvolution of medical ultrasound images using a parametric model for the point spread function. In: 2016 IEEE International Ultrasonics Symposium (IUS), pp. 1–4. IEEE

Download references

Acknowledgements

We would like to express our sincere gratitude to the anonymous reviewers for their insightful and constructive comments, which helped to improve the quality of this article. We would also like to thank the editor for their guidance and valuable feedback throughout the submission process

Funding

This work was supported by National Natural Science Foundation of China under Grant 61301250, China Scholarship Council under Grant [2020]1417, Natural Science Foundation for Young Scientists of Shanxi Province under Grant 201901D211313, Shanxi Scholarship Council of China under Grant HGKY2019080 and 2020-127, Open project of Guangdong Provincial Key Laboratory of Digital Signal and Image Processing in 2021, Shanxi Province Postgraduate Excellent Innovation Project Plan under Grant 2021Y679

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yina Guo.

Ethics declarations

Conflict of interest

All authors certify that they have no affiliations with or involvement in any organization or entity with any financial interest or non-financial interest in the subject matter or materials discussed in this manuscript

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zhang, M., Wang, H. & Guo, Y. YUVDR: A residual network for image deblurring in YUV color space. Multimed Tools Appl 83, 19541–19561 (2024). https://doi.org/10.1007/s11042-023-16284-y

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-023-16284-y

Keywords

Navigation