Skip to main content

Transformer Image Quality Assessment Based on Multi-directional Feature Extraction

  • Conference paper
  • First Online:
Pattern Recognition and Computer Vision (PRCV 2024)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 15034))

Included in the following conference series:

  • 149 Accesses

Abstract

Transformer image quality is influenced by various factors, including acquisition equipment and external environmental conditions. There is an urgent need for a quality evaluation method to assess transformer image quality, particularly to enhance the accuracy of transformer oil leakage identification tasks. Addressing this issue, this paper introduces image quality evaluation to the field of transformers for the first time and proposes a novel model called Multi-Directional Feature Extraction Transformer Image Quality Assessment (MFE-TIQA) for automatic evaluation of large-scale transformer images. MFE-TIQA comprises a main branch and a sub-branch. The main branch utilizes convolutional neural networks to extract multi-scale features from transformer images and employs a multi-directional feature extraction module to capture fine details. Meanwhile, the sub-branch employs a super-pixel segmentation model to generate local visual information from transformer images. Subsequently, a multi-branch information fusion module is constructed to comprehensively integrate the information from both branches, enhancing the model’s focus on critical details. Furthermore, comparative experiments are conducted on a self-constructed transformer image dataset and a publicly available dataset to validate the effectiveness of the proposed model. This research provides foundational evidence for subsequent transformer oil leakage identification and detection efforts.

This work was supported in part by the National Natural Science Foundation of China under Grant 62371188.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Dai, Y., Gao, Z., Zhu, Y., Zhang, W., Li, H., Wang, Y., Li, Z.: Feature grouping for no-reference image quality assessment. In: 7th International Conference on Automation, Control and Robotics Engineering, pp. 204–208 (2022)

    Google Scholar 

  2. Ghadiyaram, D., Bovik, A.C.: Massive online crowdsourced study of subjective and objective picture quality. IEEE Trans. Image Process. 25(1), 372–387 (2016)

    Article  MathSciNet  Google Scholar 

  3. Golestaneh, S.A., Dadsetan, S., Kitani, K.M.: No-reference image quality assessment via transformers, relative ranking, and self-consistency. In: IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 3989–3999 (2022)

    Google Scholar 

  4. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778. IEEE Computer Society (2016)

    Google Scholar 

  5. Larson, E.C., Chandler, D.M.: Most apparent distortion: full-reference image quality assessment and the role of strategy. J. Electron. Imaging 19(1), 1–21 (2010)

    Google Scholar 

  6. Liu, S., Li, C., Nan, N., Zong, Z., Song, R.: MMDM: multi-frame and multi-scale for image demoiréing. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 1751–1759 (2020)

    Google Scholar 

  7. Mittal, A., Moorthy, A.K., Bovik, A.C.: No-reference image quality assessment in the spatial domain. IEEE Trans. Image Process. 21(12), 4695–4708 (2012)

    Article  MathSciNet  Google Scholar 

  8. Moorthy, A.K., Bovik, A.C.: Blind image quality assessment: from natural scene statistics to perceptual quality. IEEE Trans. Image Process. 20(12), 3350–3364 (2011)

    Article  MathSciNet  Google Scholar 

  9. Nafchi, H.Z., Shahkolaei, A., Hedjam, R., Cheriet, M.: Mean deviation similarity index: efficient and reliable full-reference image quality evaluator. IEEE Access 4, 5579–5590 (2016)

    Article  Google Scholar 

  10. Pan, Q., Guo, N., Qingge, L., Zhang, J., Yang, P.: PMT-IQA: progressive multi-task learning for blind image quality assessment. In: PRICAI 2023: Trends in Artificial Intelligence - 20th Pacific Rim International Conference on Artificial Intelligence. Lecture Notes in Computer Science, vol. 14327, pp. 153–164 (2023)

    Google Scholar 

  11. Pan, Z., Zhang, H., Lei, J., Fang, Y., Shao, X., Ling, N., Kwong, S.: DACNN: blind image quality assessment via a distortion-aware convolutional neural network. IEEE Trans. Circuits Syst. Video Technol. 32(11), 7518–7531 (2022)

    Article  Google Scholar 

  12. Ponomarenko, N.N., Ieremeiev, O., Lukin, V.V., Egiazarian, K.O., Jin, L., Astola, J., Vozel, B., Chehdi, K., Carli, M., Battisti, F., Kuo, C.J.: Color image database TID2013: peculiarities and preliminary results. In: European Workshop on Visual Information Processing, pp. 106–111. IEEE (2013)

    Google Scholar 

  13. Qu, C., Zhang, Y., Ma, F., Huang, K.: Parameter optimization for point clouds denoising based on no-reference quality assessment. Measurement 211, 112–592 (2023)

    Article  Google Scholar 

  14. Su, S., Yan, Q., Zhu, Y., Zhang, C., Ge, X., Sun, J., Zhang, Y.: Blindly assess image quality in the wild guided by a self-adaptive hyper network. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3664–3673 (2020)

    Google Scholar 

  15. Wang, X., Xiong, J., Li, B., Suo, J., Gao, H.: Learning hybrid representations of semantics and distortion for blind image quality assessment. In: IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 1–5. IEEE (2023)

    Google Scholar 

  16. Xue, W., Zhang, L., Mou, X., Bovik, A.C.: Gradient magnitude similarity deviation: a highly efficient perceptual image quality index. IEEE Trans. Image Process. 23(2), 684–695 (2013)

    Article  MathSciNet  Google Scholar 

  17. Yan, J., Fang, Y., Liu, X.: The review of distortion-related image quality assessment. J. Image Graph. 27(5), 1430–1466 (2022)

    Article  Google Scholar 

  18. Yang, C., Yang, Y.: A deep neural network based on layer-by-layer fusion of multi-scale features for no-reference image quality assessment. J. South China Univ. Technol. (Natural Science Edition) 50(4), 81–89+141 (2022)

    Google Scholar 

  19. Yang, F., Sun, Q., Jin, H., Zhou, Z.: Superpixel segmentation with fully convolutional networks. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 13961–13970 (2020)

    Google Scholar 

  20. Yang, S., Jiang, Q., Lin, W., Wang, Y.: Sgdnet: An end-to-end saliency-guided deep neural network for no-reference image quality assessment. In: Proceedings of the 27th ACM International Conference on Multimedia, pp. 1383–1391 (2019)

    Google Scholar 

  21. Ye, P., Kumar, J., Kang, L., Doermann, D.S.: Unsupervised feature learning framework for no-reference image quality assessment. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, June 16–21, 2012, pp. 1098–1105 (2012)

    Google Scholar 

  22. Zhang, W., Ma, K., Yan, J., Deng, D., Wang, Z.: Blind image quality assessment using a deep bilinear convolutional neural network. IEEE Trans. Circuits Syst. Video Technol. 30(1), 36–47 (2020)

    Article  Google Scholar 

  23. Zhang, W., Ma, K., Zhai, G., Yang, X.: Task-specific normalization for continual learning of blind image quality models. IEEE Trans. Image Process. 33, 1898–1910 (2024)

    Article  Google Scholar 

  24. Zhao, W., Xu, L.: Weakly supervised target detection based on spatial attention. Vis. Intell. 2(1) (2024)

    Google Scholar 

  25. Zhao, W., Liu, L., Hu, J., Zhai, Y., Zhao, Z.: Detection of transformer oil leakage based on deep separable atrousconvolution pyramid. CAAI Trans. Intell. Syst. 1–9 (2023)

    Google Scholar 

  26. Zhou, M., Lan, X., Wei, X., Liao, X., Mao, Q., Li, Y., Wu, C., Xiang, T., Fang, B.: An end-to-end blind image quality assessment method using a recurrent network and self-attention. IEEE Trans. Broadcast. 69(2), 369–377 (2023)

    Article  Google Scholar 

  27. Zhou, Z., Zhou, Z., Tao, X., Chen, H., Yu, Z., Cao, Y.: Earnet: error-aware reconstruction network for no-reference image quality assessment. Expert Syst. Appl. 238(Part C), 122050 (2024)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Wenqing Zhao .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2025 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Zhao, W., Li, M., Ma, Y. (2025). Transformer Image Quality Assessment Based on Multi-directional Feature Extraction. In: Lin, Z., et al. Pattern Recognition and Computer Vision. PRCV 2024. Lecture Notes in Computer Science, vol 15034. Springer, Singapore. https://doi.org/10.1007/978-981-97-8505-6_25

Download citation

  • DOI: https://doi.org/10.1007/978-981-97-8505-6_25

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-97-8504-9

  • Online ISBN: 978-981-97-8505-6

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics