Abstract
Transformer image quality is influenced by various factors, including acquisition equipment and external environmental conditions. There is an urgent need for a quality evaluation method to assess transformer image quality, particularly to enhance the accuracy of transformer oil leakage identification tasks. Addressing this issue, this paper introduces image quality evaluation to the field of transformers for the first time and proposes a novel model called Multi-Directional Feature Extraction Transformer Image Quality Assessment (MFE-TIQA) for automatic evaluation of large-scale transformer images. MFE-TIQA comprises a main branch and a sub-branch. The main branch utilizes convolutional neural networks to extract multi-scale features from transformer images and employs a multi-directional feature extraction module to capture fine details. Meanwhile, the sub-branch employs a super-pixel segmentation model to generate local visual information from transformer images. Subsequently, a multi-branch information fusion module is constructed to comprehensively integrate the information from both branches, enhancing the model’s focus on critical details. Furthermore, comparative experiments are conducted on a self-constructed transformer image dataset and a publicly available dataset to validate the effectiveness of the proposed model. This research provides foundational evidence for subsequent transformer oil leakage identification and detection efforts.
This work was supported in part by the National Natural Science Foundation of China under Grant 62371188.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Dai, Y., Gao, Z., Zhu, Y., Zhang, W., Li, H., Wang, Y., Li, Z.: Feature grouping for no-reference image quality assessment. In: 7th International Conference on Automation, Control and Robotics Engineering, pp. 204–208 (2022)
Ghadiyaram, D., Bovik, A.C.: Massive online crowdsourced study of subjective and objective picture quality. IEEE Trans. Image Process. 25(1), 372–387 (2016)
Golestaneh, S.A., Dadsetan, S., Kitani, K.M.: No-reference image quality assessment via transformers, relative ranking, and self-consistency. In: IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 3989–3999 (2022)
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778. IEEE Computer Society (2016)
Larson, E.C., Chandler, D.M.: Most apparent distortion: full-reference image quality assessment and the role of strategy. J. Electron. Imaging 19(1), 1–21 (2010)
Liu, S., Li, C., Nan, N., Zong, Z., Song, R.: MMDM: multi-frame and multi-scale for image demoiréing. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 1751–1759 (2020)
Mittal, A., Moorthy, A.K., Bovik, A.C.: No-reference image quality assessment in the spatial domain. IEEE Trans. Image Process. 21(12), 4695–4708 (2012)
Moorthy, A.K., Bovik, A.C.: Blind image quality assessment: from natural scene statistics to perceptual quality. IEEE Trans. Image Process. 20(12), 3350–3364 (2011)
Nafchi, H.Z., Shahkolaei, A., Hedjam, R., Cheriet, M.: Mean deviation similarity index: efficient and reliable full-reference image quality evaluator. IEEE Access 4, 5579–5590 (2016)
Pan, Q., Guo, N., Qingge, L., Zhang, J., Yang, P.: PMT-IQA: progressive multi-task learning for blind image quality assessment. In: PRICAI 2023: Trends in Artificial Intelligence - 20th Pacific Rim International Conference on Artificial Intelligence. Lecture Notes in Computer Science, vol. 14327, pp. 153–164 (2023)
Pan, Z., Zhang, H., Lei, J., Fang, Y., Shao, X., Ling, N., Kwong, S.: DACNN: blind image quality assessment via a distortion-aware convolutional neural network. IEEE Trans. Circuits Syst. Video Technol. 32(11), 7518–7531 (2022)
Ponomarenko, N.N., Ieremeiev, O., Lukin, V.V., Egiazarian, K.O., Jin, L., Astola, J., Vozel, B., Chehdi, K., Carli, M., Battisti, F., Kuo, C.J.: Color image database TID2013: peculiarities and preliminary results. In: European Workshop on Visual Information Processing, pp. 106–111. IEEE (2013)
Qu, C., Zhang, Y., Ma, F., Huang, K.: Parameter optimization for point clouds denoising based on no-reference quality assessment. Measurement 211, 112–592 (2023)
Su, S., Yan, Q., Zhu, Y., Zhang, C., Ge, X., Sun, J., Zhang, Y.: Blindly assess image quality in the wild guided by a self-adaptive hyper network. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3664–3673 (2020)
Wang, X., Xiong, J., Li, B., Suo, J., Gao, H.: Learning hybrid representations of semantics and distortion for blind image quality assessment. In: IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 1–5. IEEE (2023)
Xue, W., Zhang, L., Mou, X., Bovik, A.C.: Gradient magnitude similarity deviation: a highly efficient perceptual image quality index. IEEE Trans. Image Process. 23(2), 684–695 (2013)
Yan, J., Fang, Y., Liu, X.: The review of distortion-related image quality assessment. J. Image Graph. 27(5), 1430–1466 (2022)
Yang, C., Yang, Y.: A deep neural network based on layer-by-layer fusion of multi-scale features for no-reference image quality assessment. J. South China Univ. Technol. (Natural Science Edition) 50(4), 81–89+141 (2022)
Yang, F., Sun, Q., Jin, H., Zhou, Z.: Superpixel segmentation with fully convolutional networks. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 13961–13970 (2020)
Yang, S., Jiang, Q., Lin, W., Wang, Y.: Sgdnet: An end-to-end saliency-guided deep neural network for no-reference image quality assessment. In: Proceedings of the 27th ACM International Conference on Multimedia, pp. 1383–1391 (2019)
Ye, P., Kumar, J., Kang, L., Doermann, D.S.: Unsupervised feature learning framework for no-reference image quality assessment. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, June 16–21, 2012, pp. 1098–1105 (2012)
Zhang, W., Ma, K., Yan, J., Deng, D., Wang, Z.: Blind image quality assessment using a deep bilinear convolutional neural network. IEEE Trans. Circuits Syst. Video Technol. 30(1), 36–47 (2020)
Zhang, W., Ma, K., Zhai, G., Yang, X.: Task-specific normalization for continual learning of blind image quality models. IEEE Trans. Image Process. 33, 1898–1910 (2024)
Zhao, W., Xu, L.: Weakly supervised target detection based on spatial attention. Vis. Intell. 2(1) (2024)
Zhao, W., Liu, L., Hu, J., Zhai, Y., Zhao, Z.: Detection of transformer oil leakage based on deep separable atrousconvolution pyramid. CAAI Trans. Intell. Syst. 1–9 (2023)
Zhou, M., Lan, X., Wei, X., Liao, X., Mao, Q., Li, Y., Wu, C., Xiang, T., Fang, B.: An end-to-end blind image quality assessment method using a recurrent network and self-attention. IEEE Trans. Broadcast. 69(2), 369–377 (2023)
Zhou, Z., Zhou, Z., Tao, X., Chen, H., Yu, Z., Cao, Y.: Earnet: error-aware reconstruction network for no-reference image quality assessment. Expert Syst. Appl. 238(Part C), 122050 (2024)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2025 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Zhao, W., Li, M., Ma, Y. (2025). Transformer Image Quality Assessment Based on Multi-directional Feature Extraction. In: Lin, Z., et al. Pattern Recognition and Computer Vision. PRCV 2024. Lecture Notes in Computer Science, vol 15034. Springer, Singapore. https://doi.org/10.1007/978-981-97-8505-6_25
Download citation
DOI: https://doi.org/10.1007/978-981-97-8505-6_25
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-97-8504-9
Online ISBN: 978-981-97-8505-6
eBook Packages: Computer ScienceComputer Science (R0)