Skip to main content
Log in

Stereoscopic video quality assessment based on visual attention and just-noticeable difference models

  • Original Paper
  • Published:
Signal, Image and Video Processing Aims and scope Submit manuscript

Abstract

With the consideration that incorporating visual saliency information appropriately can benefit image quality assessment metrics, this paper proposes an objective stereoscopic video quality assessment (SVQA) metric by incorporating stereoscopic visual attention (SVA) to SVQA metric. Specifically, based upon the multiple visual masking characteristics of HVS, a stereoscopic just-noticeable difference model is proposed to compute the perceptual visibility for stereoscopic video. Next, a novel SVA model is proposed to extract stereoscopic visual saliency information. Then, the quality maps are calculated by the similarity of the original and distorted stereoscopic videos’ perceptual visibility. Finally, the quality score is obtained by incorporating visual saliency information to the pooling of quality maps. To evaluate the proposed SVQA metric, a subjective experiment is conducted. The experimental result shows that the proposed SVQA metric achieves better performance in comparison with the existing SVQA metrics.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

References

  1. Ha, K., Kim, M.: A perceptual quality assessment metric using temporal complexity and disparity information for stereoscopic video. In: Proceedings of the ICIP, pp. 2525–2528 (2011)

  2. Bensalma, R., Larabi, M.C.: A perceptual metric for stereoscopic image quality assessment based on the binocular energy. Multidimens. Syst. Signal Process. 24(2), 281–316 (2013)

    Article  MathSciNet  MATH  Google Scholar 

  3. Shao, F., Lin, W., Gu, S., Jiang, G., Srikanthan, T.: Perceptual full-reference quality assessment of stereoscopic images by considering binocular visual characteristics. IEEE Trans. Image Process. 22(5), 1940–1953 (2013)

    Article  MathSciNet  Google Scholar 

  4. Joveluro, P., Malekmohamadi, H., Fernando, W.A.C., Kondoz, A.M.: Perceptual video quality metric for 3d video quality assessment. In: Proceedings of the 3DTV-CON, pp. 1–4 (2010)

  5. Jin, L., Boev, A., Gotchev, A., Egiazarian, K.: 3D-DCT based perceptual quality assessment of stereo vide. In: Proceedings of the ICIP, pp. 2521–2524 (2011)

  6. Lu, F., Wang, H., Ji, X., Er, G.: Quality assessment of 3D asymmetric view coding using spatial frequency dominance model. In: Proceedings of the 3DTV-CON, pp. 1–4 (2009)

  7. Han, J., Jiang, T., Ma, S.: Stereoscopic video quality assessment model based on spatial–temporal structural information. In: Proceedings of the VCIP, pp. 119–125 (2012)

  8. Chou, C.-H., Li, Y.-C.: A perceptually tuned sub-band image coder based on the measure of just-noticeable-distortion profile. IEEE Trans. Circuits Syst. Video Technol. 5(6), 467–476 (1995)

    Article  Google Scholar 

  9. Yang, X., Lin, W., Lu, Z., Ong, E.P., Yao, S.: Motion-compensated residue preprocessing in video coding based on just-noticeable-distortion profile. IEEE Trans. Circuits Syst. Video Technol. 15(6), 742–752 (2005)

    Article  Google Scholar 

  10. Zhang, X., Lin, W., Xue, P.: Just-noticeable difference estimation with pixels in images. J. Vis. Commun. Image Represent. 19, 30–41 (2008)

    Article  Google Scholar 

  11. Qi, F., Jiang, T., Fan, X., Ma, S., Zhao, D.: Stereoscopic video quality assessment based on stereo just-noticeable difference model. In: Proceedings of the ICIP, pp. 34–38 (2013)

  12. Lin, W., Jay Kuo, C.-C.: Perceptual visual quality metrics: a survey. J. Vis. Commun. Image Represent. 22(4), 297–312 (2011)

    Article  Google Scholar 

  13. Zhao, Y., Chen, Z., Zhu, C., Tan, Y., Yu, L.: Binocular just-noticeable-difference model for stereoscopic images. IEEE Signal Process. Lett. 18(1), 19–22 (2011)

    Article  Google Scholar 

  14. De. Silva, D., Fernando, W.A.C., Worrall, S.T., Yasakethu, S.L.P., Kondoz, A.M.: Just noticeable difference in depth model for stereoscopic 3D displays. In: Proceedings of the ICME, pp. 1219–1224 (2010)

  15. Li, X., Wang, Y., Zhao, D., Jiang, T., Zhang, N.: Joint just noticeable difference model based on depth perception for stereoscopic images. In: Proceedings of the VCIP, pp. 1–4 (2011)

  16. Zhai, G., Wu, X., Yang, X., Lin, W., Zhang, W.: A psychovisual quality metric in free-energy principle. IEEE Trans. Image Process. 21(1), 41–52 (2012)

    Article  MathSciNet  Google Scholar 

  17. Friston, K.: The free-energy principle: a unified brain theory? Nat. Rev. Neurosci. 11(2), 127–138 (2010)

    Article  Google Scholar 

  18. Howard, I.P., Rogers, B.J.: Binocular Vision and Stereopsis. Oxford University Press, Oxford (1995)

    Google Scholar 

  19. Wang, Z., Li, Q.: Information content weighting for perceptual image quality assessment. IEEE Trans. Image Process. 20(5), 1185–1198 (2011)

    Article  MathSciNet  Google Scholar 

  20. Zhang, Y., Jiang, G., Yu, M., Chen, K.: Stereoscopic visual attention model for 3-D video. In: Proceedings of the Multimedia Modeling, pp. 314–324 (2010)

  21. Dittrich, T., Kopf, S., Schaber, P., Guthier, B., Effelsberg, W.: Saliency detection for stereoscopic video. In: Proceedings of the 4th ACM Conference on Multimedia Systems, pp. 12–23 (2013)

  22. Wang, J., Perreira, M., Silva, D., Callet, P.L., Ricordel, V.: A computational model of stereoscopic 3D visual saliency. IEEE Trans. Image Process. 22(6), 2151–2165 (2013)

    Article  MathSciNet  Google Scholar 

  23. Zhang, L., Shen, Y., Li, H.Y.: VSI: a visual saliency induced index for perceptual image quality assessment. IEEE Trans. Image Process. 23(10), 4270–4281 (2014)

    Article  MathSciNet  Google Scholar 

  24. Aflaki, P., Hannuksela, M.M., Hakkinen, J., Lindroos, P., Gabbouj, M.: Subjective study on compressed asymmetric stereoscopic video. In: Proceedings of the ICIP, pp. 4021–4024 (2010)

  25. Fleet, D.J., Wagner, H., Heeger, D.J.: Neural encoding of binocular disparity: energy models, position shifts and phase shifts. J. Vis. Res. 36(12), 1839–1857 (1996)

    Article  Google Scholar 

  26. Cheng, M.M., Zhang, G.X., Mitra, N.J., Huang, X., Hu, S.M.: Global contrast based salient region detection. In: Proceedings of the CVPR, pp. 409–416 (2011)

  27. May, K.A., Li, Z.P., Hibbard, P.B.: Perceived direction of motion determined by adaptation to static binocular images. Curr. Biol. 22(1), 28–32 (2012)

    Article  Google Scholar 

  28. Seo, H.J., Milanfar, P.: Visual saliency for automatic target detection, boundary detection, and image quality assessment. In: Proceedings of the ICASSP (2010)

  29. Zhong, S.H., Liu, Y., Ren, F.F., Zhang, J.H., Ren, T.W.: Video saliency detection via dynamic consistent spatio-temporal attention modelling. In: Proceedings of the AAAI Conference on Artificial Intelligence (2013)

  30. Urvoy, M., Gutirrez, J., Barkowsky, M., Cousseau, R., Koudota, Y., Ricordel, V., Callet, P.L., Garca, N.: NAMA3DS1-COSPAD1: subjective video quality assessment database on coding conditions introducing freely available high quality 3D stereoscopic sequences. In: Fourth International Workshop on Quality of Multimedia Experience (2012)

  31. Qi, F.: The Illumination of SVQA Subjective Test. http://www.escience.cn/people/qifeng/index.html

Download references

Acknowledgments

This work has been supported in part by the Major State Basic Research Development Program of China (973 Program 2015CB351804), the National Science Foundation of China under Grant (Nos. 61472101, 61272386).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Xiaopeng Fan.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Qi, F., Zhao, D., Fan, X. et al. Stereoscopic video quality assessment based on visual attention and just-noticeable difference models. SIViP 10, 737–744 (2016). https://doi.org/10.1007/s11760-015-0802-4

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11760-015-0802-4

Keywords

Navigation