Skip to main content
Log in

Region of interest weighted pooling strategy for video quality metric

  • Published:
Telecommunication Systems Aims and scope Submit manuscript

Abstract

The objective video quality metrics have been researched for years and many methods have been proposed. As a main feature of the Human Visual System (HVS), visual attention (or Region of Interest—ROI) will influence viewer’s subjective feeling since artifacts on a ROI is much more annoying than those appearing on an inconspicuous area. However, little study has been taken on identifying how and to what extent ROI will influence video quality measurements. In this paper, we propose a fully automatic region of interest weighted pooling strategy considering the influence of visual attention, which is then evaluated on VQEG Phase I FR-TV test dataset. Apparent and coherent performance improvement is achieved by applying the proposed pooling strategy on PSNR and SSIM, together with a highly reduction in computation complexity.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. The Video Quality Experts Group web site. http://www.its.bldrdoc.gov/vqeg/.

  2. International Telecommunication Union (1997–2000). Final report from the Video Quality Experts Group on the validation of objective models of video quality assessment.

  3. Watson, A. B., Hu, J., & McGowan, J. F. III (2001). Digital video quality metric based on human vision. Journal of Electronic Imaging, 10(1), 20–29.

    Article  Google Scholar 

  4. Sarnoff corporation, JNDMetrix Technology (2003). http://www.sarnoff.com/productsservices/videovision/jndmetrix/download.asp.

  5. Winkler, S. (1999). Perceptual distortion metric for digital color video. In Proceedings of SPIE, the international society for optical engineering: Vol. 3644 (pp. 175–184). Bellingham: SPIE.

    Google Scholar 

  6. Koumaras, H., Lin, C. H., Shieh, C. K., & Kourtis, A. (2009). A framework for end-to-end video quality prediction of MPEG video. Journal of Visual Communication and Image Representation, 21(2), 139–154.

    Article  Google Scholar 

  7. Koumaras, H., Kourtis, A., Martakos, D., & Lauterjung, J. (2007). Quantified PQoS assessment based on fast estimation of the spatial and temporal activity level. Multimedia Tools and Applications, 34, 355–374.

    Article  Google Scholar 

  8. Wang, Z., & Bovik, A. C. (2002). A universal image quality index. IEEE Signal Processing Letters, 9, 81–84.

    Article  Google Scholar 

  9. Wang, Z., Lu, L., & Bovik, A. C. (2004). Video quality assessment based on structural distortion measurement. Signal Processing. Image Communication, 9(2), 121–132.

    Article  Google Scholar 

  10. Osberger, W., Bergmann, N., & Maeder, A. (1998). An automatic image quality assessment technique incorporating high level perceptual factors. In Proceedings of IEEE international conference on image processing (pp. 414–418).

  11. Engelke, U., Nguyen, V. X., & Zepernick, H. J. (2008). Regional attention to structural degradations for perceptual image quality metric design. In IEEE international conference on acoustics, speech, and signal processing (ICASSP) (pp. 869–872).

  12. Ninassi, A., Le Meur, O., Le Callet, P., & Barba, D. (2007). Does where you gaze on an image affect your perception of quality? Applying visual attention to image quality metric. In IEEE international conference on image processing (ICIP): Vol. 2 (pp. 169–172).

  13. Lu, Z. K., Lin, W. S., Yang, X. K., Ong, E. P., & Yao, S. S., (2005). Modeling visual attention’s modulatory aftereffects on visual sensitivity and quality evaluation. IEEE Transactions on Image Processing, 14(11), 1928–1942.

    Article  Google Scholar 

  14. Itti, L., Koch, C., & Niebur, E. (1998). A model of saliency-based visual attention for rapid scene analysis. IEEE Transactions On Pattern Analysis and Machine Intelligence, 20(11), 1254–1259.

    Article  Google Scholar 

  15. Bruce, N. D. B. (2005). Features that draw visual attention: an information theoretic perspective. Neurocomputing, 65–66, 125–133.

    Article  Google Scholar 

  16. Kadir, T., & Brady, M. (2001). Scale, saliency and image description. International Journal of Computer Vision, 45(2), 83–105.

    Article  Google Scholar 

  17. Ma, Y. F., Lu, L., Li, M. J., & Zhang, H. J. (2003). A user attention model for video summarization. In Proceedings of ACM multimedia.

  18. Zhai, Y., & Shah, M. (2006). Visual attention detection in video sequences using spatiotemporal cues. In Proceedings of ACM multimedia.

  19. Qiu, G. P., Gu, X. D. et al. (2007). An information theoretic model of spatiotemporal visual saliency. In IEEE international conference on multimedia and expo (ICME).

  20. Hu, C. C., Wu, J. L., & Cheng, W. H. (2005). A practical foveation-based rate-shaping mechanism for MPEG videos. IEEE Transactions on Circuits and Systems for Video Technology, 15, 1365–1372.

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Xiaodong Gu.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Gu, X., Qiu, G., Feng, X. et al. Region of interest weighted pooling strategy for video quality metric. Telecommun Syst 49, 63–73 (2012). https://doi.org/10.1007/s11235-010-9353-8

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11235-010-9353-8

Keywords

Navigation