Abstract
Sencogi Quality Metric (SenQM) is a novel objective metric for video quality assessment. SenQM infers video quality scores from the spatio-temporal evolution of videos. The quality model behind SenQM is based on an algorithm developed by Cogisen for modelling dynamic phenomena generated by complex systems. In the field of video compression, Cogisen’s algorithm uses machine learning to model human perception of video quality by extracting meaningful information directly from the video data domain and its frequency representation. The model has been trained over datasets of (i) x264 compressed videos as input data and (ii) the corresponding subjective Mean Opinion Scores as ground truth. This study introduces the model behind SenQM and how the proposed metric performs in subjective video quality prediction compared to the most used video quality assessment methods, i.e. PSNR, SSIM, and Netflix’s VMAF. Results indicate a significantly higher prediction performance in terms of monotonicity, consistency, and accuracy than the compared metrics. SenQM quality scores show significantly higher variations for 352 × 288 resolution videos with equivalent levels of degradation, and outstands PSNR, SSIM, and VMAF in predicting subjective scores of increasing levels of compression without being affected by either the degradation level or the video content.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
LIVE Public-Domain Subjective Video Quality Database: https://live.ece.utexas.edu/research/Quality/live_video.html, last accessed on 27 January 2020.
- 2.
Netflix Video Quality Dataset: https://github.com/Netflix/vmaf, last accessed on 27 January 2020.
- 3.
Video Quality Group at Faculty of Electrical Engineering. URL: http://www.etfos.unios.hr/vqg/ Last access: 14 February 2020.
- 4.
Prolific Academic. URL: www.prolific.co. Last access: 14 February 2020.
References
Chikkerur, S., Sundaram, V., Reisslein, M., Karam, L.J.: Objective video quality assessment methods: a classification, review, and performance comparison (2011). http://dx.doi.org/10.1109/tbc.2011.2104671
International Telecommunication Union – ITU: Recommendation ITU-R BT.500-14 (10/2019). Methodologies for the subjective assessment of the quality of television images. BT Series. Broadcasting service (television), October 2019
Zhang, Y., Kwong, S., Wang, S.: Machine learning based video coding optimizations: a survey (2020). http://dx.doi.org/10.1016/j.ins.2019.07.096
Staelens, N., et al.: Assessing quality of experience of IPTV and video on demand services in real-life environments (2010). http://dx.doi.org/10.1109/tbc.2010.2067710
Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity (2004). http://dx.doi.org/10.1109/tip.2003.819861
Wang, Z., Bovik, A.C.: Modern image quality assessment (2006). http://dx.doi.org/10.2200/s00010ed1v01y200508ivm003
Li, Z., Aaron, A., Katsavounidis, I., Moorthy, A., Manohara, M.: Toward a practical perceptual video quality metric. Netflix Tech Blog. 6, 2 (2016)
Bosse, S., Maniry, D., Muller, K.-R., Wiegand, T., Samek, W.: Deep neural networks for no-reference and full-reference image quality assessment (2018). http://dx.doi.org/10.1109/tip.2017.2760518
Kang, L., Ye, P., Li, Y., Doermann, D.: Convolutional neural networks for no-reference image quality assessment (2014). http://dx.doi.org/10.1109/cvpr.2014.224
Zhang, Y., Gao, X., He, L., Lu, W., He, R.: Objective video quality assessment combining transfer learning with CNN (2019). http://dx.doi.org/10.1109/tnnls.2018.2890310
Ye, P., Kumar, J., Kang, L., Doermann, D.: Unsupervised feature learning framework for no-reference image quality assessment (2012). http://dx.doi.org/10.1109/cvpr.2012.6247789
Shao, F., Li, K., Lin, W., Jiang, G., Yu, M., Dai, Q.: Full-reference quality assessment of stereoscopic images by learning binocular receptive field properties. IEEE Trans. Image Process. 24, 2971–2983 (2015). https://doi.org/10.1109/TIP.2015.2436332
Zhang, Y., Zhang, H., Yu, M., Kwong, S., Ho, Y.-S.: Sparse representation based video quality assessment for synthesized 3D videos. IEEE Trans. Image Process. (2019). https://doi.org/10.1109/TIP.2019.2929433
Rijnders, C.E.: U.S. Patent Application No. 15/899,331 (2018)
Seshadrinathan, K., Soundararajan, R., Bovik, A.C., Cormack, L.K.: Study of subjective and objective quality assessment of video (2010). http://dx.doi.org/10.1109/tip.2010.2042111
Shen, J., Itti, L.: Top-down influences on visual attention during listening are modulated by observer sex. Vis. Res. 65, 62–76 (2012). https://doi.org/10.1016/j.visres.2012.06.001
Rimac-Drlje, S., Vranješ, M., Žagar, D.: Foveated mean squared error—a novel video quality metric (2016). http://dx.doi.org/10.1007/s11042-009-0442-1
Vranješ, M., Rimac-Drlje, S., Grgić, K.: Review of objective video quality metrics and performance comparison using different databases. Sign. Process.-Image Commun. (2012). http://dx.doi.org/10.1016/j.image.2012.10.003
Corriveau, P., Webster, A.: The video quality experts group: evaluates objective methods of video image quality assessment (1998). http://dx.doi.org/10.5594/m00304
Mele, M.L., Millar, D., Rijnders, C.E.: The web-based subjective quality assessment of an adaptive image compression plug-in (2017). http://dx.doi.org/10.5220/0006226401330137
Mele, M.L., Millar, D., Rijnders, C.E.: Sencogi spatio-temporal saliency: a new metric for predicting subjective video quality on mobile devices. In: Kurosu, M. (ed.) HCI 2018. LNCS, vol. 10902, pp. 552–564. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-91244-8_43
Mele, M.L., Millar, D., Rijnders, C.E.: Using spatio-temporal saliency to predict subjective video quality: a new high-speed objective assessment metric. In: Kurosu, M. (ed.) HCI 2017. LNCS, vol. 10271, pp. 353–368. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-58071-5_27
Mele, M.L., Colabrese, S., Calabria, L., Millar, D., Rijnders, C.E.: The assessment of sencogi: a visual complexity model predicting visual fixations. In: Kurosu, M. (ed.) HCII 2019. LNCS, vol. 11567, pp. 332–347. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-22643-5_26
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Mele, M.L., Colabrese, S., Calabria, L., Rijnders, C.E. (2020). Is It Possible to Predict Human Perception of Video Quality? The Assessment of Sencogi Quality Metric. In: Stephanidis, C., Marcus, A., Rosenzweig, E., Rau, PL.P., Moallem, A., Rauterberg, M. (eds) HCI International 2020 - Late Breaking Papers: User Experience Design and Case Studies. HCII 2020. Lecture Notes in Computer Science(), vol 12423. Springer, Cham. https://doi.org/10.1007/978-3-030-60114-0_16
Download citation
DOI: https://doi.org/10.1007/978-3-030-60114-0_16
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-60113-3
Online ISBN: 978-3-030-60114-0
eBook Packages: Computer ScienceComputer Science (R0)