Abstract
The images and videos randomly captured by people for sharing in social media or their own use are commonly termed as the user generated content. With the growing popularity of various social media and streaming platforms along with the availability of low-cost portable devices, the amount of such user generated content is exponentially increasing. Visual quality assessment of such user-generated contents is necessary for various purpose such as estimating how the distortions induced during media transmission effects the visual quality of the non-professionally captured image or video, the social media platforms to estimate quality of a media before it gets posted, assessing quality to evaluate performance of a handheld camera or mobile phone, etc. This is a very challenging task due to the fact that the user generated content significantly suffers from multiple artifacts and distortions during both the capturing and transmission pipeline stage that eventually hinder the visual quality. This work mostly deals with the artifacts induced during video capturing stage, and subsequently, use them for estimating visual quality. A random forest-based no-reference video quality assessment metric is proposed for user generated content videos. The proposed approach is divided into two steps. Firstly, the encoding and content-based features are extracted at the frame level. Secondly, an ensemble-based prediction model is employed to exploit the extracted frame-level features for estimating the visual quality score for each frame. Finally, max pooling is performed to estimate the final video level quality score. We also study various score predictors to eventually suggest the best performing ensemble-learning method for the proposed model. Experiments are performed on the benchmark ICME grand challenge dataset of user generated content videos. The model is compared with several state-of-the-art user generated content video quality metrics. The observed results indicate that the proposed no-reference model outperforms the existing approaches for quality estimation of user generated content videos.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Barman, N., Khan, N., Martini, M.G.: Analysis of spatial and temporal information variation for 10-bit and 8-bit video sequences. In: International Workshop on Computer Aided Modeling and Design of Communication Links and Networks, pp. 1–6. IEEE (2019)
Chen, B., Zhu, L., Li, G., Lu, F., Fan, H., Wang, S.: Learning generalized spatial-temporal deep feature representation for no-reference video quality assessment. IEEE Trans. Circuits Syst. Video Technol. (2021)
Chen, Y., Wu, K., Zhang, Q.: From QoS to QoE: a tutorial on video quality assessment. IEEE Commun. Surv. Tutor. 17(2), 1126–1165 (2014)
Chikkerur, S., Sundaram, V., Reisslein, M., Karam, L.J.: Objective video quality assessment methods: a classification, review, and performance comparison. IEEE Trans. Broadcast. 57(2), 165–182 (2011)
ITU-T Recommendation, P.: 910: Subjective video quality assessment methods for multimedia applications (2008)
Korhonen, J.: Two-level approach for no-reference consumer video quality assessment. IEEE Trans. Image Process. 28(12), 5923–5938 (2019)
Lamichhane, K., Mazumdar, P., Battisti, F., Carli, M.: A no reference deep learning based model for quality assessment of UGC videos. In: International Conference on Multimedia Expo Workshops, pp. 1–5 (2021)
Li, D., Jiang, T., Jiang, M.: Quality assessment of in-the-wild videos. In: ACM International Conference on Multimedia, pp. 2351–2359 (2019)
Mazumdar, P., Arru, G., Battisti, F.: Early detection of children with autism spectrum disorder based on visual exploration of images. Signal Process. Image Commun. 94, 116184 (2021)
Mittal, A., Soundararajan, R., Bovik, A.C.: Making a “completely blind” image quality analyzer. IEEE Signal Process. Lett. 20(3), 209–212 (2012)
Mittal, A., Moorthy, A.K., Bovik, A.C.: No-reference image quality assessment in the spatial domain. IEEE Trans. Image Process. 21(12), 4695–4708 (2012)
Nguyen, D., Tran, H., Thang, T.C.: An ensemble learning-based no reference QOE model for user generated contents. In: International Conference on Multimedia Expo Workshops, pp. 1–6 (2021)
Nguyen, D., Tran, H., Thang, T.C.: An ensemble learning-based no reference QOE model for user generated contents. In: IEEE International Conference on Multimedia and Expo Workshops, pp. 1–6. IEEE (2021)
Raake, A., Garcia, M.N., Robitza, W., List, P., Göring, S., Feiten, B.: A bitstream-based, scalable video-quality model for http adaptive streaming: ITU-T, p. 1203.1. In: International Conference on Quality of Multimedia Experience, pp. 1–6 (2017)
Stanković, R.S., Falkowski, B.J.: The HAAR wavelet transform: its status and achievements. Comput. Elect. Eng. 29(1), 25–44 (2003)
Tong, H., Li, M., Zhang, H., Zhang, C.: Blur detection for digital images using wavelet transform. In: IEEE International Conference on Multimedia and Expo, vol. 1, pp. 17–20. IEEE (2004)
Tu, Z., Yu, X., Wang, Y., Birkbeck, N., Adsumilli, B., Bovik, A.C.: Rapique: rapid and accurate video quality prediction of user generated content. IEEE Open J. Signal Process. 2, 425–440 (2021)
Varga, D.: No-reference video quality assessment based on Bedford’s law and perceptual features. Electronics. 10(22) (2021). https://www.mdpi.com/2079-9292/10/22/2768
Wang, Y., et al.: Rich features for perceptual quality assessment of UGC videos. In: IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 13435–13444 (2021)
Yao, J.Y., Liu, G.: Bitrate-based no-reference video quality assessment combining the visual perception of video contents. IEEE Trans. Broadcast. 65(3), 546–557 (2018)
Yu, X., et al.: Subjective quality assessment of user-generated content gaming videos. In: IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 74–83 (2022)
Zadtootaghaj, S., et al.: Demi: deep video quality estimation model using perceptual video quality dimensions. In: International Workshop on Multimedia Signal Processing, pp. 1–6. IEEE (2020)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Kumar, K., Mazumdar, P., Jha, K.K., Lamichhane, K. (2023). A Random Forest-based No-Reference Quality Metric for UGC Videos. In: Gupta, D., Bhurchandi, K., Murala, S., Raman, B., Kumar, S. (eds) Computer Vision and Image Processing. CVIP 2022. Communications in Computer and Information Science, vol 1776. Springer, Cham. https://doi.org/10.1007/978-3-031-31407-0_41
Download citation
DOI: https://doi.org/10.1007/978-3-031-31407-0_41
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-31406-3
Online ISBN: 978-3-031-31407-0
eBook Packages: Computer ScienceComputer Science (R0)