Skip to main content

A Random Forest-based No-Reference Quality Metric for UGC Videos

  • Conference paper
  • First Online:
Computer Vision and Image Processing (CVIP 2022)

Part of the book series: Communications in Computer and Information Science ((CCIS,volume 1776))

Included in the following conference series:

  • 518 Accesses

Abstract

The images and videos randomly captured by people for sharing in social media or their own use are commonly termed as the user generated content. With the growing popularity of various social media and streaming platforms along with the availability of low-cost portable devices, the amount of such user generated content is exponentially increasing. Visual quality assessment of such user-generated contents is necessary for various purpose such as estimating how the distortions induced during media transmission effects the visual quality of the non-professionally captured image or video, the social media platforms to estimate quality of a media before it gets posted, assessing quality to evaluate performance of a handheld camera or mobile phone, etc. This is a very challenging task due to the fact that the user generated content significantly suffers from multiple artifacts and distortions during both the capturing and transmission pipeline stage that eventually hinder the visual quality. This work mostly deals with the artifacts induced during video capturing stage, and subsequently, use them for estimating visual quality. A random forest-based no-reference video quality assessment metric is proposed for user generated content videos. The proposed approach is divided into two steps. Firstly, the encoding and content-based features are extracted at the frame level. Secondly, an ensemble-based prediction model is employed to exploit the extracted frame-level features for estimating the visual quality score for each frame. Finally, max pooling is performed to estimate the final video level quality score. We also study various score predictors to eventually suggest the best performing ensemble-learning method for the proposed model. Experiments are performed on the benchmark ICME grand challenge dataset of user generated content videos. The model is compared with several state-of-the-art user generated content video quality metrics. The observed results indicate that the proposed no-reference model outperforms the existing approaches for quality estimation of user generated content videos.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

Notes

  1. 1.

    https://www.wyzowl.com/video-social-media-2020/.

  2. 2.

    http://ugcvqa.com/.

  3. 3.

    http://ugcvqa.com/.

  4. 4.

    http://2021.ieeeicme.org/2021.ieeeicme.org/conf_challenges.html.

  5. 5.

    https://github.com/slhck/siti.

  6. 6.

    https://github.com/FFmpeg/FFmpeg.

  7. 7.

    https://github.com/WillBrennan/BlurDetection2.

References

  1. Barman, N., Khan, N., Martini, M.G.: Analysis of spatial and temporal information variation for 10-bit and 8-bit video sequences. In: International Workshop on Computer Aided Modeling and Design of Communication Links and Networks, pp. 1–6. IEEE (2019)

    Google Scholar 

  2. Chen, B., Zhu, L., Li, G., Lu, F., Fan, H., Wang, S.: Learning generalized spatial-temporal deep feature representation for no-reference video quality assessment. IEEE Trans. Circuits Syst. Video Technol. (2021)

    Google Scholar 

  3. Chen, Y., Wu, K., Zhang, Q.: From QoS to QoE: a tutorial on video quality assessment. IEEE Commun. Surv. Tutor. 17(2), 1126–1165 (2014)

    Article  Google Scholar 

  4. Chikkerur, S., Sundaram, V., Reisslein, M., Karam, L.J.: Objective video quality assessment methods: a classification, review, and performance comparison. IEEE Trans. Broadcast. 57(2), 165–182 (2011)

    Article  Google Scholar 

  5. ITU-T Recommendation, P.: 910: Subjective video quality assessment methods for multimedia applications (2008)

    Google Scholar 

  6. Korhonen, J.: Two-level approach for no-reference consumer video quality assessment. IEEE Trans. Image Process. 28(12), 5923–5938 (2019)

    Article  MathSciNet  MATH  Google Scholar 

  7. Lamichhane, K., Mazumdar, P., Battisti, F., Carli, M.: A no reference deep learning based model for quality assessment of UGC videos. In: International Conference on Multimedia Expo Workshops, pp. 1–5 (2021)

    Google Scholar 

  8. Li, D., Jiang, T., Jiang, M.: Quality assessment of in-the-wild videos. In: ACM International Conference on Multimedia, pp. 2351–2359 (2019)

    Google Scholar 

  9. Mazumdar, P., Arru, G., Battisti, F.: Early detection of children with autism spectrum disorder based on visual exploration of images. Signal Process. Image Commun. 94, 116184 (2021)

    Article  Google Scholar 

  10. Mittal, A., Soundararajan, R., Bovik, A.C.: Making a “completely blind” image quality analyzer. IEEE Signal Process. Lett. 20(3), 209–212 (2012)

    Google Scholar 

  11. Mittal, A., Moorthy, A.K., Bovik, A.C.: No-reference image quality assessment in the spatial domain. IEEE Trans. Image Process. 21(12), 4695–4708 (2012)

    Article  MathSciNet  MATH  Google Scholar 

  12. Nguyen, D., Tran, H., Thang, T.C.: An ensemble learning-based no reference QOE model for user generated contents. In: International Conference on Multimedia Expo Workshops, pp. 1–6 (2021)

    Google Scholar 

  13. Nguyen, D., Tran, H., Thang, T.C.: An ensemble learning-based no reference QOE model for user generated contents. In: IEEE International Conference on Multimedia and Expo Workshops, pp. 1–6. IEEE (2021)

    Google Scholar 

  14. Raake, A., Garcia, M.N., Robitza, W., List, P., Göring, S., Feiten, B.: A bitstream-based, scalable video-quality model for http adaptive streaming: ITU-T, p. 1203.1. In: International Conference on Quality of Multimedia Experience, pp. 1–6 (2017)

    Google Scholar 

  15. Stanković, R.S., Falkowski, B.J.: The HAAR wavelet transform: its status and achievements. Comput. Elect. Eng. 29(1), 25–44 (2003)

    Article  MATH  Google Scholar 

  16. Tong, H., Li, M., Zhang, H., Zhang, C.: Blur detection for digital images using wavelet transform. In: IEEE International Conference on Multimedia and Expo, vol. 1, pp. 17–20. IEEE (2004)

    Google Scholar 

  17. Tu, Z., Yu, X., Wang, Y., Birkbeck, N., Adsumilli, B., Bovik, A.C.: Rapique: rapid and accurate video quality prediction of user generated content. IEEE Open J. Signal Process. 2, 425–440 (2021)

    Article  Google Scholar 

  18. Varga, D.: No-reference video quality assessment based on Bedford’s law and perceptual features. Electronics. 10(22) (2021). https://www.mdpi.com/2079-9292/10/22/2768

  19. Wang, Y., et al.: Rich features for perceptual quality assessment of UGC videos. In: IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 13435–13444 (2021)

    Google Scholar 

  20. Yao, J.Y., Liu, G.: Bitrate-based no-reference video quality assessment combining the visual perception of video contents. IEEE Trans. Broadcast. 65(3), 546–557 (2018)

    Article  MathSciNet  Google Scholar 

  21. Yu, X., et al.: Subjective quality assessment of user-generated content gaming videos. In: IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 74–83 (2022)

    Google Scholar 

  22. Zadtootaghaj, S., et al.: Demi: deep video quality estimation model using perceptual video quality dimensions. In: International Workshop on Multimedia Signal Processing, pp. 1–6. IEEE (2020)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Pramit Mazumdar .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Kumar, K., Mazumdar, P., Jha, K.K., Lamichhane, K. (2023). A Random Forest-based No-Reference Quality Metric for UGC Videos. In: Gupta, D., Bhurchandi, K., Murala, S., Raman, B., Kumar, S. (eds) Computer Vision and Image Processing. CVIP 2022. Communications in Computer and Information Science, vol 1776. Springer, Cham. https://doi.org/10.1007/978-3-031-31407-0_41

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-31407-0_41

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-31406-3

  • Online ISBN: 978-3-031-31407-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics