Skip to main content

Identifying Toxicity Within YouTube Video Comment

  • Conference paper
  • First Online:

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 11549))

Abstract

Online Social Networks (OSNs), once regarded as safe havens for sharing information and providing mutual support among groups of people, have become breeding grounds for spreading toxic behaviors, political propaganda, and radicalizing content. Toxic individuals often hide under the auspices of anonymity to create fruitless arguments and divert the attention of other users from the core objectives of a community. In this study, we examined five recurring forms of toxicity among the comments posted on pro- and anti-NATO channels on YouTube. We leveraged the YouTube Data API to collect video and comment data from eight channels. We then utilized Google’s Perspective API to assign toxic scores to each comment. Our analysis suggests that, on average, commenters on the anti-NATO channels are more likely to be more toxic than those on the pro-NATO channels. We further discovered that commenters on pro-NATO channels tend to use a mixture of toxic and innocuous comments. We generated word clouds to get an idea of word use frequency, as well as applied the Latent Dirichlet Allocation topic model to classify the comments into their overall topics. The topics extracted from the pro-NATO channels’ comments were primarily positive, such as “Alliance” and “United”; whereas, the topics extracted from anti-NATO channels’ comments were more geared towards geographical locations, such as “Russia”, and negative components such as “Profanity” and “Fake News”. By identifying and examining the toxic behaviors of commenters on YouTube, our analysis lends aid to the pressing need for understanding this toxicity.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   49.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   64.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. Cheng, J., Bernstein, M., Danescu-Niculescu-Mizil, C., Leskovec, J.: Anyone can become a troll: causes of trolling behavior in online discussions. In: Proceedings of the 2017 ACM Conference on Computer Supported Cooperative Work and Social Computing - CSCW 2017, pp. 1217–1230. ACM Press, Portland (2017)

    Google Scholar 

  2. Lee, S.-H., Kim, H.-W.: Why people post benevolent and malicious comments online. Commun. ACM 58, 74–79 (2015)

    Article  Google Scholar 

  3. Wulczyn, E., Thain, N., Dixon, L.: Ex machina: personal attacks seen at scale. arXiv:1610.08914 [cs] (2016)

  4. Nobata, C., Tetreault, J., Thomas, A., Mehdad, Y., Chang, Y.: Abusive language detection in online user content. Presented at the (2016)

    Google Scholar 

  5. Martens, M., Shen, S., Iosup, A., Kuipers, F.: Toxicity detection in multiplayer online games. In: 2015 International Workshop on Network and Systems Support for Games (NetGames), pp. 1–6. IEEE, Zagreb (2015)

    Google Scholar 

  6. Online Harassment | Pew Research Center (2014). http://www.pewinternet.org/2014/10/22/online-harassment/

  7. Chen, Y., Zhou, Y., Zhu, S., Xu, H.: Detecting offensive language in social media to protect adolescent online safety. In: 2012 International Conference on Privacy, Security, Risk and Trust and 2012 International Conference on Social Computing, pp. 71–80. IEEE, Amsterdam (2012)

    Google Scholar 

  8. Cao, Q., Yang, X., Yu, J., Palow, C.: Uncovering Large Groups of Active Malicious Accounts in Online Social Networks. Presented at the (2014)

    Google Scholar 

  9. Perspective. http://perspectiveapi.com/#/

  10. Varjas, K., Talley, J., Meyers, J., Parris, L., Cutts, H.: High school students’ perceptions of motivations for cyberbullying: an exploratory study. West J. Emerg. Med. 11, 269–273 (2010)

    Google Scholar 

  11. Shachaf, P., Hara, N.: Beyond vandalism: Wikipedia trolls. J. Inf. Sci. 36, 357–370 (2010)

    Article  Google Scholar 

  12. Warner, W., Hirschberg, J.: Detecting hate speech on the world wide web. In: Proceedings of the Second Workshop on Language in Social Media, pp. 19–26. Association for Computational Linguistics, Montréal (2012)

    Google Scholar 

  13. Suler, J.: The online disinhibition effect. Cyberpsychol. Behav. 7(3), 321–326 (2004)

    Article  Google Scholar 

  14. Cheng, J., Danescu-Niculescu-Mizil, C., Leskovec, J.: Antisocial behavior in online discussion communities, 10

    Google Scholar 

  15. Yin, D., Xue, Z., Hong, L.: Detection of harassment on Web 2.0., 7

    Google Scholar 

  16. Sood, S.O., Antin, J., Churchill, E.F.: Using crowdsourcing to improve profanity detection, 6

    Google Scholar 

  17. Davidson, T., Warmsley, D., Macy, M., Weber, I.: Automated hate speech detection and the problem of offensive language. In: Proceedings of the Eleventh International AAAI Conference on Web and Social Media (ICWSM 2017) (2017)

    Google Scholar 

  18. Hosseini, H., Kannan, S., Zhang, B., Poovendran, R.: Deceiving Google’s perspective API built for detecting toxic comments. arXiv:1702.08138 [cs] (2017)

  19. Gröndahl, T., Pajola, L., Juuti, M., Conti, M., Asokan, N.: All you need is “love”: evading hate-speech detection. arXiv:1808.09115 [cs] (2018)

  20. Carley, K.M., Reminga, J.: ORA: Organization Risk Analyzer: Defense Technical Information Center. Fort Belvoir, VA (2004)

    Book  Google Scholar 

  21. Responding to Cognitive Security Challenges | StratCom. https://www.stratcomcoe.org/responding-cognitive-security-challenges

  22. Blei, D.M.: Latent Dirichlet allocation, 30

    Google Scholar 

  23. van der Maaten, L., Hinton, G.: Visualizing data using t-SNE. J. Mach. Learn. Res. 9, 2579–2605 (2008)

    MATH  Google Scholar 

Download references

Acknowledgements

This research is funded in part by the U.S. National Science Foundation (IIS-1636933, ACI-1429160, and IIS-1110868), U.S. Office of Naval Research (N00014-10-1-0091, N00014-14-1-0489, N00014-15-P-1187, N00014-16-1-2016, N00014-16-1-2412, N00014-17-1-2605, N00014-17-1-2675), U.S. Air Force Research Lab, U.S. Army Research Office (W911NF-16-1-0189), U.S. Defense Advanced Research Projects Agency (W31P4Q-17-C-0059), Jerry L. Maulden/Entergy Endowment at the University of Arkansas at Little Rock and the Arkansas Research Alliance. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the funding organizations. The researchers gratefully acknowledge the support.

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to Adewale Obadimu , Esther Mead , Muhammad Nihal Hussain or Nitin Agarwal .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2019 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Obadimu, A., Mead, E., Hussain, M.N., Agarwal, N. (2019). Identifying Toxicity Within YouTube Video Comment. In: Thomson, R., Bisgin, H., Dancy, C., Hyder, A. (eds) Social, Cultural, and Behavioral Modeling. SBP-BRiMS 2019. Lecture Notes in Computer Science(), vol 11549. Springer, Cham. https://doi.org/10.1007/978-3-030-21741-9_22

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-21741-9_22

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-21740-2

  • Online ISBN: 978-3-030-21741-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics