Skip to main content

Detecting Traces of Self-harm in Social Media: A Simple and Interpretable Approach

  • Conference paper
  • First Online:

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 13068))

Abstract

Social networks have become the main means of communication and interaction between people. In them, users share information and opinions, but also their experiences, worries, and personal concerns. Because of this, there is a growing interest in analyzing this kind of content to identify people who commit self-harm, which is often one of the first signs of suicide risk. Recently, methods based on Deep Learning have shown good results in this task, however, they are opaque and do not facilitate the interpretation of decisions, something fundamental in health-related tasks. In this paper, we face the detection of self-harm in social media by applying a simple and interpretable one-class-classification approach, which, supported on the concept of the attraction force [1], produces its decisions considering both the relevance and distance between users. The results obtained in a benchmark dataset are encouraging, as they indicate a competitive performance with respect to state-of-the-art methods. Furthermore, taking advantage of the approach’s properties, we outline what could be a support tool for healthcare professionals for analyzing and monitoring self-harm behaviors in social networks.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   89.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   119.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Notes

  1. 1.

    https://www.mayoclinic.org/es-es/diseases-conditions/self-injury/symptoms-causes/syc-20350950.

  2. 2.

    https://clpsych.org/.

  3. 3.

    It is an online community of Australian youth.

  4. 4.

    https://early.irlab.org/.

  5. 5.

    https://huggingface.co/transformers/pretrained_models.html.

  6. 6.

    Martínez-Castaño et al. exploited data collected from Pushshift Reddit Dataset [5].

References

  1. Aguilera, J., González, L.C., Montes-y-Gómez, M., López, R., Escalante, H.J.: From neighbors to strengths - the k-strongest strengths (kSS) classification algorithm. Pattern Recogn. Lett. 136, 301–308 (2020)

    Article  Google Scholar 

  2. Aguilera, J., Hernández Farías, D.I., Ortega-Mendoza, R.M., Montes-y Gómez, M.: Depression and anorexia detection in social media as a one-class classification problem. Appl. Intell. 1–16 (2021)

    Google Scholar 

  3. Alhassan, M.A., Inuwa-Dutse, I., Bello, B.S., Pennington, D.R.: Self-harm: detection and support on twitter. CoRR abs/2104.00174 (2021)

    Google Scholar 

  4. Aragón, M., López-Monroy, A.P., y Gómez, M.M.: INAOE-CIMAT at eRisk 2020: detecting signs of self-harm using sub-emotions and words. In: CLEF (2020)

    Google Scholar 

  5. Baumgartner, J., Zannettou, S., Keegan, B., Squire, M., Blackburn, J.: The pushshift reddit dataset. In: Proceedings of the International AAAI Conference on Web and Social Media, vol. 14, no. 1, pp. 830–839 (2020)

    Google Scholar 

  6. Bojanowski, P., Grave, E., Joulin, A., Mikolov, T.: Enriching word vectors with subword information. Trans. Assoc. Comput. Ling. 5, 135–146 (2017)

    Google Scholar 

  7. Danilevsky, M., Qian, K., Aharonov, R., Katsis, Y., Kawas, B., Sen, P.: A survey of the state of explainable AI for natural language processing. In: Proceedings of the 1st Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 10th International Joint Conference on Natural Language Processing, pp. 447–459. ACL, Suzhou, China, December 2020

    Google Scholar 

  8. Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp. 4171–4186. ACL, Minneapolis, Minnesota, June 2019

    Google Scholar 

  9. Gkotsis, G., et al.: The language of mental health problems in social media. In: Proceedings of the Third Workshop on Computational Linguistics and Clinical Psychology, pp. 63–73. ACL, San Diego, CA, USA, June 2016

    Google Scholar 

  10. Gkotsis, G., et al.: Characterisation of mental health conditions in social media using informed deep learning. Sci. Rep. 7(1), 45141 (2017)

    Article  Google Scholar 

  11. Ive, J., Gkotsis, G., Dutta, R., Stewart, R., Velupillai, S.: Hierarchical neural model with attention mechanisms for the classification of social media text related to mental health. In: Proceedings of the Fifth Workshop on Computational Linguistics and Clinical Psychology: From Keyboard to Clinic, pp. 69–77. ACL, June 2018

    Google Scholar 

  12. Khan, S.S., Madden, M.G.: One-class classification: taxonomy of study and review of techniques. Knowl. Eng. Rev. 29(3), 345–374 (2014)

    Article  Google Scholar 

  13. Laye-Gindhu, A., Schonert-Reichl, K.: Nonsuicidal self-harm among community adolescents: understanding the “whats" and “whys" of self-harm. J. Youth Adolesc. 34, 447–457 (2005)

    Google Scholar 

  14. Losada, D.E., Crestani, F., Parapar, J.: Overview of eRisk at CLEF 2020: early risk prediction on the internet (extended overview). In: Cappellato, L., Eickhoff, C., Ferro, N., Névéol, A. (eds.) Conference and Labs of the Evaluation Forum. CEUR Workshop Proceedings (2020)

    Google Scholar 

  15. Martínez-Castaño, R., Htait, A., Azzopardi, L., Moshfeghi, Y.: Early risk detection of self-harm and depression severity using BERT-based transformers. In: Working Notes of CLEF 2020 - Conference and Labs of the Evaluation Forum, Thessaloniki, Greece, 22–25 September 2020. CEUR Workshop Proceedings, vol. 2696. CEUR-WS.org (2020)

    Google Scholar 

  16. Mikolov, T., Chen, K., Corrado, G., Dean, J.: Efficient estimation of word representations in vector space. In: Bengio, Y., LeCun, Y. (eds.) 1st Workshop Track Proceedings International Conference on Learning Representations, ICLR 2013, Scottsdale, Arizona, USA, 2–4 May 2013 (2013)

    Google Scholar 

  17. Milne, D.N., Pink, G., Hachey, B., Calvo, R.A.: CLPsych 2016 shared task: triaging content in online peer-support forums. In: Proceedings of the Third Workshop on Computational Linguistics and Clinical Psychology, pp. 118–127. ACL, June 2016

    Google Scholar 

  18. Pennington, J., Socher, R., Manning, C.D.: GloVe: global Vectors for Word representation. In: Empirical Methods in Natural Language Processing (EMNLP), pp. 1532–1543 (2014)

    Google Scholar 

  19. Scherr, S., Arendt, F., Frissen, T., Oramas, M.J.: Detecting intentional self-harm on instagram: development, testing, and validation of an automatic image-recognition algorithm to discover cutting-related posts. Soc. Sci. Comput. Rev. 38(6), 673–685 (2020)

    Google Scholar 

  20. Wang, Y., Tang, J., Li, J., Li, B., Wan, Y., Mellina, C., O’Hare, N., Chang, Y.: Understanding and discovering deliberate self-harm content in social media. In: Proceedings of the 26th International Conference on World Wide Web, pp. 93–102. WWW 2017, International World Wide Web Conferences Steering Committee, Republic and Canton of Geneva, CHE (2017)

    Google Scholar 

  21. Yates, A., Cohan, A., Goharian, N.: Depression and self-harm risk assessment in online forums. In: Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pp. 2968–2978. ACL, September 2017

    Google Scholar 

  22. Zirikly, A., Resnik, P., Uzuner, Ö., Hollingshead, K.: CLPsych 2019 shared task: predicting the degree of suicide risk in reddit posts. In: Proceedings of the Sixth Workshop on Computational Linguistics and Clinical Psychology, pp. 24–33. ACL, Minneapolis, Minnesota, June 2019

    Google Scholar 

Download references

Acknowledgments

This research was funded by the CONACYT project CB-2015-01-257383.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Delia Irazú Hernández Farías .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2021 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Aguilera, J., Farías, D.I.H., Montes-y-Gómez, M., González, L.C. (2021). Detecting Traces of Self-harm in Social Media: A Simple and Interpretable Approach. In: Batyrshin, I., Gelbukh, A., Sidorov, G. (eds) Advances in Soft Computing. MICAI 2021. Lecture Notes in Computer Science(), vol 13068. Springer, Cham. https://doi.org/10.1007/978-3-030-89820-5_16

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-89820-5_16

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-89819-9

  • Online ISBN: 978-3-030-89820-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics