Skip to main content

Self-SLAM: A Self-supervised Learning Based Annotation Method to Reduce Labeling Overhead

  • Conference paper
  • First Online:
Machine Learning and Knowledge Discovery in Databases. Applied Data Science Track (ECML PKDD 2024)

Abstract

In recent times, Deep Neural Networks (DNNs) have been effectively used to tackle various tasks such as emotion recognition, activity detection, disease prediction, and surface classification. However, a major challenge in developing models for these tasks requires a large amount of labeled data for accurate predictions. The manual annotation process for a large dataset is expensive, time-consuming, and error-prone. Thus, we present SSLAM (Self-supervised Learning-based Annotation Method) framework to tackle this challenge. SSLAM is a self-supervised deep learning framework designed to generate labels while minimizing the overhead associated with tabular data annotation. SSLAM learns valuable representations from unlabeled data that are applied to the downstream task of label generation by utilizing two pretext tasks with a novel \(log-cosh\) loss function. SSLAM outperforms supervised learning and Value Imputation and Mask Estimation (VIME) baselines on two datasets - Continuously Annotated Signals of Emotion (CASE) and wheelchair dataset. The wheelchair dataset is our novel unique surface classification dataset collected using wheelchairs showcasing our framework’s effectiveness in real-world scenarios. All these results reinforce that SSLAM significantly reduces the labeling overhead, especially when there is a vast amount of unlabeled data compared to labeled data. The code for this paper can be viewed at the following link: https://github.com/Alfiya-M-H-Shaikh/SSLAM.git

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Abdel Hakim, A.E., Deabes, W.: Can people really do nothing? handling annotation gaps in adl sensor data. Algorithms 12(10), 217 (2019)

    Article  Google Scholar 

  2. Bradley, M.M., Lang, P.J.: Measuring emotion: the self-assessment manikin and the semantic differential. J. Behav. Ther. Exp. Psychiatry 25(1), 49–59 (1994)

    Article  Google Scholar 

  3. Chatterjee, S., Chakma, A., Gangopadhyay, A., Roy, N., Mitra, B., Chakraborty, S.: Laso: exploiting locomotive and acoustic signatures over the edge to annotate imu data for human activity recognition. In: Proceedings of the 2020 International Conference on Multimodal Interaction, pp. 333–342 (2020)

    Google Scholar 

  4. Chen, Y., Chang, R., Guo, J.: Effects of data augmentation method borderline-smote on emotion recognition of eeg signals based on convolutional neural network. IEEE Access 9 (2021)

    Google Scholar 

  5. Cowie, R., Douglas-Cowie, E., Savvidou, S., McMahon, E., Sawey, M., Schröder, M.: ‘FEELTRACE’: an instrument for recording perceived emotion in real time. In: ITRW Speech-Emotion (2000)

    Google Scholar 

  6. Dissanayake, V., Seneviratne, S., Rana, R., Wen, E., Kaluarachchi, T., Nanayakkara, S.: Sigrep: toward robust wearable emotion recognition with contrastive representation learning. IEEE Access 10, 18105–18120 (2022)

    Article  Google Scholar 

  7. Garcia-Ceja, E., Riegler, M., Nordgreen, T., Jakobsen, P., Oedegaard, K.J., Tørresen, J.: Mental health monitoring with multimodal sensing and machine learning: a survey. Pervasive Mob. Comput. 51, 1–26 (2018)

    Article  Google Scholar 

  8. Girard, J.M., Wright, A.G.: Darma: software for dual axis rating and media annotation. Behav. Res. Methods 50(3), 902–909 (2018)

    Article  Google Scholar 

  9. Hossain, H.S., Khan, M.A.A.H., Roy, N.: Active learning enabled activity recognition. Pervasive Mob. Comput. 38, 312–330 (2017)

    Article  Google Scholar 

  10. Jing, L., Tian, Y.: Self-supervised visual feature learning with deep neural networks: a survey. IEEE Trans. Pattern Anal. Mach. Intell. 43(11) (2020)

    Google Scholar 

  11. Mediratta, I., Saha, S., Mathur, S.: Liparelu: arelu networks aided by lipschitz acceleration. In: 2021 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE (2021)

    Google Scholar 

  12. Muralidharan, A., Gyongyi, Z., Chi, E.: Social annotations in web search. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 1085–1094 (2012)

    Google Scholar 

  13. Nowak, S., Rüger, S.: How reliable are annotations via crowdsourcing: a study about inter-annotator agreement for multi-label image annotation. In: Proceedings of the International Conference on Multimedia Information Retrieval, pp. 557–566 (2010)

    Google Scholar 

  14. Qiu, X., Sun, T., Xu, Y., Shao, Y., Dai, N., Huang, X.: Pre-trained models for natural language processing: a survey. SCIENCE CHINA Technol. Sci. 63(10), 1872–1897 (2020)

    Article  Google Scholar 

  15. Ronao, C.A., Cho, S.B.: Human activity recognition with smartphone sensors using deep learning neural networks. Expert Syst. Appl. 59, 235–244 (2016)

    Article  Google Scholar 

  16. Russell, J.A.: A circumplex model of affect. J. Pers. Soc. Psychol. 39(6), 1161 (1980)

    Article  Google Scholar 

  17. Sarkar, P., Etemad, A.: Self-supervised ecg representation learning for emotion recognition. IEEE Trans. Affective Comput. (2020)

    Google Scholar 

  18. Settles, B.: Active learning literature survey. Technical Report, University of Wisconsin-Madison Department of Computer Sciences (2009)

    Google Scholar 

  19. Sharma, K., Castellini, C., van den Broek, E.L., Albu-Schaeffer, A., Schwenker, F.: A dataset of continuous affect annotations and physiological signals for emotion analysis. Sci. Data 6(1), 1–13 (2019)

    Article  Google Scholar 

  20. Subramanian, R., Wache, J., Abadi, M.K., Vieriu, R.L., Winkler, S., Sebe, N.: Ascertain: emotion and personality recognition using commercial sensors. IEEE Trans. Affect. Comput. 9(2), 147–160 (2016)

    Article  Google Scholar 

  21. Tang, H., Liu, W., Zheng, W.L., Lu, B.L.: Multimodal emotion recognition using deep neural networks. In: International Conference on Neural Information Processing, pp. 811–819. Springer (2017)

    Google Scholar 

  22. Wang, Y., Nazir, S., Shafiq, M.: An overview on analyzing deep learning and transfer learning approaches for health monitoring. Computational and Mathematical Methods in Medicine 2021 (2021)

    Google Scholar 

  23. Yang, J., Fan, J., Wei, Z., Li, G., Liu, T., Du, X.: Cost-effective data annotation using game-based crowdsourcing. Proc. VLDB Endowment 12(1), 57–70 (2018)

    Article  Google Scholar 

  24. Yoon, J., Zhang, Y., Jordon, J., van der Schaar, M.: Vime: extending the success of self-and semi-supervised learning to tabular domain. Adv. Neural. Inf. Process. Syst. 33, 11033–11043 (2020)

    Google Scholar 

  25. Yu, H., Raychoudhury, V., Saha, S., Edinger, J., Smith, R.O., Gani, M.O.: Automated surface classification system using vibration patterns-a case study with wheelchairs. IEEE Trans. Artif. Intell. 4(4), 884–895 (2023). https://doi.org/10.1109/TAI.2022.3190828

    Article  Google Scholar 

  26. Zhai, X., Oliver, A., Kolesnikov, A., Beyer, L.: S4l: self-supervised semi-supervised learning. In: Proceedings of the IEEE/CVF International Conference on Computer Vision (2019)

    Google Scholar 

  27. Zhang, T., El Ali, A., Wang, C., Hanjalic, A., Cesar, P.: Corrnet: fine-grained emotion recognition for video watching using wearable physiological sensors. Sensors 21(1), 52 (2020)

    Article  Google Scholar 

  28. Zhang, T., El Ali, A., Wang, C., Hanjalic, A., Cesar, P.: Rcea: real-time, continuous emotion annotation for collecting precise mobile video ground truth labels. In: Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, pp. 1–15 (2020)

    Google Scholar 

  29. Zhang, Z., Zhong, S.h., Liu, Y.: Ganser: a self-supervised data augmentation framework for eeg-based emotion recognition. IEEE Trans. Affective Comput. (2022)

    Google Scholar 

Download references

Acknowledgments

Snehanshu Saha, Surjya Ghosh and Sougata Sen would like to thank the Anuradha and Prashanth Palakurthi Center for Artificial Intelligence Research (APPCAIR), SERB-DST (SUR/2022/001965) and SERB CRG-DST (CRG/2023/003210), Govt. of India for partially supporting the work. Swarnali Banik gratefully acknowledges the Chanakya Fellowship from AI4CPS Innovation Hub, IIT Kharagpur.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Alfiya M. Shaikh .

Editor information

Editors and Affiliations

1 Electronic supplementary material

Below is the link to the electronic supplementary material.

Supplementary material 1 (pdf 207 KB)

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Shaikh, A.M. et al. (2024). Self-SLAM: A Self-supervised Learning Based Annotation Method to Reduce Labeling Overhead. In: Bifet, A., Krilavičius, T., Miliou, I., Nowaczyk, S. (eds) Machine Learning and Knowledge Discovery in Databases. Applied Data Science Track. ECML PKDD 2024. Lecture Notes in Computer Science(), vol 14949. Springer, Cham. https://doi.org/10.1007/978-3-031-70378-2_8

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-70378-2_8

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-70377-5

  • Online ISBN: 978-3-031-70378-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics