Skip to main content
Log in

EnsemConvNet: a deep learning approach for human activity recognition using smartphone sensors for healthcare applications

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

Human Activity Recognition (HAR) can be defined as the automatic prediction of the regular human activities performed in our day-to-day life, such as walking, running, cooking, performing office work, etc. It is truly beneficial in the field of medical care services, for example, personal health care assistants, old-age care services, maintaining patient records for future help, etc. Input data to a HAR system can be (a) videos or still images capturing human activities, or (b) time-series data of human body movements while performing the activities taken from sensors in the smart devices like accelerometer, gyroscope, etc. In this work, we mainly focus on the second category of the input data. Here, we propose an ensemble of three classification models, namely CNN-Net, Encoded-Net, and CNN-LSTM, which is named as EnsemConvNet. Each of these classification models is built upon simple 1D Convolutional Neural Network (CNN) but differs in terms of the number of dense layers, kernel size used along with other key differences in the architecture. Each model accepts the time series data as a 2D matrix by taking a window of data at a time in order to infer information, which ultimately predicts the type of human activity. Classification outcome of the EnsemConvNet model is decided using various classifier combination methods that include majority voting, sum rule, product rule, and a score fusion approach called adaptive weighted approach. Three benchmark datasets, namely WISDM activity prediction, UniMiB SHAR, MobiAct, are used for evaluating our proposed model. We have compared our EnsemConvNet model with some existing deep learning models such as Multi Headed CNN, hybrid of CNN, and Long Short Term Memory (LSTM) models. The results obtained here establish the supremacy of the EnsemConvNet model over the other mentioned models.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15
Fig. 16
Fig. 17
Fig. 18
Fig. 19
Fig. 20
Fig. 21
Fig. 22

Similar content being viewed by others

References

  1. Aggarwal JK, Xia L (2014) Letters, and undefined, “Human activity recognition from 3d data: A review,” Elsevier. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S0167865514001299. Accessed 01 Jul 2020

  2. Akhavian R, Behzadan AH (2016) Smartphone-based construction workers’ activity recognition and classification. Autom Constr 71(Part 2):198–209. https://doi.org/10.1016/j.autcon.2016.08.015

    Article  Google Scholar 

  3. Anguita D, Ghio A, Oneto L, Parra X, Reyes-Ortiz JL (2013) A Public Domain Dataset for Human Activity Recognition Using Smartphones

  4. Barua S, Islam M, Murase K (2011) On Neural, and undefined, “A novel synthetic minority oversampling technique for imbalanced data set learning,” Springer, Accessed 01 Jul 2020. [Online]. Available: https://doi.org/10.1007/978-3-642-24958-7_85.

  5. Bhattacharya S, Shaw V, Singh PK, Sarkar R, Bhattacharjee D (2020) SV-NET: A Deep Learning Approach To Video Based Human Activity Recognition. Proceedings of the Eleventh International Conference on Soft Computing and Pattern Recognition, SoCPaR 2.”

  6. Breiman L (2001) Random forests. Mach Learn 45(1):5–32. https://doi.org/10.1023/A:1010933404324

    Article  MATH  Google Scholar 

  7. Bulling A (2014) “33 A Tutorial on Human Activity Recognition Using Body-Worn Inertial Sensors,” dl.acm.org 46(3). https://doi.org/10.1145/2499621.

  8. Casilari E, Santoyo-Ramón J-A, Cano-García J-M (2017) “Analysis of Public Datasets for Wearable Fall Detection Systems,” mdpi.com. https://doi.org/10.3390/s17071513.

  9. Cook DJ, Krishnan NC (2020) “Activity Learning: Discovering, Recognizing, and Predicting Human Behavior. Google Books.” https://books.google.co.in/books?hl=en&lr=&id=TMZ9BgAAQBAJ&oi=fnd&pg=PR11&dq=1.%09Cook,+D.J.+and+Krishnan,+N.C.,+2015.+Activity+learning:+discovering,+recognizing,+and+predicting+human+behavior+from+sensor+data.+John+Wiley+%26+Sons&ots=9qtJ5vnknl&sig=q9o1. Accessed 01 Jul 2020

  10. Dietterich TG (2000) Ensemble methods in machine learning In: Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), vol. 1857 LNCS, pp. 1–15. https://doi.org/10.1007/3-540-45014-9_1.

  11. Ehatisham-ul-Haq M et al (2017) Authentication of smartphone users based on activity recognition and mobile sensing. Sensors 17(9):2043. https://doi.org/10.3390/s17092043

    Article  Google Scholar 

  12. Frigo M, Johnson SG (2020) “FFTW: An adaptive software architecture for the FFT.” [Online]. Available: https://ieeexplore.ieee.org/abstract/document/681704/. Accessed 01 Jul 2020

  13. Ghosh K, Ghosh S, Sen S, Sarkar R, Maulik U (2020) “A two-stage approach towards protein secondary structure classification,” Springer, Accessed 01 Jul 2020. [Online]. Available: https://doi.org/10.1007/s11517-020-02194-w.pdf.

  14. Kwapisz JR, Weiss GM, Moore SA (2011) Activity recognition using cell phone accelerometers. ACM SIGKDD Explor Newsl 12(2):74–82. https://doi.org/10.1145/1964897.1964918

    Article  Google Scholar 

  15. Maekawa T, Nakai D, Ohara K, Namioka Y (2016) Toward practical factory activity recognition: Unsupervised understanding of repetitive assembly work in a factory. In: UbiComp 2016 - Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing, pp. 1088–1099. https://doi.org/10.1145/2971648.2971721.

  16. Mi C, Xu R, Lin C-T (2019) “Real-time Recognition of Smartphone User Behavior Based on Prophet Algorithms,” no. 1, pp. 4–6. [Online]. Available: http://arxiv.org/abs/1909.08997. Accessed 01 Jul 2020

  17. Micucci DI, Mobilio MI, Napoletano P (2017) UniMiB SHAR: a dataset for human activity recognition using acceleration data from smartphones. Appl Sci 7:1101. https://doi.org/10.3390/app7101101

    Article  Google Scholar 

  18. Mubashir M, Shao L, Seed L (2013) A survey on fall detection: principles and approaches. Neurocomputing 100:144–152. https://doi.org/10.1016/j.neucom.2011.09.037

    Article  Google Scholar 

  19. Panhwar M, Shah S, Syed M (2017) Journal of, and undefined 2017, “Smart phone based fall detection using auto regression modeling in a non-restrictive setting,” safetylit.org. [Online]. Available: https://www.safetylit.org/citations/index.php?fuseaction=citations.viewdetails&citationIds[]=citjournalarticle_553203_7. Accessed 01 Jul 2020

  20. Quispe KM, Lima WS, DM Batista, Souto E (2018) MBOSS: A symbolic representation of human activity recognition using mobile sensors. Sensors 18(12): 4354. https://doi.org/10.3390/s18124354.

    Article  Google Scholar 

  21. Rish I (2020) An empirical study of the naive Bayes classifier. [Online]. Available: https://www.cc.gatech.edu/~isbell/reading/papers/Rish.pdf. Accessed 01 Jul 2020

  22. Rokach L, Maimon O (2005) Top-down induction of decision trees classifiers - a survey. IEEE Trans Syst Man Cybern Part C Appl Rev 35(4):476–487. https://doi.org/10.1109/TSMCC.2004.843247

    Article  Google Scholar 

  23. Ronneberger O, Fischer P, Brox T (2020) U-Net: Convolutional Networks for Biomedical Image Segmentation. [Online]. Available: http://lmb.informatik.uni-freiburg.de/. Accessed 01 Jul 2020

  24. Rosenblatt F 1958 Review and undefined 1958, “The perceptron: a probabilistic model for information storage and organization in the brain”. [Online]. Available: https://psycnet.apa.org/journals/rev/65/6/386/. Accessed 01 Jul 2020

  25. Ruck DW, Rogers SK, Kabrisky M (1990) Feature Selection Using a Multilayer Perceptron. [Online]. Available: https://www.researchgate.net/profile/Teresa_Hawkes/project/Neurophysiology-based-computer-program-design-for-cyber-security/attachment/58f16b381042bf333c68ad87/AS:483205733195776@1492216632627/download/Ruck%2C+Rogers%2C+Kabrisky.pdf?context=ProjectUpdatesLog. Accessed 01 Jul 2020

  26. Sadhukhan S, Mallick S, Singh PK, Sarkar R, Bhattacharjee D (2020) A comparative study of different feature descriptors for video-based human action recognition. Springer, Singapore, pp 35–52

    Google Scholar 

  27. Saunders C, Stitson MO, Weston J, Bottou L, Schölkopf B, Smola A (1998) Support Vector Machine Reference Manual. [Online]. Available: https://eprints.soton.ac.uk/258959/1/SVM_Reference.pdf. Accessed 01 Jul 2020

  28. Sazonov E, Metcalfe K, Lopez-Meyer P, Tiffany S (2011) RF hand gesture sensor for monitoring of cigarette smoking. In: Proceedings of the International Conference on Sensing Technology, ICST, pp. 426–430. https://doi.org/10.1109/ICSensT.2011.6137014.

  29. Shoaib M, Bosch S, Incel OD, Scholten H, Havinga PJM (2016) “Complex Human Activity Recognition Using Smartphone and Wrist-Worn Motion Sensors,” mdpi.com. https://doi.org/10.3390/s16040426.

  30. Vavoulas G, Chatzaki C, Malliotakis T, Pediaditis M, Tsiknakis M 92016) The MobiAct Dataset: Recognition of Activities of Daily Living using Smartphones

  31. Walse KH, Dharaskar RV (2016) Performance Evaluation of Classifiers on WISDM Dataset for Human Activity Recognition. Accessed 01 Jul 2020. [Online]. Available: https://doi.org/10.1145/2905055.2905232.

  32. Wang J, Chen Y, Hao S, Peng X, Hu L (2019) Deep learning for sensor-based activity recognition: a survey. Pattern Recogn Lett 119:3–11. https://doi.org/10.1016/j.patrec.2018.02.010

    Article  Google Scholar 

  33. Wiskott L, Sejnowski TJ (2002) Slow feature analysis: unsupervised learning of invariances. Neural Comput 14(4):715–770. https://doi.org/10.1162/089976602317318938

    Article  MATH  Google Scholar 

  34. Xu Y, Lu Y (2015) Neurocomputing, and undefined 2015, “Adaptive weighted fusion: a novel fusion approach for image classification,” Elsevier. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S0925231215007687. Accessed 01 Jul 2020

  35. Xu J, He Z, Zhang Y (2019) CNN-LSTM combined network for iot enabled fall detection applications. J Phys 12044. https://doi.org/10.1088/1742-6596/1267/1/012044.

  36. Zhang Y, Zhang Y, Zhang Z, Bao J, Song Y (2020) Human activity recognition based on time series analysis using U-Net. [Online]. Available: https://arxiv.org/abs/1809.08113. Accessed 01 Jul 2020

  37. Zheng Z, Du J, Sun L, Huo M, Chen Y (2018) “TASG: An Augmented Classification Method for Impersonal HAR,” hindawi.com. https://doi.org/10.1155/2018/6751363.

Download references

Acknowledgements

Authors are grateful to the Center for Microprocessor Application for Training Education and Research (CMATER) of Computer Science & Engineering Department, Jadavpur University, for providing infrastructure facilities during progress of the work.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Pawan Kumar Singh.

Ethics declarations

Conflict of interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Mukherjee, D., Mondal, R., Singh, P.K. et al. EnsemConvNet: a deep learning approach for human activity recognition using smartphone sensors for healthcare applications. Multimed Tools Appl 79, 31663–31690 (2020). https://doi.org/10.1007/s11042-020-09537-7

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-020-09537-7

Keywords

Navigation