Abstract
Recognizing a person’s physical activity with certainty makes an important aspect of intelligent computing. Modern smart devices are equipped with powerful sensors that are suitable for sensor-based human activity recognition (AR) task. Traditional approaches to human activity recognition has made significant progress but most of those methods rely upon manual feature extraction. The design and selection of relevant features is the most challenging task in sensor-based human AR problem. Using manually extracted features for this task hinders the generalization of performance and these handcrafted features are also incapable of handling similar and complex activities with certainty. In this paper, we propose a deep learning based method for human activity recognition problem. The method uses convolutional neural networks to automatically extract features from raw sensor data and classify six basic human activities. Furthermore, transfer learning is used to reduce the computational cost involved in training the model from scratch for a new user. The model uses the labelled information from supervised learning, to mutually enhance the feature extraction and classification. Experiments carried on benchmark dataset verified the strong advantage of proposed method over the traditional human AR algorithms such as Random Forest (RF) and multiclass Support Vector Machine (SVM).







Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Notes
AIM ’94 (1994). Diabetes Data Set [online]. Website https://archive.ics.uci.edu/ml/datasets/Diabetes [accessed 22 November 2018].
References
AM (2013) Recognizing physical activities using wii remote. Int J Inf Educ Technol 3(1):60
Angelov PP, Zhou X (2008) Evolving fuzzy-rule-based classifiers from data streams. IEEE Trans Fuzzy Syst 16(6):1462–1475
Angelov P, Filev DP, Kasabov N (2010) Evolving intelligent systems: methodology and applications. Wiley, Hoboken
Bagct U, Bai L (2007)A comparison of daubechies and gabor wavelets for classification of MR images. In: 2007 IEEE International Conference on Signal Processing and Communications. IEEE, 2007
Bao L, Intille SS (2004) Activity recognition from user-annotated acceleration data. In: International Conference on Pervasive Computing. Springer, pp 1–17
Bengio Y (2009) Learning deep architectures for ai. Found Trends® Mach Learn 2(1):1–127
Bhattacharya S, Nurmi P, Hammerla N, Plötz T (2014) Using unlabeled data in a sparse-coding framework for human activity recognition. Pervasive Mob Comput 15:242–262
Brezmes T, Gorricho J-L, Cotrina J (2009) Activity recognition from accelerometer data on a mobile phone. In: International Work-Conference on Artificial Neural Networks. Springer, pp 796–799
Chennuru S, Chen P-W, Zhu J, Zhang JY (2010) Mobile lifelogger–recording, indexing, and understanding a mobile user’s life. In: International Conference on Mobile Computing, Applications, and Services. Springer, pp 263–281
Choudhary P, Hazra A (2019) Chest disease radiography in twofold: using convolutional neural networks and transfer learning. Evol Syst 1–13
Figo D, Diniz PC, Ferreira DR, Cardoso JM (2010) Preprocessing techniques for context recognition from accelerometer data. Pers Ubiquit Comput 14(7):645–662
Huynh T, Schiele B (2005) Analyzing features for activity recognition. In: Proceedings of the 2005 joint conference on Smart objects and ambient intelligence innovative context-aware services: usages and technologies-sOc-EUSAI 05. ACM Press
Krause A, Siewiorek D, Smailagic A, Farringdon J (2003) Unsupervised, dynamic identification of physiological and activity context in wearable computing. In: Seventh IEEE International Symposium on Wearable Computers, 2003. Proceedings. IEEE
Krizhevsky A, Sutskever I, Hinton GE (2017) ImageNet classification with deep convolutional neural networks. Commun ACM 60(6):84–90
Kwapisz JR, Weiss GM, Moore SA (2011) Activity recognition using cell phone accelerometers. ACM SigKDD Explor Newslett 12(2):74–82
Lowe D (1999) Object recognition from local scale-invariant features. In: Proceedings of the Seventh IEEE International Conference on Computer Vision. IEEE, 1999
Ordonez FJ, Roggen D (2016) Deep convolutional and lstm recurrent neural networks for multimodal wearable activity recognition. Sensors 16(1):115
Peterek T, Penhaker M, Gajdoš P, Dohnálek P (2014) Comparison of classification algorithms for physical activity recognition. In: Innovations in bio-inspired computing and applications. Springer, pp 123–131
Plotz T, Hammerla NY, Olivier P (2011) Feature learning for activity recognition in ubiquitous computing. In: IJCAI Proceedings-International Joint Conference on Artificial Intelligence, vol. 22, no. 1, 2011, p 1729
Reddy S, Mun M, Burke J, Estrin D, Hansen M, Srivastava M (2010) Using mobile phones to determine transportation modes. ACM Trans Sens Netw (TOSN) 6(2):13
Rokni SA, Ghasemzadeh H (2017) Synchronous dynamic view learning: a framework for autonomous training of activity recognition models using wearable sensors. In: Proceedings of the 16th ACM-IEEE International Conference on Information Processing in Sensor Networks - IPSN. ACM Press
Ronao CA, Cho S-B (2016) Human activity recognition with smartphone sensors using deep learning neural networks. Expert Syst Appl 59:235–244
Ronaoo CA, Cho S-B (2015) Evaluation of deep convolutional neural network architectures for human activity recognition with smartphone sensors, pp 858–860
Sharma A, Lee Y-D, Chung W-Y (2008) High accuracy human activity monitoring using neural network. In: Convergence and Hybrid Information Technology, 2008. ICCIT’08. Third International Conference on, vol. 1. IEEE, 2008, pp 430–435
Tang Y, Salakhutdinov R, Hinton G (2012) Robust boltzmann machines for recognition and denoising. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition. IEEE, jun 2012
Thomas Platz NYH, Olivier P (2012) Feature learning for activity recognition in ubiquitous computing. IJCAI
Vollmer C, Gross H-M, Eggert JP (2013) Learning features for activity recognition with shift-invariant sparse coding. In: International conference on artificial neural networks. Springer, pp 367–374
Wang Z, Chen Y (2017) Recognizing human concurrent activities using wearable sensors: a statistical modeling approach based on parallel hmm. Sens Rev 37(3):330–337
Wu P, Zhu J, Zhang JY (2013) Mobisens: a versatile mobile sensing platform for real-world applications. Mob Netw Appl 18(1):60–80
Wu P, Peng H-K, Zhu J, Zhang Y (2011) Senscare: Semi-automatic activity summarization system for elderly care. In: International Conference on Mobile Computing, Applications, and Services. Springer, pp 1–19
Yang J, Nguyen MN, San PP, Li X, Krishnaswamy S (2015) Deep convolutional neural networks on multichannel time series for human activity recognition. Ijcai 15(2015):3995–4001
Yuan G, Wang Z, Meng F, Yan Q, Xia S (2018) An overview of human activity recognition based on smartphone. Sens Rev
Funding
This research did not receive any specific grant from funding agencies in the public, commercial, or not-for-profit sectors.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The Authors declare that there is no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Bhat, O., Khan, D.A. Evaluation of deep learning model for human activity recognition. Evolving Systems 13, 159–168 (2022). https://doi.org/10.1007/s12530-021-09373-6
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s12530-021-09373-6