Skip to main content

Advertisement

Log in

Sparse representation based classification scheme for human activity recognition using smartphones

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

The availability of built-in sensors in mobile devices have paved a way for researchers to accurately determine human activities through these sensors. In this paper, we present a novel action recognition system based on sparse representation wherein, eight different human activities were classified. Our proposed classifier employs data of accelerometer, gyroscope, magnetometer and orientation sensor equipped in smartphones for recognizing human activities. Time-domain and frequency-domain features are derived from the acquired sensor data. We have introduced a novel algorithm for fusing the data from the four sensors using a sparse representation based technique that aid in achieving the best classification performance. In the proposed algorithm, if the majority of the sensors indicate a particular class as the output, then that specific class is assigned as the actual test class. However, if there is a disagreement between the classified output of different sensors, then a novel weighted fusion scheme is introduced to fuse the scores and the residue produced by different sensors. The weight used in fusion is chosen to be the standard deviation of the score vector. Thus, the features of excellent sensors are made to bestow more on to the result of action recognition. Finally, the action label is recognized based on an activity metric that maximizes the score while minimizing the residue. The performance analysis of the proposed system is performed using leave-one-subject-out approach. Performance evaluation metrics like recall, precision, specificity, F-score and accuracy are utilized in projecting the performance of the proposed system. It was shown that the proposed system attained a high overall accuracy of about 97.13%.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4

Similar content being viewed by others

References

  1. Aggarwal JK, Xia L (2014) Human activity recognition from 3d data: a review. Pattern Recogn Lett 48:70–80

    Article  Google Scholar 

  2. Akhavian R, Behzadan AH (2016) Smartphone-based construction workers' activity recognition and classification. Autom Constr 71:198–209

    Article  Google Scholar 

  3. Alam MA, Wang W, Ahamed SI, Chu W (2013) Elderly safety: a smartphone based real time approach. In: Proceedings of the international conference on smart homes and health telematics. Springer, Berlin, pp 134–142

    Google Scholar 

  4. Altun K, Barshan B, Tuncel O (2010) Comparative study on classifying human activities with miniature inertial and magnetic sensors. Pattern Recogn 43(10):3605–3620

    Article  Google Scholar 

  5. Atallah L, Lo B, King R, Yang GZ (2011) Sensor positioning for activity recognition using wearable accelerometers. IEEE T Biomed Circ Syst 5(4):320–329

    Article  Google Scholar 

  6. Bao L, Intille S (2004) Activity recognition from user-annotated acceleration data. In: Ferscha A, Mattern F (eds) Pervasive computing. Lecture notes in computer science, vol 3001. Springer Berlin / Heidelberg, pp 1–17

  7. Baraniuk RG, Candes E, Elad M, Ma Y (2010) Applications of sparse representation and compressive sensing [scanning the issue]. Proc IEEE 98(6):906–909

    Article  Google Scholar 

  8. Bruckstein AM, Donoho DL, Elad M (2007) From sparse solutions of systems of equations to sparse modeling of signals and images. SIAM Rev 51(1):34–81

    Article  MathSciNet  Google Scholar 

  9. Catal C, Tufekci S, Pirmit E, Kocabag G (2015) On the use of ensemble of classifiers for accelerometer-based activity recognition. Appl Soft Comput 37:1018–1022

    Article  Google Scholar 

  10. Chen Z, Huang W, Lv Z (2017) Towards a face recognition method based on uncorrelated discriminant sparse preserving projection. Multimed Tools Appl 76(17):17669–17683

    Article  Google Scholar 

  11. Cheng H, Liu Z, Zhao Y, Ye G, Sun X (2014) Real world activity summary for senior home monitoring. Multimed Tools Appl 70(1):177–197

    Article  Google Scholar 

  12. Figueiredo IN, Leal C, Pinto L, Bolito J, Lemos A (2016) Exploring smartphone sensors for fall detection. mUX: J Mob User Exp 5(1):2

    Google Scholar 

  13. Fuentes D, Gonzalez-Abril L, Angulo C, Ortega JA (2012) Online motion recognition using an accelerometer in a mobile device. Expert Syst Appl 39(3):2461–2465

    Article  Google Scholar 

  14. Gao L, Bourke AK, Nelson J (2014) Evaluation of accelerometer based multi-sensor versus single-sensor activity recognition systems. Med Eng Phys 36(6):779–785

    Article  Google Scholar 

  15. Gravina R, Alinia P, Ghasemzadeh H, Fortino G (2017) Multi-sensor fusion in body sensor networks: state-of-the-art and research challenges. Inform Fusion 35:68–80

    Article  Google Scholar 

  16. Guha T, Ward RK (2012) Learning sparse representations for human action recognition. IEEE Trans Pattern Anal Mach Intell 34(8):1576–1588

    Article  Google Scholar 

  17. Huynh QT, Nguyen UD, Irazabal LB, Ghassemian N, Tran BQ (2015) Optimization of an accelerometer and gyroscope-based fall detection algorithm. J Sensors 452078

  18. Ignatov AD, Strijov VV (2016) Human activity recognition using quasiperiodic time series collected from a single tri-axial accelerometer. Multimed Tools Appl 75(12):7257–7270

    Article  Google Scholar 

  19. Jain A, Kanhangad V (2015) Exploring orientation and accelerometer sensor data for personal authentication in smartphones using touchscreen gestures. Pattern Recogn Lett 68:351–360

    Article  Google Scholar 

  20. Jansi R, Amutha R (2018) A novel chaotic map based compressive classification scheme for human activity recognition using a tri-axial accelerometer. Multimed Tools Appl 2018:1–20

    Google Scholar 

  21. Jansi R, Amutha R, Gokulakrishnan A (2017) A novel framework for action recognition based on histogram of oriented gradients and sparsity-inducing shared dictionary. In: Proceedings of the IEEE International Conference on Wireless Communications, Signal Processing and Networking (WiSPNET), 493–497

  22. Lara OD, Labrador MA (2013) A survey on human activity recognition using wearable sensors. IEEE Commun Surv Tut 15(3):1192–1209

    Article  Google Scholar 

  23. Lee SW, Mase K (2002) Activity and location recognition using wearable sensors. IEEE Pervas Comput 1(3):24–32

    Article  Google Scholar 

  24. Liu Y, Nie L, Liu L (2016) Rosenblum DS. From action to activity: sensor-based activity recognition. Neurocomputing 181:108–115

    Article  Google Scholar 

  25. Lu Y, Wei Y, Liu L, Zhong J, Sun L, Liu Y (2017) Towards unsupervised physical activity recognition using smartphone accelerometers. Multimed Tools Appl 76(8):10701–10719

    Article  Google Scholar 

  26. Luo X, Guan Q, Tan H, Gao L, Wang Z, Luo X (2017) Simultaneous indoor tracking and activity recognition using pyroelectric infrared sensors. Sensors 17(8):1738

    Article  Google Scholar 

  27. Miao F, Cheng Y, He Y, He Q, Li Y (2015) A wearable context-aware ECG monitoring system integrated with built-in kinematic sensors of the smartphone. Sensors 15(5):11465–11484

    Article  Google Scholar 

  28. Mitchell E, Monaghan D, O'Connor NE (2013) Classification of sporting activities using smartphone accelerometers. Sensors 13(4):5317–5337

    Article  Google Scholar 

  29. Moonon AU, Hu J, Li S (2015) Remote sensing image fusion method based on nonsubsampled shearlet transform and sparse representation. Sens Imaging 16(1):23

    Article  Google Scholar 

  30. Ordóñez FJ, de Toledo P, Sanchis A (2015) Sensor-based Bayesian detection of anomalous living patterns in a home setting. Pers Ubiquit Comput 19(2):259–270

  31. De Pessemier T, Dooms S, Martens L (2014) Context-aware recommendations through context and activity recognition in a mobile environment. Multimed Tools Appl 72(3):2925–2948

  32. Preece SJ, Goulermas JY, Kenney LP, Howard D (2009) A comparison of feature extraction methods for the classification of dynamic activities from accelerometer data. IEEE Trans Biomed Eng 56(3):871–879

    Article  Google Scholar 

  33. Ronao CA, Cho SB (2016) Human activity recognition with smartphone sensors using deep learning neural networks. Expert Syst Appl 59:235–244

    Article  Google Scholar 

  34. Roy D, Srinivas M, Mohan CK (2016) Sparsity-inducing dictionaries for effective action classification. Pattern Recogn 59:55–62

    Article  Google Scholar 

  35. Seo JJ, Kim HI, De Neve W, Ro YM (2017) Effective and efficient human action recognition using dynamic frame skipping and trajectory rejection. Image Vis Comput 58:76–85

    Article  Google Scholar 

  36. Tapia EM, Intille SS, Larson K (2004) Activity recognition in the home using simple and ubiquitous sensors. In: Pervasive computing. Springer, Berlin, pp 158–175

    Chapter  Google Scholar 

  37. Varkey JP, Pompili D, Walls TA (2012) Human motion recognition using a wireless sensor-based wearable system. Pers Ubiquit Comput 6(7):897–910

    Article  Google Scholar 

  38. Villalonga C, Pomares H, Rojas I, Banos O (2017) MIMU-Wear: ontology-based sensor selection for real-world wearable activity recognition. Neurocomputing 250:76–100

    Article  Google Scholar 

  39. Wang S (2015) Zhou G. a review on radio based activity recognition. Dig Commun Netw 1(1):20–29

    Article  Google Scholar 

  40. Wang A, Chen G, Yang J, Zhao S, Chang CY (2016) A comparative study on human activity recognition using inertial sensors in a smartphone. IEEE Sensors J 16(11):4566–4578

    Article  Google Scholar 

  41. Xiao L, Li R, Luo J, Xiao Z (2016) Energy-efficient recognition of human activity in body sensor networks via compressed classification. Int J Distrib Sens N 12(12):1–8

    Google Scholar 

  42. Xing X, Qiu F, Xu X, Qing C, Wu Y (2017) Robust object tracking based on sparse representation and incremental weighted PCA. Multimed Tools Appl 76(2):2039–2057

    Article  Google Scholar 

  43. Xu Y, Lu Y (2015) Adaptive weighted fusion: a novel fusion approach for image classification. Neurocomputing 168:566–574

    Article  Google Scholar 

  44. Xu Y, Zhu Q, Zhang D (2011) Combine crossing matching scores with conventional matching scores for bimodal biometrics and face and palmprint recognition experiments. Neurocomputing 74(18):3946–3952

    Article  Google Scholar 

  45. Yang AY, Jafari R, Sastry SS, Bajcsy R (2009) Distributed recognition of human actions using wearable motion sensor networks. J Amb Intel Smart En 1(2):103–115

    Google Scholar 

  46. Yao T, Wang Z, Xie Z, Gao J, Feng DD (2017) Learning universal multiview dictionary for human action recognition. Pattern Recogn 64:236–244

    Article  Google Scholar 

  47. Yin L, Dong M, Duan Y, Deng W, Zhao K, Guo J (2014) A high-performance training-free approach for hand gesture recognition with accelerometer. Multimed Tools Appl 72(1):843–864

    Article  Google Scholar 

  48. Yun X, Calusdian J, Bachmann ER, McGhee RB (2012) Estimation of human foot motion during normal walking using inertial and magnetic sensor measurements. IEEE Trans Instrum Meas 61(7):2059–2072

    Article  Google Scholar 

  49. Zhang M, Sawchuk AA (2013) Human daily activity recognition with sparse representation using wearable sensors. IEEE J Biomed Health Inform 17(3):553–556

    Article  Google Scholar 

  50. Zhang Z, Xu Y, Yang J, Li X, Zhang D (2015) A survey of sparse representation: algorithms and applications. IEEE Access 3:490–530

    Article  Google Scholar 

  51. Zhao X, Zhang S, Lei B (2014) Robust emotion recognition in noisy speech via sparse representation. Neural Comput & Applic 24(7–8):1539–1553

    Article  Google Scholar 

  52. Zualkernan I, Aloul F, Shapsough S, Hesham A, El-Khorzaty Y (2017) Emotion recognition using mobile phones. Comput Electr Eng 60:1–13

    Article  Google Scholar 

Download references

Acknowledgements

The authors wish to thank all the volunteers who contributed towards data collection.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to R. Jansi.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Jansi, R., Amutha, R. Sparse representation based classification scheme for human activity recognition using smartphones. Multimed Tools Appl 78, 11027–11045 (2019). https://doi.org/10.1007/s11042-018-6662-5

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-018-6662-5

Keywords

Navigation