ABSTRACT
Benefitting from the development of pervasive computing, recent years have witnessed a variety of meaningful human-centric applications, where automating the recognition of human activities plays a central role in bridging the gap between sensing data and high-level services. Accelerometer-based activity recognizer often remains a priority due to its recognition performance, low costs, and portability, however, few studies systematically investigate how to extract and use features from the time-series sensor data and further compare their discriminant power. To this end, we herein propose two different ways of extracting features and exploring their combinations. Specifically, we take as a resultant axis or separate channels the accelerometer axes and then extract axes-resultant and axis-wise features. Afterwards, we evaluate the cases where the two feature sets are used separately or jointly. Finally, we conduct comparative experiments on two public activity recognition datasets with five different classification models in terms of four performance metrics. Results show that the use of axis-wise features outperforms its competitor in the majority across the datasets and that their joint use generally leads to enhanced accuracy.
- A Haque, A Milstein and L Fei-Fei (2020). Illuminating the dark spaces of healthcare with ambient intelligence. Nature, 585(7824), 193-202.Google ScholarCross Ref
- A Wang, S Zhao, C Zheng, J Yang, G Chen and CY Chang (2021). Activities of daily living recognition with binary environment sensors using deep learning: A comparative study. IEEE Sensors Journal, 21(4), 5423-5433.Google ScholarCross Ref
- A Bulling, U Blanke and B. Schiele (2014). A tutorial on human activity recognition using body-worn inertial sensors. ACM Computing Surveys, 46(3), 1-33.Google ScholarDigital Library
- A Wang, G Chen, J Yang, S Zhao and CY Chang (2016). A comparative study on human activity recognition using inertial sensors in a smartphone. IEEE Sensors Journal,16(11), 4566-4578.Google ScholarCross Ref
- X Zhou, W Liang, I Kevin, K Wang, H Wang, LT Yang and Q Jin (2020). Deep-learning-enhanced human activity recognition for Internet of healthcare things. IEEE Internet of Things Journal, 7(7), 6429-6438.Google ScholarCross Ref
- A Wang, G Chen, X Wu, L Liu, N An and CY Chang (2018). Towards human activity recognition: A hierarchical feature selection framework. Sensors, 18(11), 3629.Google ScholarCross Ref
- J Kwapisz, G Weiss and S Moore (2011). Activity recognition using cell phone accelerometers. ACM SigKDD Explorations Newsletter, 12(2), 74-82.Google ScholarDigital Library
- P Zappi, C Lombriser, T Stiefmeier, E Farella, D Roggen, L Benini and G Tröster. Activity recognition from on-body sensors: Accuracy-power trade-off by dynamic sensor selection. Proceeding of European Conference on Wireless Sensor Networks, 2008, pp. 17-33.Google ScholarCross Ref
- S Rosati, G Balestra and M Knaflitz (2018). Comparison of different sets of features for human activity recognition by wearable sensors. Sensors, 18(12), 4189.Google ScholarCross Ref
- A Ignatov (2018). Real-time human activity recognition from accelerometer data using convolutional neural networks. Applied Soft Computing, 62, 915-922.Google ScholarCross Ref
- C Lübbe, B Friedrich, S Fudickar, S Hellmers and A Hein. Feature based random forest nurse care activity recognition using accelerometer data. Adjunct Proceedings of the 2020 ACM International Joint Conference on Pervasive and Ubiquitous Computing and Proceedings of the 2020 ACM International Symposium on Wearable Computers, 2020, pp. 408-413.Google ScholarDigital Library
- D Anguita, A Ghio, L Oneto, X Parra and JL Reyes-Ortiz. A public domain dataset for human activity recognition using smartphones. Proceedings of 21st European Symposium on Artificial Neural Networks, 2013, pp. 437–442.Google Scholar
- A Wang, S Zhao, C Zheng, H Chen, L Liu and G Chen (2021). HierHAR: Sensor-based data-driven hierarchical human activity recognition, IEEE Sensors Journal, 21(3), 3353-3365.Google ScholarCross Ref
- A Wang, X Wu, L Zhao, H Chen, S Zhao (2021). Physical activity recognition from accelerometer data using multi-view aggregation. Journal of Applied Science and Engineering, 24(4), 611-620.Google Scholar
- C Ronao and S Cho (2016). Human activity recognition with smartphone sensors using deep learning neural networks. Expert Systems with Applications, 59, 235-244.Google ScholarDigital Library
- H Ma, W Li, X Zhang, S Gao and S Lu. AttnSense: Multi-level attention mechanism for multimodal human activity recognition. International Joint Conference on Artificial Intelligence, 2019, pp. 3109-3115.Google ScholarCross Ref
Index Terms
- Human Activity Recognition from Accelerometer Data: Axis-Wise Versus Axes-Resultant Feature Extraction
Recommendations
Multi-view stacking for activity recognition with sound and accelerometer data
A multi-view stacking based method for sensor fusion in activity recognition.Home-tasks activities recognition from accelerometer and audio signals.Combining heterogeneous wearable sensors using stacked generalization. Many Ambient Intelligence (AmI) ...
Physical activity recognition using a single triaxial accelerometer and a barometric sensor for baby and child care in a home environment
In this paper, we describe and evaluate an activity recognition system using a single 3-axis accelerometer and a barometric sensor worn on a waist of the body. The purpose of this work is to prevent child accidents such as unintentional injuries at ...
Artificial intelligence in human activity recognition: a review
The various activities of human movements have been discussed for several years, such as sports activities, daily life activities, and so on. Their detection and classification have given crucial information about a person's behaviour and health status. ...
Comments