ABSTRACT
For human activity recognition based on phone sensors, the position of the phone is an important factor of the recognition accuracy. To improve the recognition accuracy of behavioral activities and the position of the phone placed, this paper proposes a classification recognition algorithm based on accelerometer and gyroscope sensors. First, sensor data collected from seven different body positions are used as inputs to a deep stacked bidirectional long and short-term memory neural network; then the activity type and the phone position are used as labels to train the neural network for simultaneous recognition of human activity and phone position; finally, the performance of the proposed method is evaluated by cross-validation. The experimental results show that placing the phone on the waist and thigh achieves the highest recognition accuracy rate. The accuracy of the simultaneous recognition of activity and position is over 90%, which is 18% higher than existing algorithms.
- K. Chen, L. Yao, D. Zhang, X. Wang, X. Chang and F. Nie, "A Semisupervised Recurrent Convolutional Attention Model for Human Activity Recognition," in IEEE Transactions on Neural Networks and Learning Systems, vol. 31, no. 5, pp. 1747-1756, May 2020.Google ScholarCross Ref
- I. Bisio, A. Delfino, F. Lavagetto and A. Sciarrone, "Enabling IoT for in-home rehabilitation: Accelerometer signals classification methods for activity and movement recognition", IEEE Internet Things J., vol. 4, no. 1, pp. 135-146, Feb. 2017.Google ScholarCross Ref
- J. Huang, S. Lin, N. Wang, G. Dai, Y. Xie and J. Zhou, "TSE-CNN: A Two-Stage End-to-End CNN for Human Activity Recognition," in IEEE Journal of Biomedical and Health Informatics, vol. 24, no. 1, pp. 292-299, Jan. 2020.Google ScholarCross Ref
- H. Bi, M. Perello-Nieto, R. Santos-Rodriguez and P. Flach, "Human Activity Recognition Based on Dynamic Active Learning," in IEEE Journal of Biomedical and Health Informatics, vol. 25, no. 4, pp. 922-934, April 2021.Google ScholarCross Ref
- G. De Leonardis , "Human Activity Recognition by Wearable Sensors: Comparison of different classifiers for real-time applications," 2018 IEEE International Symposium on Medical Measurements and Applications (MeMeA), Rome, Italy, 2018, pp. 1-6.Google Scholar
- M. Kheirkhahan, S. Nair, A. Davoudi, P. Rashidi, A. A. Wanigatunga, D. B. Corbett, , "A smartwatch-based framework for real-time and online assessment and mobility monitoring", J. Biomed. Informat., vol. 89, pp. 29-40, Jan. 2019.Google ScholarCross Ref
- C. -T. Yen, J. -X. Liao and Y. -K. Huang, "Human Daily Activity Recognition Performed Using Wearable Inertial Sensors Combined With Deep Learning Algorithms," in IEEE Access, vol. 8, pp. 174105-174114, 2020.Google ScholarCross Ref
- X. Wang , "PerRNN: Personalized Recurrent Neural Networks for Acceleration-Based Human Activity Recognition," ICC 2019 - 2019 IEEE International Conference on Communications (ICC), Shanghai, China, pp. 1-6, 2019.Google Scholar
- M. Abdel-Basset, H. Hawash, R. K. Chakrabortty, M. Ryan, M. Elhoseny and H. Song, "ST-DeepHAR: Deep Learning Model for Human Activity Recognition in IoHT Applications, " in IEEE Internet of Things Journal, vol. 8, no. 6, pp. 4969-4979, 15 March15, 2021.Google ScholarCross Ref
- A. Shrestha, H. Li, J. Le Kernec and F. Fioranelli, "Continuous Human Activity Classification From FMCW Radar With Bi-LSTM Networks," in IEEE Sensors Journal, vol. 20, no. 22, pp. 13607-13619, 15 Nov.15, 2020.Google ScholarCross Ref
- A. Alajaji, W. Gerych, K. Chandrasekaran, L. Buquicchio, E. Agu and E. Rundensteiner, "DeepContext: Parameterized Compatibility-Based Attention CNN for Human Context Recognition, " 2020 IEEE 14th International Conference on Semantic Computing (ICSC), San Diego, CA, USA, 2020, pp. 53-60.Google Scholar
- S. W. Pienaar and R. Malekian, "Human activity recognition using LSTM-RNN deep neural network architecture", Proc. IEEE 2nd Wireless Afr. Conf. (WAC), pp. 1-5, Aug. 2019.Google ScholarCross Ref
- S. Wang and X. Zhu, "A Hybrid Deep Neural Networks for Sensor-based Human Activity Recognition," 2020 12th International Conference on Advanced Computational Intelligence (ICACI), Dali, China, 2020, pp. 486-491.Google Scholar
- D. Coskun, O. D. Incel and A. Ozgovde, "Phone location/placement detection using accelerometer: Impact on activity recognition", Proc. IEEE 10th Int. Conf. Intell. Sensors Sensor Netw. Inf. Process. (ISSNIP), pp. 1-6, Apr. 2015.Google Scholar
- M. Ehatisham-Ul-Haq, M. A. Azam, Y. Amin and U. Naeem, "C2FHAR: Coarse-to-Fine Human Activity Recognition With Behavioral Context Modeling Using Smart Inertial Sensors," in IEEE Access, vol. 8, pp. 7731-7747, 2020.Google ScholarCross Ref
- Hnoohom N, Ji Tp A Tt Anakul A, Inluergsri P, Multi-sensor-based fall detection and activity daily living classification by using ensemble learning[C]// 2018:111-115.Google Scholar
- I. A. Lawal and S. Bano, "Deep Human Activity Recognition With Localisation of Wearable Sensors, " in IEEE Access, vol. 8, pp. 155060-155070, 2020.Google ScholarCross Ref
- T. Sztyler, H. Stuckenschmidt and W. Petrich, " Position-aware activity recognition with wearable devices", Pervasive Mobile Comput., vol. 38, pp. 281-295, Jul. 2017.Google ScholarCross Ref
- A. Aich, A. Krishna, V. Akhilesh and C. Hegde, "Encoding Web-based Data for Efficient Storage in Machine Learning Applications," 2019 Fifteenth International Conference on Information Processing (ICINPRO), 2019, pp. 1-6.Google Scholar
- A. Shrestha, H. Li, J. Le Kernec and F. Fioranelli, "Continuous Human Activity Classification From FMCW Radar With Bi-LSTM Networks," in IEEE Sensors Journal, vol. 20, no. 22, pp. 13607-13619, 15 Nov.15, 2020.Google ScholarCross Ref
- W. Qi, H. Su and A. Aliverti, "A Smartphone-Based Adaptive Recognition and Real-Time Monitoring System for Human Activities," in IEEE Transactivities on Human-Machine Systems, vol. 50, no. 5, pp. 414-423, Oct. 2020.Google ScholarCross Ref
- Simultaneous Recognition Algorithm of Human Activity and Phone Position Based on Multi-sensor Data Fusion
Recommendations
Human activity recognition based on transformed accelerometer data from a mobile phone
Mobile phones are equipped with a rich set of sensors, such as accelerometers, magnetometers, gyroscopes, photometers, orientation sensors, and gravity sensors. These sensors can be used for human activity recognition in the ubiquitous computing domain. ...
Human Activity Recognition Using Multi-modal Data Fusion
Progress in Pattern Recognition, Image Analysis, Computer Vision, and ApplicationsAbstractThe automated recognition of human activity is an important computer vision task, and it has been the subject of an increasing number of interesting home, sports, security, and industrial applications. Approaches using a single sensor have ...
Human activity recognition using deep transfer learning of cross position sensor based on vertical distribution of data
AbstractSensor-based human activity recognition and health monitoring are attaining great interest in the eye of the researcher as it maintained the privacy of an individual. A model based on the transfer learning for the vertical distribution of cross ...
Comments