Abstract
With the gradual improvement of the intelligent degree of smart-home devices, its popularity is also multiplying. These devices dramatically improve the richness of available data, including a large number of indoor family life visual data. At the same time, it also has higher requirements on effectively using these data and further improving the intelligence of smart-home devices. In this paper, we mainly verify the performance of frameworks and feasibility of deploying the human activity recognition models when the computing power of edge computing devices is limited. It includes typical deep learning methods, CNN and GCN-based recognition methods, and a single activity judgment recognition method based on CTW. We also analyze the help of different data preprocessing steps in improving time efficiency and accuracy. In addition, a human activity dataset is built based on the actual home fitness equipment. The experimental results verify the feasibility and effectiveness of deploying the activity recognition model on IoT devices with limited computing capability.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Artacho, B., Savakis, A.: UniPose: unified human pose estimation in single images and videos. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition pp. 7035–7044 (2020)
Baradel, F., Wolf, C., Mille, J.: Pose-conditioned spatio-temporal attention for human action recognition. arXiv preprint arXiv:1703.10106 (2017)
Berndt, D.J., Clifford, J.: Using dynamic time warping to find patterns in time series. In: KDD Workshop, pp. 359–370 (1994)
Bruna, J., Zaremba, W., Szlam, A., LeCun, Y.: Spectral networks and locally connected networks on graphs. arXiv preprint arXiv:1312.6203 (2013)
Cao, Z., Simon, T., Wei, S.E., Sheikh, Y.: Realtime multi-person 2D pose estimation using part affinity fields. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7291–7299 (2017)
Ji, S., Xu, W., Yang, M., Yu, K.: 3D convolutional neural networks for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 35(1), 221–231 (2013)
Kendall, A., Grimes, M., Cipolla, R.: PoseNet: a convolutional network for real-time 6-DOF camera relocalization. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2938–2946 (2015)
Li, Y., He, Z., Ye, X., He, Z., Han, K.: Spatial temporal graph convolutional networks for skeleton-based dynamic hand gesture recognition. EURASIP J. Image Video Process. 2019(1), 1–7 (2019). https://doi.org/10.1186/s13640-019-0476-x
Lugaresi, C., et al.: Mediapipe: A framework for building perception pipelines. arXiv preprint arXiv:1906.08172 (2019)
Luvizon, D.C., Picard, D., Tabia, H.: 2D/3D pose estimation and action recognition using multitask deep learning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5137–5146 (2018)
Lysenkov, I., Eruhimov, V., Bradski, G.: Recognition and pose estimation of rigid transparent objects with a kinect sensor. Robotics 273(273–280), 2 (2013)
Obdržálek, Å., et al.: Accuracy and robustness of kinect pose estimation in the context of coaching of elderly population. In: 2012 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, pp. 1188–1193 (2012)
Osokin, D.: Real-time 2D multi-person pose estimation on CPU: Lightweight openpose. arXiv preprint arXiv:1811.12004 (2018)
Peres-Neto, P.R., Jackson, D.A.: How well do multivariate data sets match? the advantages of a procrustean superimposition approach over the mantel test. Oecologia 129(2), 169–178 (2001). https://doi.org/10.1007/s004420100720
Sun, J.J., Zhao, J., Chen, L.-C., Schroff, F., Adam, H., Liu, T.: View-invariant probabilistic embedding for human pose. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020, Part V. LNCS, vol. 12350, pp. 53–70. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-58558-7_4
Toshev, A., Szegedy, C.: DeepPose: Human pose estimation via deep neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1653–1660 (2014)
Vrigkas, M., Nikou, C., Kakadiaris, I.A.: A review of human activity recognition methods. Front. Robot. AI 2, 28 (2015)
Yan, S., Xiong, Y., Lin, D.: Spatial temporal graph convolutional networks for skeleton-based action recognition. In: Thirty-second AAAI Conference on Artificial Intelligence, pp. 1–9 (2018)
Zhou, F., Torre, F.: Canonical time warping for alignment of human behavior. In: Bengio, Y., Schuurmans, D., Lafferty, J., Williams, C., Culotta, A. (eds.) Advances in Neural Information Processing Systems, vol. 22, pp. 2286–2294. Curran Associates, Inc. (2009)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
He, T. (2022). Human Pose-Based Activity Recognition Approaches on Smart-Home Devices. In: Streitz, N.A., Konomi, S. (eds) Distributed, Ambient and Pervasive Interactions. Smart Environments, Ecosystems, and Cities. HCII 2022. Lecture Notes in Computer Science, vol 13325. Springer, Cham. https://doi.org/10.1007/978-3-031-05463-1_19
Download citation
DOI: https://doi.org/10.1007/978-3-031-05463-1_19
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-05462-4
Online ISBN: 978-3-031-05463-1
eBook Packages: Computer ScienceComputer Science (R0)