Abstract
Human pose estimation has gained significant attention from researchers of the present era. Personal exercise sessions can be monitored and supervised with the help of pose recognition. Existing work on exercise classification primarily relies on external or wearable sensors for recognizing poses. However, such sensors often fail to differentiate amongst similar exercises. Some essential extensions of human pose estimation are activity detection and activity prediction. In this paper, we first classify an individual’s exercises and then predict whether the pose corresponding to an exercise is correct or not. The tasks mentioned above are performed with the help of 2-dimensional pose coordinates. We have used an RGB camera to capture the poses during exercises performed by individuals. We formulate our model with 2D coordinates obtained from the 2D pose. We consider 2D coordinates of 18 joints of a human body as the primary features to classify different exercises and predict correctness about the poses. We have developed a benchmark dataset consisting of human subjects of various age groups with varying heights. An accuracy of 97.01% has been obtained, and it is better than existing work when tested on our dataset.
Similar content being viewed by others
References
Andriluka M, Pishchulin L, Gehler P, Schiele B (2014) 2d human pose estimation: New benchmark and state of the art analysis. In: Proceedings of the IEEE Conference on computer Vision and Pattern Recognition, pp 3686–3693
Cao Z, Hidalgo G, Simon T, Wei S-E, Sheikh Y (2019) Openpose: realtime multi-person 2d pose estimation using part affinity fields. IEEE Trans Pattern Anal Mach Intell 43(1):172–186
Cao Z, Simon T, Wei S-E, Sheikh Y (2017) Realtime multi-person 2d pose estimation using part affinity fields. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 7291–7299
Dhar P, Alvarez J M, Roy P P (2017) Efficient framework for action recognition using reduced fisher vector encoding. In: Proceedings of International Conference on Computer Vision and Image Processing, pp 343–354
Feichtenhofer C, Pinz A, Wildes R P (2017) Spatiotemporal multiplier networks for video action recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4768–4777
Hidalgo G, Raaj Y, Idrees H, Xiang D, Joo H, Simon T, Sheikh Y (2019) Single-network whole-body pose estimation. In: Proceedings of the IEEE International Conference on Computer Vision, pp 6982–6991
Hu J-F, Zheng W-S, Lai J, Gong S, Xiang T (2015) Exemplar-based recognition of human–object interactions. IEEE Trans Circ Syst Video Technol 26(4):647–660
Huang G, Liu Z, Van Der Maaten L, Weinberger K Q (2017) Densely connected convolutional networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4700–4708
Hwang D-H, Kim S, Monet N, Koike H, Bae S (2020) Lightweight 3d human pose estimation network training using teacher-student learning. In: Proceedings of the IEEE Winter Conference on Applications of Computer Vision, pp 479–488
Ke S-R, Zhu L, Hwang J-N, Pai H-I, Lan K-M, Liao C-P (2010) Real-time 3d human pose estimation from monocular view with applications to event detection and video gaming. In: Proceedings of the IEEE International Conference on Advanced Video and Signal Based Surveillance, pp 489–496
Koskimäki H, Siirtola P (2014) Recognizing gym exercises using acceleration data from wearable sensors. In: Proceedings of the IEEE Symposium on Computational Intelligence and Data Mining (CIDM), pp 321–328
Kumar P, Mukherjee S, Saini R, Kaushik P, Roy P P, Dogra D P (2018) Multimodal gait recognition with inertial sensor data and video using evolutionary algorithm. IEEE Trans Fuzzy Syst 27(5):956–965
Kumar P, Saini R, Roy P P, Dogra D P (2018) A position and rotation invariant framework for sign language recognition using kinect. Multimed Tools Appl 77(7):8823–8846
Kumar P, Saini R, Yadava M, Roy P P, Dogra D P, Balasubramanian R (2017) Virtual trainer with real-time feedback using kinect sensor. In: Proceedings of the IEEE Region 10 Symposium (TENSYMP), pp 1–5
Kunze K, Lukowicz P (2008) Dealing with sensor displacement in motion-based onbody activity recognition systems. In: Proceedings of the 10th international conference on Ubiquitous computing, pp 20–29
Lester J, Choudhury T, Kern N, Borriello G, Hannaford B (2005) A hybrid discriminative/generative approach for modeling human activities. In: Proceedings of the 19th international joint conference on Artificial intelligence, pp 766–772
Linna M, Kannala J, Rahtu E (2018) Real-time human pose estimation with convolutional neural networks. In: Proceedings of the VISIGRAPP, pp 335–342
Lukowicz P, Ward J A, Junker H, Stäger M, Tröster G, Atrash A, Starner T (2004) Recognizing workshop activity using body worn microphones and accelerometers. In: International conference on pervasive computing. Springer, pp 18–32
Maurer U, Smailagic A, Siewiorek D P, Deisher M (2006) Activity recognition and monitoring using multiple sensors on different body positions. In: International Workshop on Wearable and Implantable Body Sensor Networks. IEEE, pp 1–4
O’Reilly M, Caulfield B, Ward T, Johnston W, Doherty C (2018) Wearable inertial sensor systems for lower limb exercise detection and evaluation: a systematic review. Sports Med 48(5):1221–1246
Parkka J, Ermes M, Korpipaa P, Mantyjarvi J, Peltola J, Korhonen I (2006) Activity classification using realistic data from wearable sensors. IEEE Trans Inf Technol Biomed 10(1):119–128
Saini R, Kumar P, Kaur B, Roy P P, Dogra D P, Santosh KC (2019) Kinect sensor-based interaction monitoring system using the blstm neural network in healthcare. Int J Mach Learn Cybern 10(9):2529–2540
Saini R, Kumar P, Roy P P, Dogra D P (2018) A novel framework of continuous human-activity recognition using kinect. Neurocomputing 311:99–111
Seeger C, Buchmann A, Van Laerhoven K (2011) myhealthassistant: a phone-based body sensor network that captures the wearer’s exercises throughout the day. In: Proceedings of the 6th International Conference on Body Area Networks. Institute for Computer Sciences and Social-Informatics, pp 1–7
Shotton J, Girshick R, Fitzgibbon A, Sharp T, Cook M, Finocchio M, Moore R, Kohli P, Criminisi A, Kipman A et al (2012) Efficient human pose estimation from single depth images. IEEE Trans Pattern Anal Mach Intell 35(12):2821–2840
Simon T, Joo H, Matthews I, Sheikh Y (2017) Hand keypoint detection in single images using multiview bootstrapping. In: Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, pp 1145–1153
Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556
Toshev A, Szegedy C (2014) Deeppose: Human pose estimation via deep neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 1653–1660
Ukita N, Uematsu Y (2018) Semi-and weakly-supervised human pose estimation. Comput Vis Image Underst 170:67–78
Wang H, Ullah M M, Klaser A, Laptev I, Schmid C (2009) Evaluation of local spatio-temporal features for action recognition. In: Bmvc 2009-british machine vision conference. BMVA Press, pp 124–1
Wang L, Xiong Y, Wang Z, Qiao Y, Lin D, Tang X, Van Gool L (2016) Temporal segment networks: Towards good practices for deep action recognition. In: Proceedings of the European conference on computer vision. Springer, pp 20–36
Wei S-E, Ramakrishna V, Kanade T, Sheikh Y (2016) Convolutional pose machines. In: Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, pp 4724–4732
Xiaohan Nie B, Xiong C, Zhu S-C (2015) Joint action recognition and pose estimation from video. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 1293–1301
Zhang Z, Wang N, Cui L (2018) Fine-fit: A fine-grained gym exercises recognition system. In: Proceedings of the 24th Asia-Pacific Conference on Communications (APCC), pp 492–497
Zhu A, Zhang S, Huang Y, Hu F, Cui R, Hua G (2019) Exploring hard joints mining via hourglass-based generative adversarial network for human pose estimation. AIP Adv 9(035321-3):1–9
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflicts of Interest
The authors declared that they have no conflicts of interest to this work.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Rangari, T., Kumar, S., Roy, P.P. et al. Video based exercise recognition and correct pose detection. Multimed Tools Appl 81, 30267–30282 (2022). https://doi.org/10.1007/s11042-022-12299-z
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-022-12299-z