Abstract
In the performing arts context, video analysis plays a significant role in analyzing the artists’ movement in educational and professional artistic environments. However, current research lacks user-centered systems for artists to directly engage with movement detection in their respective practice recordings. Our framework presents an annotation tool capable of automatically analyzing human poses of individual examples in real-time and presenting motion data in the form of simple reports.
At the core of this system, augmenting multimedia content by manually adding different types of annotations to the intended video segment is possible. Therefore, we developed novel features that combine automatic intelligent movement analysis with manual annotation capabilities for a more rewarding and efficient user experience. We demonstrate how pose estimation techniques can be employed to set up customized sets of human poses that can be automatically identified in videos. In addition, this paper discusses the results obtained from two system evaluation phases, where we report our findings based on expert feedback.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
A version of the annotation tool is available at https://motion-notes.di.fct.unl.pt/.
References
Torre, I., Galluccio, I., Coccoli, M.: Video augmentation to support video-based learning. In: ACM International Conference Proceeding Series (2022)
Dias Pereira Dos, A., Loke, L., Martinez-Maldonado, R.: Exploring video annotation as a tool to support dance teaching. In: ACM International Conference Proceeding Series, pp. 448–452 (2018)
Paul, S., Saha, R., Padhi, S., Majumdar, S., Das, P.P., Rao, S.: NrityaManch: an annotation and retrieval system for bharatanatyam dance ACM reference format. Proceedings of the 14th Annual Meeting of the Forum for Information Retrieval Evaluation, p. 22 (2022). https://doi.org/10.1145/3574318
Singh, V., Latulipe, C., Carroll, E., Lottridge, D.: The choreographer’s notebook-a video annotation system for dancers and choreographers. C and C 2011 - Proceedings of the 8th ACM Conference on Creativity and Cognition, pp. 197–206 (2011). https://doi.org/10.1145/2069618.2069653
Lagrue, S., et al.: An ontology web application-based annotation tool for intangible culture heritage dance videos. SUMAC 2019 - Proceedings of the 1st Workshop on Structuring and Understanding of Multimedia heritAge Contents, co-located with MM 2019, pp. 75–81 (2019). https://doi.org/10.1145/3347317.3357245
Chen, K., Zhang, D., Yao, L., Guo, B., Yu, Z., Liu, Y.: Deep learning for sensor-based human activity recognition. ACM Computing Surveys (CSUR), p. 54 (2021). https://doi.org/10.1145/3447744
Kuffner dos Anjos, R., Ribeiro, C., Fernandes, C., Sartor, J.B.: Three-Dimensional Visualization of Movement Qualities in Contemporary Dance (2018). https://doi.org/10.1145/3212721.3212812
Ho, C., Tsai, W.T., Lin, K.S., Chen, H.H.: Extraction and alignment evaluation of motion beats for street dance. ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing – Proceedings, pp. 2429–2433 (2013). https://doi.org/10.1109/ICASSP.2013.6638091
Ashwin, T.S., Prakash, V., Rajendran, R.: A systematic review of intelligent tutoring sys-tems based on Gross body movement detected using computer vision. Computers and Education: Artificial Intell. 4, 100125 (2023). https://doi.org/10.1016/J.CAEAI.2023.100125
Kong, Y., Fu, Y.: Human action recognition and prediction: a survey. Int J Comput Vis. 130, 1366–1401 (2022). https://doi.org/10.1007/S11263-022-01594-9
Munea, T.L., Jembre, Y.Z., Weldegebriel, H.T., Chen, L., Huang, C., Yang, C.: The progress of human pose estimation: a survey and taxonomy of models applied in 2D human pose estimation. IEEE Access. 8, 133330–133348 (2020). https://doi.org/10.1109/ACCESS.2020.3010248
Difini, G.M., Martins, M.G., Barbosa, J.L.V.: Human pose estimation for training assistance: a systematic literature review. ACM International Conference Proceeding Series, pp.189–196 (2021). https://doi.org/10.1145/3470482.3479633
Cao, Z., Simon, T., Wei, S.E., Sheikh, Y.: Realtime multi-person 2D pose estimation using part affinity fields. In: Proceedings - 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017 (2017)
Li, Y.C., Chang, C.T., Cheng, C.C., Huang, Y.L.: Baseball swing pose estimation using openpose. 2021 IEEE International Conference on Robotics, Automation and Artificial Intelligence, RAAI 2021, pp. 6–9 (2021). https://doi.org/10.1109/RAAI52226.2021.9507807
Potempski, F., Sabo, A., Patterson, K.K.: Technical note: quantifying music-dance synchrony during salsa dancing with a deep learning-based 2D pose estimator. J Biomech 141, 111178 (2022). https://doi.org/10.1016/J.JBIOMECH.2022.111178
Muhammad, K., et al.: Human action recognition using attention based LSTM network with dilated CNN features. Future Generation Computer Syst. 125, 820–830 (2021). https://doi.org/10.1016/J.FUTURE.2021.06.045
Yen, C.C., Pin, T., Xu, H.: Bilateral pose transformer for human pose estimation. ACM International Conference Proceeding Series. Par F180473, pp. 21–29 (2022). https://doi.org/10.1145/3532342.3532346
Xia, H., Zhang, Q.: VitPose: multi-view 3D human pose estimation with vision transformer. 2022 IEEE 8th International Conference on Computer and Communications, ICCC 2022, pp. 1922–1927 (2022). https://doi.org/10.1109/ICCC56324.2022.10065997
Janardhanan, J., Umamaheswari, S.: A comprehensive study on human pose estimation. 8th International Conference on Advanced Computing and Communication Systems, ICACCS 2022, pp. 535–541 (2022). https://doi.org/10.1109/ICACCS54159.2022.9784965
Raheb, K.E., Stergiou, M., Katifori, A., Ioannidis, Y.: Dance interactive learning systems. ACM Comput Surv. 52, 1–37 (2020). https://doi.org/10.1145/3323335
Towey, D., et al.: Students as partners in a multi-media note-taking app development: best practices. In: Proceedings - 2017 IEEE/ACM 39th International Conference on Software Engineering Companion, ICSE-C, pp. 334–335. Institute of Electrical and Electronics Engineers Inc. (2017)
Rodrigues, R., Neves Madeira, R.: Studying natural user interfaces for smart video annotation towards ubiquitous environments. In: 20th International Conference on Mobile and Ubiquitous Multimedia (MUM 2021), pp. 1–18. ACM, New York, NY, USA, Leuven, Belgium (2021)
de Sousa, L., Richter, B., Nel, C.: The effect of multimedia use on the teaching and learning of Social Sciences at tertiary level: a case study. Yesterday and Today, pp. 1–22 (2017). https://doi.org/10.17159/2223-0386/2017/n17a1
Sidi, Y., Blau, I., Shamir-Inbal, T.: Mapping active and collaborative learning in higher education through annotations in hyper-video by learning analytics. J. Comput. Assist Learn. 38, 1752–1764 (2022). https://doi.org/10.1111/JCAL.12714
Tseng, S.S.: The influence of teacher annotations on student learning engagement and video watching behaviors. Int. J. Educ. Technol. High. Educ. 18, 1–17 (2021). https://doi.org/10.1186/S41239-021-00242-5/FIGURES/5
Mirriahi, N., Jovanović, J., Lim, L.A., Lodge, J.M.: Two sides of the same coin: video an-notations and in-video questions for active learning. Education Tech. Research Dev. 69, 2571–2588 (2021). https://doi.org/10.1007/S11423-021-10041-4/TABLES/3
Rijmer, S.: Negotiating deliberate choice-making: Insights from an interdisciplinary and mul-timodal encounter during the making of a new contemporary dance. In: Dance Data, Cognition, and Multimodal Communication, pp. 15–37. Routledge, London (2022)
Jurgens, S.: Three methods of designing a workflow with multimodal video annotation in interdisciplinary choreographic processes. Multimodality and Performance, Fernandes, C. (Ed.). Cambridge Scholars Publishing, Newcastle upon Tyne, pp. 159–178 (2016)
Butterworth, J., Wildschut, L.: Contemporary Choreography: A Critical Reader. Routledge (2009)
Harvie, J., Lavender, A.: Making Contemporary Theatre: International Rehearsal Processes, p. 252 (2010)
Jürgens, S., Fernandes, C.: Choreographic practice-as-research. In: Performance as Research, pp. 249–274. Routledge, First edition. Routledge, New York (2017)
Bargeron, D., Gupta, A., Grudin, J., Sanocki, E.: Annotations for streaming video on the web. Conference on Human Factors in Computing Systems – Proceedings, pp. 278–279 (1999). https://doi.org/10.1145/632716.632887
Wittenburg, P., Brugman, H., Russel, A., Klassmann, A., Sloetjes, H.: ELAN: a professional framework for multimodality research. In: Proceedings of the 5th International Conference on Language Resources and Evaluation, LREC 2006, pp. 1556–1559 (2006)
Buono, P., Desolda, G., Lanzilotti, R.: Scenes extraction from videos of telementored surgeries. Proceedings: DMS 2013 - 19th International Conference on Distributed Multimedia Systems, pp. 106–111 (2013)
Pless, A., Hari, R., Brem, B., Woermamm, U., Schnabel, K.P.: Using self and peer video annotations of simulated patient encounters in communication training to facilitate the reflection of communication skills: an implementation study. GMS J Med Educ. 38 (2021). https://doi.org/10.3205/ZMA001451
Deng, D., et al.: EventAnchor: reducing Human Interactions in Event Annotation of Racket Sports Videos KEYWORDS. https://doi.org/10.1145/3411764.3445431
Risko, E.F., Foulsham, T., Dawson, S., Kingstone, A.: The collaborative lecture annotation system (CLAS): a new TOOL for distributed learning. IEEE Trans. Learn. Technol. 6, 4–13 (2013). https://doi.org/10.1109/TLT.2012.15
Cabral, D., Valente, J., Silva, J., Aragão, U., Fernandes, C., Correia, N.: A creation-tool for contemporary dance using multimodal video annotation. MM’11 - Proceedings of the 2011 ACM Multimedia Conference and Co-Located Workshops, pp. 905–908 (2011). https://doi.org/10.1145/2072298.2071899
Ribeiro, C., Kuffner, R., Fernandes, C., Pereira, J.: 3D annotation in contemporary dance. In: Proceedings of the 3rd International Symposium on Movement and Computing, pp. 1–4. ACM, New York, NY, USA (2016)
de Lahunta, S., Jenett, F.: Making digital choreographic objects interrelate. In: Performing the Digital, pp. 63–80. transcript Verlag (2016)
Rizzo, A., et al.: WhoLoDancE: whole-body interaction learning for dance education. In: CEUR Workshop Proceedings, pp. 41–50 (2018)
Lagrue, S., et al.: An Ontology Web Application-based Annotation Tool for Intangible Culture Heritage Dance Videos (2019). https://doi.org/10.1145/3347317.3357245
Rivière, J.P., Alaoui, S.F., Caramiaux, B., Mackay, W.E.: Capturing movement decomposition to support learning and teaching in contemporary dance. Proc ACM Hum Comput Interact. 3 (2019). https://doi.org/10.1145/3359188
Rodrigues, R., Madeira, R.N., Correia, N.: Exploring the user interaction with a multimodal web-based video annotator. Lecture Notes of the Institute for Computer Sciences, Social-Informatics and Telecommunications Engineering, LNICST. 429 LNICST, pp. 13–22 (2022). https://doi.org/10.1007/978-3-030-99188-3_2/COVER
Rodrigues, R., Madeira, R.N., Correia, N., Fernandes, C., Ribeiro, S.: Multimodal web based video annotator with real-time human pose estimation. Lecture Notes in Comput-er Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). 11872 LNCS, pp. 23–30 (2019). https://doi.org/10.1007/978-3-030-33617-2_3/COVER
Diogo, J., Rodrigues, R., Madeira, R., Correia, N.: Video annotation tool using human pose estimation for sports training. ACM International Conference Proceeding Series, pp. 262–265 (2022). https://doi.org/10.1145/3568444.3570592
Acknowledgments
This work is funded by Fundação para a Ciência e Tecnologia through Studentship grants: 2020.09417.BD (Ph.D.), UIDB/04516/2020/BIM/11 (MSc). It was supported by the project WEAVE, Grant Agreement Number: INEA/CEF/ICT/A2020/ 2288018. It is also supported by NOVA LINCS (UIDB/04516/2020) with the financial support of FCT.IP. Lastly, we would like to thank Virgínia Gonçalves, Ana Carvalho, Alina Teles, Bárbara Salvador, Berta Berton, Carolina Ramalho, Inês Costa and Mariana Colaço from Centro Cultural e Recreativo do Alto do Moinho for their feedback.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Rodrigues, R., Diogo, J., Jurgens, S., Fernandes, C., Correia, N. (2023). Sample-Based Human Movement Detection for Interactive Videos Applied to Performing Arts. In: Abdelnour Nocera, J., Kristín Lárusdóttir, M., Petrie, H., Piccinno, A., Winckler, M. (eds) Human-Computer Interaction – INTERACT 2023. INTERACT 2023. Lecture Notes in Computer Science, vol 14144. Springer, Cham. https://doi.org/10.1007/978-3-031-42286-7_32
Download citation
DOI: https://doi.org/10.1007/978-3-031-42286-7_32
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-42285-0
Online ISBN: 978-3-031-42286-7
eBook Packages: Computer ScienceComputer Science (R0)