Abstract
We present an approach that uses visual hand detection, tracking and gesture recognition to provide an interactive interface between the user and a multimedia application. We use the Viola–Jones cascade detector to locate the hand and then utilise the unscented Kalman filter and a pre-constructed hand model for efficient object tracking. We then use a semantic-probabilistic method to recognise hand gestures that provide a simple and user-friendly way for the user to interact with the application. We perform experimental testing of our tracking system, validate the gesture recognition approach and evaluate the proposed approach in an augmented reality application. Finally, we discuss several possible applications that can use our gesture recognition approach, including a virtual reality application, a smart home control system and an interface to interact with museum exhibits.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Itkarkar, R.R., Nandi, A.V.: A survey of 2D and 3D imaging used in hand gesture recognition for human-computer interaction (HCI). In: 2016 IEEE International WIE Conference on Electrical and Computer Engineering (WIECON-ECE), Pune, pp. 188–193 (2016)
De Jesus Oliveira, V.A., Nedel, L., Maciel, A.: Proactive haptic articulation for intercommunication in collaborative virtual environments. In: 2016 IEEE Symposium on 3D User Interfaces (3DUI), Greenville, SC, pp. 91–94 (2016)
Sreejith, M., Rakesh, S., Gupta, S., Biswas, S., Das, P.P.: Real-time hands-free immersive image navigation system using Microsoft Kinect 2.0 and Leap Motion Controller. In: 2015 Fifth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG), Patna, pp. 1–4 (2015)
Invitto, S., Faggiano, C., Sammarco, S., De Luca, V., De Paolis, L.T.: Interactive entertainment, virtual motion training and brain ergonomy. In: 7th International Conference on Intelligent Technologies for Interactive Entertainment (INTETAIN), 2015, Turin, pp. 88–94 (2015)
Chang, C.-C., Chen, I.-Y., Huang, Y.-S.: Hand pose recognition using curvature scale space. In: IEEE International Conference on Pattern Recognition (2002)
Utsumi, A., Miyasato, T., Kishino, F.: Multi-camera hand pose recognition system using skeleton image. In: IEEE International Workshop on Robot and Human Communication, pp. 219–224 (1995)
Aoki, Y., Tanahashi, S., Xu, J.: Sign language image processing for intelligent communication by communication satellite. In: IEEE International Conference on Acoustics, Speech, and Signal Processing (1994)
Yang, J., Zhu, C., Yuan, J.: Real time hand gesture recognition via finger-emphasized multi-scale description. In: 2017 IEEE International Conference on Multimedia and Expo (ICME), Hong Kong, Hong Kong, pp. 631–636 (2017)
Zhao, S., Tan, W., Wu, C., Wen, L.: A Novel Interactive Method of Virtual Reality System Based on Hand Gesture Recognition, pp. 5879–5882. IEEE (2009). ISBN: 978-1-4244-2723-9/09
Sepehri, A., Yacoob, Y., Davis, L.: Employing the hand as an interface device. J. Multimed. 1(7), 18–29 (2006)
Guan, Y., Zheng, M.: Real-time 3D pointing gesture recognition for natural HCI. In: Proceedings of the World Congress on Intelligent Control and Automation, China, pp. 2433–2436 (2008)
Freeman, W., Weissman, C.: Television control by hand gesture. In: IEEE International Workshop on Automatic Face and Gesture Recognition, Zurich (1995)
Kumar, P., Saini, R., Behera, S.K., Dogra, D.P., Roy, P.P.: Real-time recognition of sign language gestures and air-writing using leap motion. In: 2017 Fifteenth IAPR International Conference on Machine Vision Applications (MVA), Nagoya, pp. 157–160 (2017)
Erol, A., Bebis, G., Nicolescu, M., Boyle, R., Twombly, X.: Vision-based hand pose estimation: a review. Comput. Vis. Image Underst. 108, 52–73 (2007)
Bao, P., Binh, N., Khoa, T.: A new approach to hand tracking and gesture recognition by a new feature type and HMM. In: International Conference on Fuzzy Systems and Knowledge Discovery. IEEE Computer Society (2009)
Kovalenko, M., Antoshchuk, S., Sieck, J.: Real-time hand tracking and gesture recognition using a semantic-probabilistic network. In: Proceedings of 16th International Conference on Computer Modelling and Simulation, Cambridge, pp. 269–274 (2014)
Viola, P., Jones, M.: Robust real-time object detection. Int. J. Comput. Vis. 57(2), 137–154 (2004)
Open Computer Vision Library. Retrieved from 26 June 2016. http://sourceforge.net/projects/opencvlibrary/
Tayal, Y., Lamba, R., Padhee, S.: Automatic face detection using colour based segmentation. Int. J. Sci. Res. Publ. 2(6), 1–7 (2012)
Wang, X., Wang, J.: Simulation analysis of EKF and UKF implementations in PHD filter. In: 2016 IEEE 13th International Conference on Networking, Sensing, and Control (ICNSC), Mexico City, pp. 1–6 (2016)
Cooper, G., Herskovits, E.: A Bayesian method for the induction of probabilistic networks from data. Mach. Learn. 9(4), 309–347 (1992)
Murphy, K.: Dynamic bayesian networks: representation, inference and learning. Ph.D. thesis, University of California at Berkley (2002)
Wobbrock, J.O., Wilson, A.D., Li, Y.: Gestures without libraries, toolkits or training: a $1 recognizer for user interface prototypes. In: Proceedings of the ACM Symposium on User Interface Software and Technology, Newport, Rhode Island, pp. 159–168, 7–10 Oct 2007
Apolinário, A.L., Giraldi, G.A., Macedo, M.C., Souza, A.C.: A markerless augmented reality approach based on real-time 3D reconstruction using Kinect. In: Workshop of Works in Progress (WIP) in SIBGRAPI 2013, XXVI Conference on Graphics, Patterns and Images, Salvador, Brazil (2014)
Khandelwal, P., Swarnalatha, P., Bisht, N., Prabu, S.: Detection of features to track objects and segmentation using GrabCut for application in marker-less augmented reality. Procedia Comput. Sci. 58, 698–705 (2015)
Sridhar, S., Oulasvirta, A., Theobalt, C.: Interactive markerless articulated hand motion tracking using RGB and depth data. In: Proceedings of the IEEE International Conference on Computer Vision (2013)
Baumann, A., Boltz, M., Ebling, J.: A review and comparison of measures for automatic video surveillance systems. EURASIP J. Image Video Process. 3, 280–312 (2008)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Singapore Pte Ltd.
About this chapter
Cite this chapter
Antoshchuk, S., Kovalenko, M., Sieck, J. (2018). Gesture Recognition-Based Human–Computer Interaction Interface for Multimedia Applications. In: Jat, D., Sieck, J., Muyingi, HN., Winschiers-Theophilus, H., Peters, A., Nggada, S. (eds) Digitisation of Culture: Namibian and International Perspectives. Springer, Singapore. https://doi.org/10.1007/978-981-10-7697-8_16
Download citation
DOI: https://doi.org/10.1007/978-981-10-7697-8_16
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-10-7696-1
Online ISBN: 978-981-10-7697-8
eBook Packages: Computer ScienceComputer Science (R0)