Abstract
In the design of gesture-based user interfaces, continuously recognizing complex dynamic gestures is a challenging task, because of the high-dimensional information of gestures, ambiguous semantic meanings of gestures, and the presence of unpredictable non-gesture body motions. In this paper, we propose a hybrid model that can leverage the time-series modeling ability of hidden Markov model and the fuzzy inference ability of fuzzy neural network. First, a complex dynamic gesture is decomposed and fed into the hybrid model. The likelihood probability of an observation sequence estimated by the hidden Markov model is used as fuzzy membership degree of the corresponding fuzzy class variable in fuzzy neural network. Next, fuzzy rule modeling and fuzzy inference are performed by fuzzy neural network for gesture classification. To spot key gestures accurately, a threshold model is introduced to calculate the likelihood threshold of an input pattern and provide a reliability measure of whether to accept the pattern as a gesture. Finally, the proposed method is applied to recognize ten user-defined dynamic gestures for controlling interactive digital television in a smart room. Results of our experiment show that the proposed method performed better in terms of spotting reliability and recognition accuracy than conventional gesture recognition methods.






Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Feng, Z.Q., Zhang, M.M., Pan, Z.G., Yang, B., Xu, T., Tang, H.K., Li, Y.: 3D-freehand-pose initialization based on operator’s cognitive behavioral models. Vis. Comput. 26(6–8), 607–617 (2010)
Kulshreshth, A., LaViola Jr., J.J.: Exploring the usefulness of finger-based 3D gesture menu selection. In: Proceedings of the ACM Conference on Human Factors in Computing Systems (CHI’14), pp. 1093–1102 (2014)
Pfeil, K.P., Koh, S.L., LaViola Jr. J.J.: Exploring 3D gesture metaphors for interaction with unmanned aerial vehicles. In: Proceedings of the 18th International Conference on Intelligent User Interfaces (IUI’13), pp. 257–266 (2013)
LaViola, J.J.: An introduction to 3D gestural interfaces. In: Proceedings of SIGGRAPH’14, ACM SIGGRAPH 2014 Course, Vancouver, British Columbia, Canada. doi:10.1145/2614028.2615424
Mitra, S., Acharya, T.: Gesture recognition: a survey. IEEE Trans. Syst. Man Cybern. Part C: Appl. Rev. 37(3), 311–324 (2007)
Rautaray, S.S., Agrawal, A.: Vision based hand gesture recognition for human computer interaction: a survey. Artif. Intell. Rev. (2012). doi:10.1007/s10462-012-9356-9
Kölsch, M., Turk, M., Höllerer, T.: Vision-based interfaces for mobility. In: Proceedings of IEEE International Conference on Mobile and Ubiquitous Systems (Mobiquitous’04), pp. 86–94 (2004)
Ren, Z., Yuan, J.S., Meng, J.J., Zhang, Z.Y.: Robust part-based hand gesture recognition using Kinect sensor. IEEE Trans. Multimed. 15(5), 1110–1120 (2013)
Stenger, B., Torr, P.H.S., Cipolla, R.: Model-based hand tracking using a hierarchical Bayesian filter. IEEE Trans. Pattern Anal. Mach. Intell. 28(9), 1372–1384 (2006)
Stergiopoulou, E., Papamarkos, N.: Hand gesture recognition using a neural network shape fitting technique. Eng. Appl. Artif. Intell. 22(8), 1141–1158 (2009)
Bobick, A.F., Davis, J.W.: The recognition of human movement using temporal templates. IEEE Trans. Pattern Anal. Mach. Intell. 23(3), 257–267 (2001)
Kim, I.C., Chien, S.I.: Analysis of 3D hand trajectory gestures using stroke-based composite hidden Markov models. Appl. Intell. 15, 131–143 (2001)
Lee, H., Kim, J.: An HMM-based threshold model approach for gesture recognition. IEEE Trans. Pattern Anal. Mach. Intell. 21(10), 961–973 (1999)
Oz, C., Leu, M.C.: American sign language word recognition with a sensory glove using artificial neural networks. Eng. Appl. Artif. Intell. 24, 1204–1213 (2011)
Seo, H.J., Milanfar, P.: Action recognition from one example. IEEE Trans. Pattern Anal. Mach. Intell. 33(5), 867–882 (2011)
Sohn, M.K., Lee, S.H., Kim, D.J., Kim, B., Kim, H.: 3D hand gesture recognition from one example. In: IEEE International Conference on Consumer Electronics, pp. 171–172 (2013)
Yang, H.D., Park, A.Y., Lee, S.W.: Gesture spotting and recognition for human–robot interaction. IEEE Trans. Robot. 23(2), 256–270 (2007)
Jacob, R.J.K.: Eye Movement-Based Human–Computer Interaction Techniques: Toward Non-Command Interfaces. In: Advances in Human–Computer Interaction, vol. 4. Ablex Publishing Co., Norwood, pp. 151–190 (1993)
Peng, B., Qian, G.: Online gesture spotting from visual hull data. IEEE Trans. Pattern Anal. Mach. Intell. 33(6), 1175–1188 (2011)
Bilal, S., Akmeliawati, R., Shafie, A.A., Salami, M.J.E.: Hidden Markov model for human to computer interaction: a study on human hand gesture recognition. Artif. Intell. Rev. 40(4), 495–516 (2013)
Wang, X.Y., Dai, G.Z. : A Novel Method to Recognize Complex Dynamic Gesture by Combing HMM and FNN Models. In: Proceedings of the 2007 IEEE Symposium on Computational Intelligence in Image and Signal Processing (CIIP’07), pp. 13–18 (2007)
Pedersoli, F., Benini, S., Adami, N., Leonardi, R.: XKin: an open source framework for hand pose and gesture recognition using Kinect. Vis. Comput. 30(10), 1107–1122 (2014)
Zhu, J.Y., Wang, X.Y., Wang, W.X., Dai, G.Z.: Hand gesture recognition based on structure analysis. Chin. J. Comput. 29(12), 2130–2137 (2006)
Wang, X.Y., Zhang, X.W., Dai, G.Z.: Continuous Input for Gesture-based Interaction. In: Proceedings of the 12th International Conference on Intelligent User Interfaces (IUI’07), pp. 235–242 (2007)
Coupe, S.: Machine learning of projected 3D shape. Ph.D. Thesis, University of Manchester, Manchester (2009)
Linde, Y., Buzo, A., Gray, R.M.: An algorithm for vector quantizer design. IEEE Trans. Commun. 28(1), 84–95 (1980)
Rabiner, L.R.: A tutorial on hidden Markov models and selected applications in speech recognition. Proc. IEEE 77(2), 257–286 (1989)
Viterbi, A.J.: Error bounds for convolution codes and an asymptotically optimum decoding algorithm. IEEE Trans. Inf. Theory 13, 260–269 (1967)
Yang, H.D., Sclaroff, S., Lee, S.W.: Sign language spotting with a threshold model based on conditional random fields. IEEE Trans. Pattern Anal. Mach. Intell. 31(7), 1264–1277 (2009)
Wu, H.Y., Wang, J.M., Zhang, X.L.: User-centered gesture development in TV viewing environment. Multimed. Tools Appl. (2014). doi:10.1007/s11042-014-2323-5
Colaco, A., Kirmani, A., Yang, H.S., Gong, N.W., Schmandt, C., Goyal, V.K.: Mime: compact, low-power 3D gesture sensing for interaction with head-mounted displays. In: Proceedings of the ACM Symposium of User Interface Software and Technology (UIST’ 13), pp. 227–236 (2013)
Mo, Z.Y., Lewis, J.P., Neumann, U.: SmartCanvas: a gesture-driven intelligent drawing desk. In: Proceedings of the ACM Symposium of User Interface Software and Technology (UIST’05), pp. 239-243 (2005)
Wilson, A.D.: Robust computer vision-based detection of pinching for one and two-handed gesture input. In: Proceedings of the ACM Symposium of User Interface Software and Technology (UIST’ 06), pp. 255–258 (2006)
Hilliges, O., Izadi, S., Wilson, A.D., Hodges, S., Mendoza, A.G., Butz, A.: Interactions in the air: adding further depth to interactive tabletops. In: Proceedings of the ACM Symposium of User Interface Software and Technology (UIST’ 09). pp. 139–148 (2009)
Song, P., Goh, W.B., Hutama, W., Fu, C.W., Liu, X.P.: A handle bar metaphor for virtual object manipulation with mid-air interaction. In: Proceedings of the ACM Conference on Human Factors in Computing Systems (CHI’12), pp. 1297–1306 (2012)
Wobbrock, J.O., Morris, M.R., Wilson, A.D.: User-defined gestures for surface computing. In: Proceedings of the ACM Conference on Human Factors in Computing Systems (CHI’ 09), pp. 1083–1092 (2009)
Acknowledgments
We thank the financial support from the National Natural Science Foundation of China, No. 61202344; the Fundamental Research Funds for the Central Universities, Sun Yat-Sen University, No. 1209119; Special Project on the Integration of Industry, Education and Research of Guangdong Province, No. 2012B091000062; the Fundamental Research Funds for the Central Universities, Tongji University, Nos. 0600219052, 0600219053. We would like to express our great appreciation to editor and reviewers.
Author information
Authors and Affiliations
Corresponding authors
Rights and permissions
About this article
Cite this article
Wu, H., Wang, J. & Zhang, X. Combining hidden Markov model and fuzzy neural network for continuous recognition of complex dynamic gestures. Vis Comput 33, 1265–1278 (2017). https://doi.org/10.1007/s00371-015-1147-2
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00371-015-1147-2