ABSTRACT
In this paper, we propose a method combining visual local features and depth image information to recognize cooking gestures. We employ the feature calculation method[2] which used extended FAST detector and a compact descriptor CHOG3D to calculate visual local features. We pack the local features by BoW in frame sequences to represent the cooking gestures. In addition, the depth images of hands gestures are extracted and integrated spatio-temporally to represent the position and trajectory information of cooking gestures. The two kinds of features are used to describe cooking gestures, and recognition is realized by employing the SVM. In our method, we determine the gesture class for each frame in cooking sequences. By analyzing the results of frames, we recognize cooking gestures in a continue frame sequences of cooking menus, and find the temporal positions of the recognized gestures.
- A. F. Bobick and J. W. Davis. The recognition of human movement using temporal templates. Transactions on Pattern Analysis and Machine Intelligence, 23(3):257--267, 2001 Google ScholarDigital Library
- Y. Ji, A. Shimada, and R. Taniguchi. A compact 3d descriptor in roi for human action recognition. In IEEE TENCON, 2010.Google Scholar
- B. Packer, K. Saenko, and D. Koller. A combined pose, object, and feature model for action understanding. In CVPR, 2012. Google ScholarDigital Library
- M. Rohrbach, S. Amin, M. Andriluka, and B. Schiele. A database for fine grained activity detection of cooking activities. In CVPR, 2012.Google ScholarCross Ref
- E. Rosten, R. Porter, and T. Drummond. Faster and better: A machine learning approach to corner detection. IEEE Trans. Pattern Analysis and Machine Intelligence, 32:105--119, Jan. 2010. Google ScholarDigital Library
- L. Rybok, S. Friedberger, U. D. Hanebeck, and R. Stiefelhagen. The kit robo-kitchen data set for the evaluation of view-based activity recognition systems. In IEEE-RAS International Conference on Humanoid Robots, 2011.Google ScholarCross Ref
Index Terms
- Cooking gesture recognition using local feature and depth image
Recommendations
Multi-scenario gesture recognition using Kinect
CGAMES '12: Proceedings of the 2012 17th International Conference on Computer Games: AI, Animation, Mobile, Interactive Multimedia, Educational & Serious Games (CGAMES)Hand gesture recognition (HGR) is an important research topic because some situations require silent communication with sign languages. Computational HGR systems assist silent communication, and help people learn a sign language. In this article, a ...
Combining multiple depth-based descriptors for hand gesture recognition
The hand is reliably extracted from the scene by jointly using color and depth data.Features extracted from depth data allow a reliable hand gesture recognition.Multiple features capturing different properties of the gestures are combined together.The ...
Hand gesture recognition with depth data
ARTEMIS '13: Proceedings of the 4th ACM/IEEE international workshop on Analysis and retrieval of tracked events and motion in imagery streamDepth data acquired by current low-cost real-time depth cameras provide a very informative description of the hand pose, that can be effectively exploited for gesture recognition purposes. This paper introduces a novel hand gesture recognition scheme ...
Comments