skip to main content
10.1145/3373509.3373577acmotherconferencesArticle/Chapter ViewAbstractPublication PagesiccprConference Proceedingsconference-collections
research-article

Multi-view Human Action Recognition by Cell Summary Descriptor and Decision Fusion

Published: 25 March 2020 Publication History

Abstract

Multi-view human action recognition (MVHAR) is essential for many applications including smart video surveillance in shopping malls, airports, railway stations and other public places, as well as the ambient assisted living systems. Currently, a lot of methods realized MVHAR by constructing high dimensional features or by complex calculation process, which causes the recognition speed difficult to meet the needs of real-time application system. To address the problem, a cell summary (CS) descriptor is constructed for each frame by dividing each frame into several cells and further dividing each cell into several radial bins and then extracting spatiotemporal features from each radial bin. Then a video is represented by a sequence of dimension-reduced CS descriptors. The CS descriptor is not only discriminative but also low computational cost. A probabilistic classifier is learned for each view of each action category, and then action classification is carried out independently in each view. A probability estimation based decision fusion algorithm is proposed to make a final decision. Experimental results on the two publically available multi-view human action datasets MuHAVi-MAS-14 and IXMAS show that the proposed approach is superior to state-of-the-art methods in recognition accuracy and moreover, the approach is so computationally efficient that it is appropriate for real-time applications.

References

[1]
Singh S, Velastin S A, Ragheb H. MuHAVi: A multicamera human action video dataset for the evaluation of action recognition methods[C]. IEEE International Conference on Advanced Video and Signal Based Surveillance. Boston, MA, USA: IEEE, 2010: 48--55.
[2]
Chaaraoui A A, Climent-Pérez P, Flórez-Revuelta F. An efficient approach for multi-view human action recognition based on bag-of-key-poses[C]. International Workshop on Human Behavior Understanding. Vilamoura, Portugal: Springer, 2012: 29--40.
[3]
Modarres A F A, Soryani M. A pyramidal layered HMM for multiview human behavior recognition in asynchronous video streams [J]. International Journal of Computer Applications, 2014, 96(7).
[4]
Chaaraoui A A, Flórez-Revuelta F. A low-dimensional radial silhouette-based feature for fast human action recognition fusing multiple views [J]. International scholarly research notices, 2014, 2014:1--11.
[5]
Yan P, Khan S M, Shah M. Learning 4D action feature models for arbitrary view action recognition[C]. IEEE Conference on Computer Vision and Pattern Recognition. Alaska, USA: IEEE, 2008: 1--7.
[6]
Holte M B, Chakraborty B, Gonzalez J, et al. A local 3-D motion descriptor for multi-view human action recognition from 4-D spatio-temporal interest points [J]. IEEE Journal of Selected Topics in Signal Processing, 2012, 6(5): 553--565.
[7]
Sargano A, Angelov P, Habib Z. Human action recognition from multiple views based on view-invariant feature descriptor using support vector machines[J]. Applied Sciences, 2016, 6(10): 309--322.
[8]
Kong Y, Ding Z, Li J, et al. Deeply learned view-invariant features for cross-view action recognition [J]. IEEE Transactions on Image Processing, 2017, 26(6): 3028--3037.
[9]
Aryanfar A, Yaakob R, Halin A A, et al. Multi-view human action recognition using wavelet data reduction and multi-class classification [J]. Procedia Computer Science, 2015, 62: 585--592.
[10]
Liu C, Li Z, Shi X, et al. Learning a mid-level representation for multiview action recognition [J]. Advances in Multimedia, 2018, 2018: 1--10.
[11]
Cilla R, Patricio M A, Berlanga A, et al. A probabilistic, discriminative and distributed system for the recognition of human actions from multiple views [J]. Neurocomputing, 2012, 75(1): 78--87.
[12]
Iosifidis A, Tefas A, Pitas I. Multi-view action recognition based on action volumes, fuzzy distances and cluster discriminant analysis[J]. Signal Processing, 2013, 93(6): 1445--1457.
[13]
Weinland D, Ronfard R, Boyer E. Free viewpoint action recognition using motion history volumes [J]. Computer vision and image understanding, 2006, 104(2--3): 249--257.
[14]
Chaaraoui A A, Climent-Pérez P, Flórez-Revuelta F. Silhouette-based human action recognition using sequences of key poses [J]. Pattern Recognition Letters, 2013, 34(15): 1799--1807.
[15]
Alcantara M F, Moreira T P, Pedrini H. Real-time action recognition using a multilayer descriptor with variable size [J]. Journal of Electronic Imaging, 2016, 25(1): 013020.
[16]
Putra P U, Shima K, Shimatani K. Markerless human activity recognition method based on deep neural network model using multiple cameras[C]. IEEE International Conference on Control, Decision and Information Technologies. Thessaloniki, Greece: IEEE, 2018: 13--18.
[17]
Cheema S, Eweimidwi A, Thurau C, et al. Action recognition by learning discriminative key poses. In Proceedings of the 2011 IEEE International Conference on Computer Vision Workshops. Barcelona, Spain: IEEE, 2011. 1302--1309.

Index Terms

  1. Multi-view Human Action Recognition by Cell Summary Descriptor and Decision Fusion

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    ICCPR '19: Proceedings of the 2019 8th International Conference on Computing and Pattern Recognition
    October 2019
    522 pages
    ISBN:9781450376570
    DOI:10.1145/3373509
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    In-Cooperation

    • Hebei University of Technology
    • Beijing University of Technology

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 25 March 2020

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Decision Fusion
    2. Human Action Recognition
    3. Motion Tendency
    4. Probability Estimation
    5. Spatial Distribution

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Funding Sources

    • National Nature Science Foundation of China
    • National Key Research and Development Plan

    Conference

    ICCPR '19

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • 0
      Total Citations
    • 60
      Total Downloads
    • Downloads (Last 12 months)1
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 28 Feb 2025

    Other Metrics

    Citations

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media