skip to main content
10.1145/3487075.3487125acmotherconferencesArticle/Chapter ViewAbstractPublication PagescsaeConference Proceedingsconference-collections
research-article

Effective Indoor Fire Detection with Channel Shuffle Module

Authors Info & Claims
Published:07 December 2021Publication History

ABSTRACT

In recent years, methods based on computer vision and deep learning become the mainstream approaches in fire detection. However, the expensive computation cost of 3D convolutional neutral network (CNN) is unbearable and it is difficult for them to capture the fire regions of videos in time. In this paper, we design a module named channel shuffle module (CSM) based on 2D CNN to keep the balance between computation cost and accuracy. By fusing RGB frame and differential frame, CSM improves the ability of 2D CNN in temporal information extraction which much less cost than methods based on 3D CNN. Four different structures of CSM are proposed and we choose the best one by experiment results. Also, experiments prove that the performances of TSN and TSM are improved with CSM in sequence classification. The accuracy of TSM with CSM is 99.2045%, false positive rate reaches 0.7890% and false negative rate reaches 0.4530%, which demonstrates the efficiency of CSM in temporal feature modeling.

References

  1. Tran D, Bourdev L, Fergus R, Torresani L, Paluri M (2015). Learning Spatiotemporal Features with 3D Convolutional Networks. IEEE International Conference on Computer Vision. IEEE.Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Carreira J, Zisserman A (2017). Quo vadis, action recognition? a new model and the kinetics dataset. IEEE.Google ScholarGoogle ScholarCross RefCross Ref
  3. Qiu Z, Yao T, Mei T (2017). Learning Spatio-Temporal Representation with Pseudo-3D Residual Networks. 2017 IEEE International Conference on Computer Vision (ICCV). IEEE.Google ScholarGoogle ScholarCross RefCross Ref
  4. Feichtenhofer C, Fan H, Malik J, K He (2019). SlowFast Networks for Video Recognition. 2019 IEEE/CVF International Conference on Computer Vision (ICCV). IEEE.Google ScholarGoogle Scholar
  5. X Wang, Girshick R, Gupta A, He K (2018). Non-local Neural Networks. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). IEEE.Google ScholarGoogle Scholar
  6. K He, X Zhang, S Ren, and J Sun (2016). Deep residual learning for image recognition. 2016 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). IEEE.Google ScholarGoogle ScholarCross RefCross Ref
  7. Szegedy C, Wei L, Jia Y, Sermanet P, Rabinovich A (2015). Going deeper with convolutions. 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE.Google ScholarGoogle ScholarCross RefCross Ref
  8. Wang L, Xiong Y, Wang Z, Qiao Y, Lin D, Tang X, (2016). Temporal segment networks: towards good practices for deep action recognition. European Conference on Computer Vision.Google ScholarGoogle ScholarCross RefCross Ref
  9. Zhou B, Andonian A, Oliva A, Torralba A (2018). Temporal relational reasoning in videos. European Conference on Computer Vision.Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Lin J, Gan C, Han S. 2019. TSM: Temporal Shift Module for Efficient Video Understanding. 2019 IEEE/CVF International Conference on Computer Vision (ICCV). IEEE.Google ScholarGoogle Scholar
  11. Zolfaghari M, Singh K, Brox T (2018). Eco: efficient convolutional network for online video understanding. European Conference on Computer Vision.Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Du T, Wang H, Torresani L, Ray J, Paluri M (2018). A Closer Look at Spatiotemporal Convolutions for Action Recognition. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). IEEE.Google ScholarGoogle Scholar
  13. Xie S, Sun C, Huang J, Tu Z, Murphy K (2017). Rethinking spatiotemporal feature learning: speed-accuracy trade-offs in video classification.Google ScholarGoogle Scholar
  14. Sun L, Jia K, Yeung D, Shi B (2015). Human Action Recognition Using Factorized Spatio-Temporal Convolutional Networks. IEEE International Conference on Computer Vision pp 4597-4605. IEEE.Google ScholarGoogle Scholar
  15. Lin T, Dollar P, Girshick R, He K, Hariharan B, Belongie S (2017). Feature Pyramid Networks for Object Detection. 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE Computer Society.Google ScholarGoogle Scholar

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Other conferences
    CSAE '21: Proceedings of the 5th International Conference on Computer Science and Application Engineering
    October 2021
    660 pages
    ISBN:9781450389853
    DOI:10.1145/3487075

    Copyright © 2021 ACM

    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 7 December 2021

    Permissions

    Request permissions about this article.

    Request Permissions

    Check for updates

    Qualifiers

    • research-article
    • Research
    • Refereed limited

    Acceptance Rates

    Overall Acceptance Rate368of770submissions,48%
  • Article Metrics

    • Downloads (Last 12 months)17
    • Downloads (Last 6 weeks)4

    Other Metrics

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format .

View HTML Format