Skip to main content
Log in

A temporal and channel-combined attention block for action segmentation

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

The task of video action segmentation is to classify an untrimmed long video at the frame level. With the requirement of processing long-term feature sequences containing much information, many computing units and auxiliary-training structures are required. The redundant information in these features can interfere with classification inference. It is an effective feature optimization mechanism to distinguish between useful and useless information by adjusting weight distribution. Such a method can adaptively calibrate complex features without increasing many calculations and improve frame-wise classification performance. Therefore, this study proposes a temporal and channel-combined attention block (TCB) that can be used for temporal sequences. It combines the attention of temporal and channel dimensions to reasonably assign weights to features. TCB contains two submodules: multi-scale temporal attention (MTA) and channel attention (CHA). MTA can adapt to different action instances with varying durations in a video using multilayer dilated convolution to capture multi-scale temporal relations and generate frame-wise attention weights. CHA captures the dependencies between channels and generates channel-wise attention weights to selectively increase the weights of important features. We combined the two attention modules to form a two-dimensional attention mechanism to improve action segmentation performance. We inserted TCB on boundary-aware cascade networks for simulation testing. The results show that our attention mechanism can improve action segmentation performance. In the three action segmentation datasets GTEA, 50Salads, and Breakfast, the accuracy Acc increased by an average of 1.4%, the Edit score increased by an average of 2.1%, and the F1 score increased by an average of approximately 2.1%.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8

Similar content being viewed by others

References

  1. Du Z, Mukaidani H. (2021) Linear dynamical systems approach for human action recognition with dual-stream deep features[J]. Applied Intelligence, pp 1–19

  2. Jiang G, Jiang X, Fang Z, et al. (2021) An efficient attention module for 3d convolutional neural networks in action recognition[J]. Applied Intelligence, pp 1–15

  3. Zhang XY, Huang YP, Mi Y, et al. (2021) Video sketch: A middle-level representation for action recognition[J]. Appl Intell 51(4):2589–2608

    Article  Google Scholar 

  4. Lea C, Flynn MD, Vidal R et al (2017) Temporal convolutional networks for action segmentation and detection[C]. Proceedings of the IEEE conference on computer vision and pattern recognition, pp 156–165

  5. Farha Y A, Gall J (2019) Ms-tcn: Multi-stage temporal convolutional network for action segmentation[C]. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 3575–3584

  6. Singh Li SJ, AbuFarha Y, Liu Y et al (2020) Ms-tcn++: Multi-stage temporal convolutional network for action segmentation[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence

  7. Wang Z, Gao Z, Wang L et al (2020) Boundary-aware cascade networks for temporal action segmentation[C]. In: European conference on computer vision. Springer, Cham, pp 34– 51

  8. Ishikawa Y, Kasai S, Aoki Y et al (2021) Alleviating over-segmentation errors by detecting action boundaries[C]. In: Proceedings of the IEEE/CVF winter conference on applications of computer vision, pp 2322–2331

  9. Singhania D, Rahaman R, Yao A (2021) Coarse to Fine Multi-Resolution Temporal Convolutional Network[J]. arXiv preprint arXiv:2105.10859

  10. Singh B, Marks TK, Jones M et al (2016) A multi-stream bi-directional recurrent neural network for fine-grained action detection[C]. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1961–1970

  11. Richard A, Kuehne H, Gall J (2017) Weakly supervised action learning with rnn based fine-to-coarse modeling[C]. In: Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, pp 754–763

  12. Hu J, Shen L, Sun G (2018) Squeeze-and-excitation networks[C]. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7132–7141

  13. Woo S, Park J, Lee JY et al (2018) Cbam: Convolutional block attention module[C]. In: Proceedings of the European conference on computer vision (ECCV), pp 3–19

  14. Qin Z, Zhang P, Wu F et al (2020) FcaNet: Frequency Channel Attention Networks[J]. arXiv e-prints. arXiv:2012.11879

  15. Wang Q, Wu B, Zhu P, et al. (2020) ECA-Net: efficient channel attention for deep convolutional neural networks, 2020 IEEE[C]. In: CVF conference on computer vision and pattern recognition (CVPR), IEEE

  16. Zhang H, Zu K, Lu J et al (2021) Epsanet: An efficient pyramid split attention block on convolutional neural network[J]. arXiv preprint arXiv:2105.14447

  17. Vaswani A, Shazeer N, Parmar N et al (2017) Attention is all you need[C]. In: Advances in neural information processing systems, pp 5998–6008

  18. Wang Z, She Q, Smolic A (2021) ACTION-Net: Multipath Excitation for Action Recognition[C]. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 13214–13223

  19. He K, Zhang X, Ren S et al (2016) Deep residual learning for image recognition[C]. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778

  20. Carreira J, Zisserman A (2017)

  21. Fathi A, Ren X, Rehg JM (2011) Learning to recognize objects in egocentric activities[C]. In: CVPR 2011. IEEE, pp 3281–3288

  22. Stein S, McKenna SJ (2013) Combining embedded accelerometers with computer vision for recognizing food preparation activities[C]. In: Proceedings of the 2013 ACM ACM international joint conference on Pervasive and ubiquitous computing, pp 729–738

  23. Kuehne H, Arslan A, Serre T (2014) The language of actions: Recovering the syntax and semantics of goal-directed human activities[C]. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 780–787

  24. Richard A, Gall J (2016) Temporal action detection using a statistical language model[C]. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3131–3140

  25. Lea C, Reiter A, Vidal R, et al. (2016) Segmental spatiotemporal cnns for fine-grained action segmentation[C]. In: European conference on computer vision. Springer, Cham, pp 36–52

  26. Ronneberger O, Fischer P, Brox T (2015) U-net: Convolutional networks for biomedical image segmentation[C]. In: International conference on medical image computing and computer-assisted intervention. Springer, Cham, pp 234–241

  27. Ding L, Xu C (2017) Tricornet: A hybrid temporal convolutional and recurrent network for video action segmentation[J]. arXiv preprint arXiv:1705.07818

  28. Lei P, Todorovic S (2018) Temporal deformable residual networks for action segmentation in videos[C]. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 6742–6751

  29. Gammulle H, Fernando T, Denman S, et al. (2019) Coupled generative adversarial network for continuous fine-grained action segmentation[C]. In: 2019 IEEE winter conference on applications of computer vision (WACV), IEEE, pp 200–209

  30. Ding L, Xu C (2018) Weakly-supervised action segmentation with iterative soft boundary assignment[C]. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 6508–6516

  31. Kuehne H, Gall J, Serre T. (2016) An end-to-end generative framework for video segmentation and recognition[C]. In: 2016 IEEE winter conference on applications of computer vision (WACV), IEEE, pp 1–8

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Zhe Cao.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Yang, D., Cao, Z., Mao, L. et al. A temporal and channel-combined attention block for action segmentation. Appl Intell 53, 2738–2750 (2023). https://doi.org/10.1007/s10489-022-03569-2

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10489-022-03569-2

Keywords

Navigation