Abstract
Applications for gait recognition are numerous especially in security surveillance. However, due to the variety of individual walking behaviours and the complexities of external variables during data gathering, gait identification continues to face several obstacles. Among these, shallow learning-based gait recognition algorithms struggle to attain the correct rate of recognition crucial by numerous applications, while the volume of gait training data available cannot match the demands of deep learning-based model training. In order to comply with the problem outlined above, this work offers a visual gait detection system based on an attention-based multi-scale convolutional network with sequential learning. As a first step, the approach takes multi-recurrent gait energy images (MR-GEIs) as an input in a frame by frame manner and runs each frame through the convolutional network to extract entire gait features. In the second step, the key attributes of the extracted features from multi-scale convolutions are highlighted by the attention block in order to improve prediction performance. Thirdly, the bidirectional gated recurrent unit (Bi-GRU) layer is applied to obtain the temporal relationships among the different frames of MR-GEIs in the sequential learning block. The proposed network achieves average accuracies of 93.4% on CASIA-B and 97.4% on OULP gait datasets, demonstrating superior recognition performance and improved generalizability compared to previous state-of-the-art techniques.
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11760-024-03765-2/MediaObjects/11760_2024_3765_Fig1_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11760-024-03765-2/MediaObjects/11760_2024_3765_Fig2_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11760-024-03765-2/MediaObjects/11760_2024_3765_Fig3_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11760-024-03765-2/MediaObjects/11760_2024_3765_Fig4_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11760-024-03765-2/MediaObjects/11760_2024_3765_Fig5_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11760-024-03765-2/MediaObjects/11760_2024_3765_Fig6_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs11760-024-03765-2/MediaObjects/11760_2024_3765_Fig7_HTML.png)
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data availability
The authors do not have permission to share data and proper citations are provided.
References
Singh, J.P., Jain, S., Arora, S., Singh, U.P.: A survey of behavioral biometric gait recognition: current success and future perspectives. Arch. Comput. Methods Eng. 28, 107–148 (2021)
Singh, J.P., Jain, S., Arora, S., Singh, U.P.: Vision-based gait recognition: a survey. IEEE Access 6, 70497–70527 (2018)
Wang, X., Yan, W.Q.: Cross-view gait recognition through ensemble learning. Neural Comput. Appl. 32, 7275–7287 (2020)
Yiwei, H., Junping, Z.: Deep learning of gait recognition: A review [j]. Pattern Recognit. Artif. Intell. 31(05), 442–452 (2018)
Han, J., Bhanu, B.: Individual recognition using gait energy image. IEEE Trans. Pattern Anal. Mach. Intell. 28(2), 316–322 (2005)
Connie, T., Goh, M.K.O., Teoh, A.B.J.: A Grassmannian approach to address view change problem in gait recognition. IEEE Trans. Cybern. 47(6), 1395–1408 (2016)
Guan, Y., Li, C.T., Roli, F.: On reducing the effect of covariate factors in gait recognition: a classifier ensemble method. IEEE Trans. Pattern Anal. Mach. Intell. 37(7), 1521–1528 (2014)
Theekhanont, P., Kurutach, W., Miguet, S.: In: 2012 International Symposium on Information Technologies in Medicine and Education, vol. 2 (IEEE, 2012), pp. 936–940
Wan, M., Yang, G., Gai, S., Yang, Z.: Two-dimensional discriminant locality preserving projections (2ddlpp) and its application to feature extraction via fuzzy set. Multimed. Tools Appl. 76, 355–371 (2017)
Wan, M., Lai, Z., Yang, G., Yang, Z., Zhang, F., Zheng, H.: Local graph embedding based on maximum margin criterion via fuzzy set. Fuzzy Sets Syst. 318, 120–131 (2017)
Mahfouf, Z., Merouani, H.F., Bouchrika, I., Harrati, N.: Investigating the use of motion-based features from optical flow for gait recognition. Neurocomputing 283, 140–149 (2018)
Shiraga, K., Makihara, Y., Muramatsu, D., Echigo, T., Yagi, Y.: Geinet: view-invariant gait recognition using a convolutional neural network. 2016 international conference on biometrics (ICB) pp. 1–8 (2016)
Li, C., Min, X., Sun, S., Lin, W., Tang, Z.: Deepgait: a learning deep convolutional representation for view-invariant gait recognition using joint Bayesian. Appl. Sci. 7(3), 210 (2017)
Takemura, N., Makihara, Y., Muramatsu, D., Echigo, T., Yagi, Y.: On input/output architectures for convolutional neural network-based cross-view gait recognition. IEEE Trans. Circuits Syst. Video Technol. 29(9), 2708–2719 (2017)
Wu, Z., Huang, Y., Wang, L., Wang, X., Tan, T.: A comprehensive study on cross-view gait based human identification with deep cnns. IEEE Trans Pattern Anal. Mach. Intell. 39(2), 209–226 (2016)
Gao, J., Gao, X., Wu, N., Yang, H.: Bi-directional lstm with multi-scale dense attention mechanism for hyperspectral image classification. Multimed. Tools Appl. 81(17), 24003–24020 (2022)
Kuan, L., Yan, Z., Xin, W., Yan, C., Xiangkun, P., Wenxue, S., Zhe, J., Yong, Z., Nan, X., Xin, Z.: Short-term electricity load forecasting method based on multilayered self-normalizing gru network. 2017 IEEE Conference on Energy Internet and Energy System Integration (EI2) pp. 1–5 (2017)
Xiong, C., Merity, S., Socher, R.: Dynamic memory networks for visual and textual question answering. International conference on machine learning pp. 2397–2406 (2016)
Yu, S., Tan, D., Tan, T.: A framework for evaluating the effect of view angle, clothing and carrying condition on gait recognition. 18th international conference on pattern recognition (ICPR’06) 4: 441–444 (2006)
Iwama, H., Okumura, M., Makihara, Y., Yagi, Y.: The ou-isir gait database comprising the large population dataset and performance evaluation of gait recognition. IEEE Trans. Inf. Forensics Secur. 7(5), 1511–1521 (2012)
Li, X., Makihara, Y., Xu, C., Yagi, Y., Yu, S., Ren, M.: End-to-end model-based gait recognition. Proceedings of the Asian conference on computer vision (2020)
Zhang, Z., Tran, L., Liu, F., Liu, X.: On learning disentangled representations for gait recognition. IEEE Trans. Pattern Anal. Mach. Intell. 44(1), 345–360 (2020)
Chao, H., Wang, K., He, Y., Zhang, J., Feng, J.: Gaitset: cross-view gait recognition through utilizing gait as a deep set. IEEE Trans. Pattern Anal. Mach. Intell. 44(7), 3467–3478 (2021)
Huo, W., Wang, K., Tang, J., Wang, N., Liang, D.: Gaitscm: causal representation learning for gait recognition. Comput. Vis. Image Underst. 243, 103995 (2024)
Acknowledgements
This work was supported by Science and Engineering Research Board (grant no. CRG/2019/001414).
Author information
Authors and Affiliations
Contributions
Mohammad Iman Junaid: Writing – original draft, Methodology, Investigation, Conceptualization. Sandeep Madarapu: Writing – review & editing. Samit Ari: Writing – review & editing, Supervision, Resources.
Corresponding author
Ethics declarations
Conflict of interest
The authors declare no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Junaid, M.I., Madarapu, . & Ari, S. Human gait recognition using attention based convolutional network with sequential learning. SIViP 19, 157 (2025). https://doi.org/10.1007/s11760-024-03765-2
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1007/s11760-024-03765-2