Abstract
Pedestrian trajectory prediction is crucial across a wide range of applications like self-driving vehicles and social robots. Such prediction is challenging because crowd behavior is inherently determined by various factors, such as obstacles, stationary crowd groups and destinations which were difficult to effectively represent. Especially pedestrians tend to be greatly affected by the pedestrians in front of them more than those behind them, which were often ignored in literature. In this paper, we propose a novel framework of Social-Scene-Aware Generative Adversarial Networks (SSA-GAN), which includes three modules, to predict the future trajectory of pedestrians in dynamic scene. Specifically, in the Scene module, we model the original scene image into a scene energy map by combining various scene factors and calculating the probability of pedestrians passing at each location. And the modeling formula is inspired by the distance relationship between pedestrians and scene factors. Moreover, the Social module is used to aggregate neighbors’ interactions on the basis of the correlation between the motion history of pedestrians. This correlation is captured by the self-attention pooling module and limited by the field of view. And then the Generative Adversarial module with variety loss can solve the multimodal problem of pedestrian trajectory. Extensive experiments on publicly available datasets validate the effectiveness of our method for crowd behavior understanding and trajectory prediction.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Alahi, A., Goel, K., Ramanathan, V., et al.: Social LSTM: human trajectory prediction in crowded spaces. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 961–971 (2016)
Radenović, F., Tolias, G., Chum, O.: CNN image retrieval learns from BoW: unsupervised fine-tuning with hard examples. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9905, pp. 3–20. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46448-0_1
Chandra, R., Bhattacharya, U., Bera, A., et al.: Traphic: trajectory prediction in dense and heterogeneous traffic using weighted interactions. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 8483–8492 (2019)
Tang, Y.C., Salakhutdinov, R.: Multiple futures prediction. arXiv preprint arXiv:1911.00997 (2019)
Ziebart, B.D., Ratliff, N., Gallagher, G., et al.: Planning-based prediction for pedestrians. In: IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 3931–3936 (2009)
Shao, J., Change, L.C., Wang, X.: Scene-independent group profiling in crowd. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2219–2226 (2014)
Yi, S., Li, H., Wang, X.: Understanding pedestrian behaviors from stationary crowd groups. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3488–3496 (2015)
Helbing, D., Molnar, P.: Social force model for pedestrian dynamics. Phys. Rev. E 51(5), 4282 (1995)
Mehran, R., Oyama, A., Shah, M.: Abnormal crowd behavior detection using social force model. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 935–942 (2009)
Xu, Y., Piao, Z., Gao, S.: Encoding crowd interaction with deep neural network for pedestrian trajectory prediction. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5275–5284 (2018)
Zhang, P., Ouyang, W., Zhang, P., et al.: SR-LSTM: State refinement for LSTM towards pedestrian trajectory prediction. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 12085–12094 (2019)
Yi, S., Wang, X., Lu, C., et al.: L0 regularized stationary time estimation for crowd group analysis. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2211–2218 (2014)
Gupta, A., Johnson, J., Fei-Fei, L., et al.: Social gan: Socially acceptable trajectories with generative adversarial networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2255–2264 (2018)
Vaswani, A., Shazeer, N., Parmar, N., et al.: Attention is all you need. arXiv preprint arXiv:1706.03762 (2017)
Sadeghian, A., Kosaraju, V., Sadeghian, A., et al.: Sophie: an attentive gan for predicting paths compliant to social and physical constraints. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 1349–1358 (2019)
Huang, Y., Bi, H., Li, Z., et al.: STGAT: modeling spatial-temporal interactions for human trajectory prediction. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 6272–6281 (2019)
Goodfellow, I.J., Pouget-Abadie, J., Mirza, M., et al.: Generative Adversarial Nets. MIT Press (2014)
Ho, J., Ermon, S.: Generative adversarial imitation learning. arXiv preprint arXiv:1606.03476 (2016)
Tao, C., Jiang, Q., Duan, L., et al.: Dynamic and static context-aware LSTM for multi-agent motion prediction. arXiv preprint arXiv:2008.00777 (2020)
Zou, H., Su, H., Song, S., et al.: Understanding human behaviors in crowds by imitating the decision-making process. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32(1) (2018)
Yang, W., Sheng, B., et al.: Deep color guided coarse-to-fine convolutional network cascade for depth image super-resolution. IEEE Trans. Image Process. Public. IEEE Signal Process. Soc. (2018)
Da, K.: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)
Pellegrini, S., Ess, A., Van Gool, L.: Improving data association by joint modeling of pedestrian trajectories and groupings. In: Daniilidis, K., Maragos, P., Paragios, N. (eds.) ECCV 2010. LNCS, vol. 6311, pp. 452–465. Springer, Heidelberg (2010). https://doi.org/10.1007/978-3-642-15549-9_33
Lerner, A., Chrysanthou, Y., Lischinski, D.: Crowds by example. In: Computer Graphics Forum. Blackwell Publishing Ltd, Oxford, UK, vol. 26(3), pp. 655–664 (2007)
He, K., Zhang, X., Ren, S., et al.: Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778 (2016)
Chen, Z., Hu, Z., Sheng, B., et al.: Simplified non-locally dense network for single-image dehazing. Visual Comput. 36(9), 2189–2200 (2020)
Zhang, B., Sheng, B., Li, P., et al.: Depth of field rendering using multilayer-neighborhood optimization. In: IEEE Transactions on Visualization and Computer Graphics, p. 1 (2019)
Acknowledgement
This work was supported by the Natural Science Foundation of Shanghai (Grant 19ZR1415800), Shanghai Science and Technology Commission (Grant 21511100700), the Research Project of Shanghai Science and Technology Commission (Grant 20dz2260300), the Fundamental Research Funds for the Central Universities.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Huang, B., Ma, Z., Chen, L., He, G. (2021). Social-Scene-Aware Generative Adversarial Networks for Pedestrian Trajectory Prediction. In: Magnenat-Thalmann, N., et al. Advances in Computer Graphics. CGI 2021. Lecture Notes in Computer Science(), vol 13002. Springer, Cham. https://doi.org/10.1007/978-3-030-89029-2_15
Download citation
DOI: https://doi.org/10.1007/978-3-030-89029-2_15
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-89028-5
Online ISBN: 978-3-030-89029-2
eBook Packages: Computer ScienceComputer Science (R0)