Skip to main content
Log in

Deep Q-network-based multi-criteria decision-making framework for virtual simulation environment

  • S. I : Hybridization of Neural Computing with Nature Inspired Algorithms
  • Published:
Neural Computing and Applications Aims and scope Submit manuscript

Abstract

Deep learning improves the realistic expression of virtual simulations specifically to solve multi-criteria decision-making problems, which are generally rely on high-performance artificial intelligence. This study was inspired by the motivation theory and natural life observations. Recently, motivation-based control has been actively studied for realistic expression, but it presents various problems. For instance, it is hard to define the relation among multiple motivations and to select goals based on multiple motivations. Behaviors should generally be practiced to take into account motivations and goals. This paper proposes a deep Q-network (DQN)-based multi-criteria decision-making framework for virtual agents in real time to automatically select goals based on motivations in virtual simulation environments and to plan relevant behaviors to achieve those goals. All motivations are classified according to the five-level Maslow’s hierarchy of needs, and the virtual agents train a double DQN by big social data, select optimal goals depending on motivations, and perform behaviors relying on a predefined hierarchical task networks (HTNs). Compared to the state-of-the-art method, the proposed framework is efficient and reduced the average loss from 0.1239 to 0.0491 and increased accuracy from 63.24 to 80.15%. For behavioral performance using predefined HTNs, the number of methods has increased from 35 in the Q network to 1511 in the proposed framework, and the computation time of 10,000 behavior plans reduced from 0.118 to 0.1079 s.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14

Similar content being viewed by others

References

  1. Wang J, Gu X, Liu W, Sangaiah AK, Kim HJ (2019) An empower Hamilton loop based data collection algorithm with mobile agent for WSNs. Hum-Centric Comput Inf Sci 9(1):18

    Article  Google Scholar 

  2. Yao J, Ni Y, Zhao J, Niu H, Liu S, Zheng Y, Wang J (2019) Data based violated behavior analysis of taxi driver in metropolis in China. CMC Comput Mater Contin 60(3):1109–1122

    Google Scholar 

  3. Long M, Zeng Y (2019) Detecting iris liveness with batch normalized convolutional neural network. Comput Mater Contin 58(2):493–504

    Article  Google Scholar 

  4. Luck M, d’Inverno M (1998) Motivated behaviour for goal adoption. In: Zhang C, Lukose D (eds) Multi-agent systems: theories, languages and applications Lecture notes in computer science, vol 1544. Springer, Berlin

    Google Scholar 

  5. Maslow AH (1943) A theory of human motivation. Psychol Rev 50:370–396

    Article  Google Scholar 

  6. The Sims 4. https://www.ea.com/games/the-sims/the-sims-4

  7. Assassin’s Creed Origins. https://www.ubisoft.com/en-gb/game/assassins-creed-origins

  8. Thalmann D (2007) Crowd simulation. In: Wah BW (ed) Wiley encyclopedia of computer science and engineering. Wiley, New York

    Google Scholar 

  9. Torren PM (2012) Moving agent pedestrians through space and time. Ann Assoc Am Geogr 102:35–66

    Article  Google Scholar 

  10. Pan X, Han CS, Law KH (2007) A multi-agent based simulation framework for the study of human and social behavior in egress analysis. Comput Civ Eng 22:113–132

    Google Scholar 

  11. Musse SR, Thalmann D (2001) Hierarchical model for real time simulation of virtual human crowds. IEEE Trans Vis Comput Graph 7:152–164

    Article  Google Scholar 

  12. Murakami Y, Minami K, Kawasoe T, Ishida T (2002) Multi-agent simulation for crisis management. In: Knowledge media networking, pp 135–139

  13. Lee WS (2012) Smart house simulator design using automatic scenario generation. M.S. thesis, Department of Multimedia Engineering, Dongguk University

  14. Song W, Cho KE, Um KH (2008) Motivation-based hierarchical behavior planning. J Korea Game Soc 8:79–90

    Google Scholar 

  15. Trescak T, Bogdanovych A, Simoff S (2014) City of Uruk 3000 BC: using genetic algorithms, dynamic planning and crowd simulation to re-enact everyday life of ancient Sumerians. In: Social simulation conference

  16. Sevin ED, Thalmann D (2005) A motivational model of action selection for virtual humans. In: Computer graphics international, pp 213–220

  17. Sevin ED, Thalmann D (2005) An affective model of action selection for virtual humans. In: Proceedings of agents that want and like: motivational and emotional roots of cognition and action symposium at the artificial intelligence and social behaviors 2005 conference

  18. Nguyen HT (2018) A study on motivation generation based on Q-network for agent simulations. M.S. thesis, Department of Multimedia Engineering, Dongguk University

  19. Guy SJ, Chhugani J, Kim C, Satish N, Lin M, Manocha D, Dubey P (2009) Clearpath: highly parallel collision avoidance for multi-agent simulation. In: 2009 ACM SIGGRAPH/Eurographics symposium on computer animation, pp 177–187

  20. Bogdanovych A, Trescak T (2017) To plan or not to plan: lessons learned from building large scale social simulations. In: International conference on intelligent virtual agents, pp 53–62

  21. Feldman R, Dagan I, Hirsh H (1998) Mining text using keyword distributions. J Intell Inf Syst 10:281–300

    Article  Google Scholar 

  22. Heimerl F, Lohmann S, Lange S, Ertl T (2014) Word cloud explorer: text analytics based on word clouds. In: 2014 47th Hawaii international conference on system sciences, pp 1833–1842

  23. Gambette P, Véronis J (2010) Visualising a text with a tree cloud. In: Locarek-Junge H, Weihs C (eds) Classification as a tool for research. Springer, Berlin, pp 561–569

    Chapter  Google Scholar 

  24. Aggarwal CC, Wang H (2011) Text mining in social networks. In: Aggarwal C (ed) Social network data analytics. Springer, Boston, pp 353–378

    Chapter  Google Scholar 

  25. Rose S, Engel D, Cramer N, Cowley W (2010) Automatic keyword extraction from individual documents. In: Berry MW, Kogan J (eds) Text mining: applications and theory. Wiley, New York, pp 1–20

    Google Scholar 

  26. Najafabadi MM, Villanustre F, Khoshgoftaar TM, Seliya N, Wald R, Muharemagic E (2015) Deep learning applications and challenges in big data analytics. J Big Data 2:1

    Article  Google Scholar 

  27. Lim CU, Harrell FD (2013) Modeling player preferences in avatar customization using social network data: a case-study using virtual items in Team Fortress 2. In: 2013 IEEE conference on computational intelligence in games (CIG), pp 1–8

  28. Thawonmas R, Kurashige M, Iizuka K, Kantardzic M (2006) Clustering of online game users based on their trails using self-organizing map. In: International conference on entertainment computing, pp 366–369

  29. Christakis NA, Fowler JH (2013) Social contagion theory: examining dynamic social networks and human behavior. Stat Med 32:556–577

    Article  MathSciNet  Google Scholar 

  30. Youyou W, Michal K, David S (2015) Computer-based personality judgments are more accurate than those made by humans. Proc Natl Acad Sci 112:1036–1040

    Article  Google Scholar 

  31. Xu Z, Tang J, Meng J, Zhang W, Wang Y, Liu CH, Yang D (2018) Experience-driven networking: a deep reinforcement learning based approach. In: IEEE Infocom 2018, 15–19 April 2018, Honolulu, HI, USA, pp 1–9

  32. Song Y, Yang G, Xie H, Zhang D, Xingming S (2017) Residual domain dictionary learning for compressed sensing video recovery. Multimed Tools Appl 76(7):10083–10096

    Article  Google Scholar 

  33. Rao Y, Lu J, Zhou J (2017) Attention-aware deep reinforcement learning for video face recognition. In: 2017 IEEE international conference on computer vision (ICCV), 22–29 October 2017, Venice, Italy, pp 3931–3940

  34. Xia Z, Hu Z, Luo J (2017) UPTP vehicle trajectory prediction based on user preference under complexity environment. Wirel Pers Commun 97(3):4651–4665

    Article  Google Scholar 

  35. Hossain MS, Muhammad G (2018) Environment classification for urban big data using deep learning. IEEE Commun Mag 56(11):44–50

    Article  Google Scholar 

  36. Gui Y, Zeng G (2019) Joint learning of visual and spatial features for edit propagation from a single image. Vis Comput 36:1–14

    Google Scholar 

  37. He S, Li Z, Tang Y, Liao Z, Wang J, Kim HJ (2019) Parameters compressing in deep learning. Comput Mater Contin 62:1–16

    Google Scholar 

  38. Zhou S, Ke M, Luo P (2019) Multi-camera transfer GAN for person re-identification. J Vis Commun Image Represent 59:393–400

    Article  Google Scholar 

  39. Meyer JA (1996) Artificial life and the animat approach to artificial intelligence. In: Artificial intelligence. Elsevier Inc, pp 325–354. https://doi.org/10.1016/B978-012161964-0/50013-3

  40. Hasselt HV, Guez A, Silver D (2016) Deep reinforcement learning with double Q-learning. AAAI 2:5

    Google Scholar 

Download references

Acknowledgements

This research was supported by a grant from Defense Acquisition Program Administration and Agency for Defense Development, under contract #UE171095RD, and this work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIP) (2018R1A2B2007934).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Kyungeun Cho.

Ethics declarations

Conflict of interest

We certify that there is no actual or potential conflict of interest in relation to this article.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix

Appendix

The virtual simulation environment for this study, shown in Fig. 15, was implemented using the commercial game engine Unity 2017 (Unity Technologies ApS, San Francisco, CA, USA). The virtual environment demonstrates a small city. In the virtual city, we simulate actions of humans and animals on the road areas. In addition, there are eight junctions including six T junctions and two cross-junctions. Each T junction contains nine traffic lights, and each cross junction contains twelve. Therefore, a total of 78 traffic lights are utilized in the simulator.

Fig. 15
figure 15

Virtual simulation environment

Table 19 lists screens of behaviors, results, motivations, goals, and behavior outcomes when the virtual human agent was activated in the simulation after learning using the proposed framework.

Table 19 Motivation, goal, and behavior result of the virtual human gent in the proposed framework

For instance, Fig. 16a shows virtual human agents commuting to offices or schools in the morning. Figure 16b–d illustrate the virtual human agents having meals at noon, returning home in the evening, and going to sleep later in the evening, respectively.

Fig. 16
figure 16

Virtual human agents selecting goals and performing behaviors

Table 20 lists the behaviors, goals, and top-priority motivations of the virtual animal during simulation after learning using the proposed framework.

Table 20 Motivation–goal–behavior results of the virtual animal in the proposed framework

Figure 17a, b illustrate that multiple virtual animal agents tend to move in groups using the proposed framework. An interaction between virtual animal and human agents is shown in Fig. 17c, and the behavior of only one virtual animal agent is illustrated in Fig. 17d.

Fig. 17
figure 17

Goal selection and behavior performance of the virtual animal

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Jang, H., Hao, S., Chu, P.M. et al. Deep Q-network-based multi-criteria decision-making framework for virtual simulation environment. Neural Comput & Applic 33, 10657–10671 (2021). https://doi.org/10.1007/s00521-020-04918-3

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00521-020-04918-3

Keywords

Navigation