skip to main content
10.1145/3631085.3631229acmotherconferencesArticle/Chapter ViewAbstractPublication PagessbgamesConference Proceedingsconference-collections
research-article

Investigating Reinforcement Learning for Dynamic Difficulty Adjustment

Published:19 January 2024Publication History

ABSTRACT

Dynamic Difficulty Adjustment (DDA) is a technique to automatically adjust various game factors, such as items, maps, or opponent behavior, to provide players with a challenging and engaging experience. The goal is to maintain a balance ensuring an optimal level of enjoyment. In this work, we propose a reinforcement learning agent in a fighting game to create an opponent that matches the player’s skill level. We propose a reward function that leads the player to have similar relative skill to his opponent and maintain a balanced match. Additionally, we introduce a penalty given to the agent during training to constrain its win rate. Therefore, creating an opponent that is not too wear nor too strong. We also explore regularization techniques to improve the agent’s performance and adaptability. We show that regularization improves over the baseline in generalizing its behavior to handle opponents not encountered during training.

References

  1. Mihaly Csikszentmihalyi. 1990. Flow: The Psychology of Optimal Experience.Google ScholarGoogle Scholar
  2. Simon Demediuk, Marco Tamassia, William L. Raffe, Fabio Zambetta, Xiaodong Li, and Florian Mueller. 2017. Monte Carlo tree search based algorithms for dynamic difficulty adjustment. In 2017 IEEE Conference on Computational Intelligence and Games (CIG). 53–59. https://doi.org/10.1109/CIG.2017.8080415Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Simon Demediuk, Marco Tamassia, William L. Raffe, Fabio Zambetta, Xiaodong Li, and Florian Mueller. 2017. Monte Carlo Tree Search Based Algorithms for Dynamic Difficulty Adjustment. In 2017 IEEE Conference on Computational Intelligence and Games (CIG) (New York, NY, USA). IEEE Press, 53–59. https://doi.org/10.1109/CIG.2017.8080415Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Jesse Farebrother, Marlos C. Machado, and Michael Bowling. 2020. Generalization and Regularization in DQN. arxiv:1810.00123 [cs.LG]Google ScholarGoogle Scholar
  5. Miguel González-Duque, Rasmus Berg Palm, and Sebastian Risi. 2021. Fast Game Content Adaptation Through Bayesian-based Player Modelling. https://doi.org/10.48550/ARXIV.2105.08484Google ScholarGoogle ScholarCross RefCross Ref
  6. Juho Hamari and Lauri Keronen. 2017. Why do people play games? A meta-analysis. International Journal of Information Management 37, 3 (2017), 125–141. https://doi.org/10.1016/j.ijinfomgt.2017.01.006Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Robin Hunicke and Vernell Chapman. 2004. AI for dynamic difficulty adjustment in games. Challenges in game artificial intelligence AAAI workshop 2 (01 2004).Google ScholarGoogle Scholar
  8. Robert Kirk, Amy Zhang, Edward Grefenstette, and Tim Rocktäschel. 2023. A Survey of Zero-shot Generalisation in Deep Reinforcement Learning. Journal of Artificial Intelligence Research 76 (jan 2023), 201–264. https://doi.org/10.1613/jair.1.14174Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Raphaël Koster and Will Wright. 2004. A Theory of Fun for Game Design.Google ScholarGoogle Scholar
  10. Changchun Liu, Pramila Agrawal, Nilanjan Sarkar, and Shuo Chen. 2009. Dynamic Difficulty Adjustment in Computer Games Through Real-Time Anxiety-Based Affective Feedback. Int. J. Hum. Comput. Interaction 25 (08 2009), 506–529. https://doi.org/10.1080/10447310902963944Google ScholarGoogle ScholarCross RefCross Ref
  11. Zhuang Liu, Xuanlin Li, Bingyi Kang, and Trevor Darrell. 2021. Regularization Matters in Policy Optimization. arxiv:1910.09191 [cs.LG]Google ScholarGoogle Scholar
  12. Thomas W. Malone. 1981. What Makes Things Fun to Learn? A Study of Intrinsically Motivating Computer Games.Google ScholarGoogle Scholar
  13. Ashey Noblega., Aline Paes., and Esteban Clua.2019. Towards Adaptive Deep Reinforcement Game Balancing. In Proceedings of the 11th International Conference on Agents and Artificial Intelligence - Volume 2: ICAART,. INSTICC, SciTePress, 693–700. https://doi.org/10.5220/0007395406930700Google ScholarGoogle ScholarCross RefCross Ref
  14. OpenAI, :, Christopher Berner, Greg Brockman, Brooke Chan, Vicki Cheung, Przemysław Dębiak, Christy Dennison, David Farhi, Quirin Fischer, Shariq Hashme, Chris Hesse, Rafal Józefowicz, Scott Gray, Catherine Olsson, Jakub Pachocki, Michael Petrov, Henrique P. d. O. Pinto, Jonathan Raiman, Tim Salimans, Jeremy Schlatter, Jonas Schneider, Szymon Sidor, Ilya Sutskever, Jie Tang, Filip Wolski, and Susan Zhang. 2019. Dota 2 with Large Scale Deep Reinforcement Learning. arxiv:1912.06680 [cs.LG]Google ScholarGoogle Scholar
  15. Dvir Ben Or, Michael Kolomenkin, and Gil Shabat. 2021. DL-DDA – Deep Learning based Dynamic Difficulty Adjustment with UX and Gameplay constraints. https://doi.org/10.48550/ARXIV.2106.03075Google ScholarGoogle ScholarCross RefCross Ref
  16. Antonin Raffin, Ashley Hill, Adam Gleave, Anssi Kanervisto, Maximilian Ernestus, and Noah Dormann. 2021. Stable-Baselines3: Reliable Reinforcement Learning Implementations. Journal of Machine Learning Research 22, 268 (2021), 1–8. http://jmlr.org/papers/v22/20-1364.htmlGoogle ScholarGoogle Scholar
  17. John Schulman, Filip Wolski, Prafulla Dhariwal, Alec Radford, and Oleg Klimov. 2017. Proximal Policy Optimization Algorithms. arxiv:1707.06347 [cs.LG]Google ScholarGoogle Scholar
  18. Yoones A. Sekhavat. 2017. MPRL: Multiple-Periodic Reinforcement Learning for difficulty adjustment in rehabilitation games. In 2017 IEEE 5th International Conference on Serious Games and Applications for Health (SeGAH). 1–7. https://doi.org/10.1109/SeGAH.2017.7939260Google ScholarGoogle ScholarCross RefCross Ref
  19. Mirna Silva, Victor Silva, and Luiz Chaimowicz. 2016. Dynamic Difficulty Adjustment on MOBA Games. Entertainment Computing 18 (10 2016). https://doi.org/10.1016/j.entcom.2016.10.002Google ScholarGoogle ScholarCross RefCross Ref
  20. Adi Stein, Yair Yotam, Rami Puzis, Guy Shani, and Meirav Taieb-Maimon. 2018. EEG-triggered dynamic difficulty adjustment for multiplayer games. Entertainment Computing 25 (2018), 14–25. https://doi.org/10.1016/j.entcom.2017.11.003Google ScholarGoogle ScholarCross RefCross Ref
  21. Penelope Sweetser and Peta Wyeth. 2005. GameFlow: A Model for Evaluating Player Enjoyment in Games. Comput. Entertain. 3, 3 (jul 2005), 3. https://doi.org/10.1145/1077246.1077253Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Oriol Vinyals, Igor Babuschkin, Junyoung Chung, Michael Mathieu, Max Jaderberg, Wojtek Czarnecki, Andrew Dudzik, Aja Huang, Petko Georgiev, Richard Powell, Timo Ewalds, Dan Horgan, Manuel Kroiss, Ivo Danihelka, John Agapiou, Junhyuk Oh, Valentin Dalibard, David Choi, Laurent Sifre, Yury Sulsky, Sasha Vezhnevets, James Molloy, Trevor Cai, David Budden, Tom Paine, Caglar Gulcehre, Ziyu Wang, Tobias Pfaff, Toby Pohlen, Dani Yogatama, Julia Cohen, Katrina McKinney, Oliver Smith, Tom Schaul, Timothy Lillicrap, Chris Apps, Koray Kavukcuoglu, Demis Hassabis, and David Silver. 2019. AlphaStar: Mastering the Real-Time Strategy Game StarCraft II. https://deepmind.com/blog/alphastar-mastering-real-time-strategy-game-starcraft-ii/.Google ScholarGoogle Scholar
  23. Su Xue, Meng Wu, John Kolen, Navid Aghdaie, and Kazi A. Zaman. 2017. Dynamic Difficulty Adjustment for Maximized Engagement in Digital Games. In Proceedings of the 26th International Conference on World Wide Web Companion (Perth, Australia) (WWW ’17 Companion). International World Wide Web Conferences Steering Committee, Republic and Canton of Geneva, CHE, 465–471. https://doi.org/10.1145/3041021.3054170Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. Alexander Zook and Mark Riedl. 2021. A Temporal Data-Driven Player Model for Dynamic Difficulty Adjustment. Proceedings of the AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment 8, 1 (Jun. 2021), 93–98. https://ojs.aaai.org/index.php/AIIDE/article/view/12504Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Investigating Reinforcement Learning for Dynamic Difficulty Adjustment

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Other conferences
      SBGames '23: Proceedings of the 22nd Brazilian Symposium on Games and Digital Entertainment
      November 2023
      176 pages

      Copyright © 2023 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 19 January 2024

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article
      • Research
      • Refereed limited
    • Article Metrics

      • Downloads (Last 12 months)15
      • Downloads (Last 6 weeks)7

      Other Metrics

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format .

    View HTML Format