Skip to main content

Human-Autonomy Teaming and Explainable AI Capabilities in RTS Games

  • Conference paper
  • First Online:
Book cover Engineering Psychology and Cognitive Ergonomics. Cognition and Design (HCII 2020)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 12187))

Included in the following conference series:

Abstract

Real-time strategy games often times mimic the appearance and feel of military-like command and control systems. Artificial intelligence and machine learning research enjoys utilizing the environments produced by these games and are often focused on creating intelligent agents to beat them or accomplish high scores. Instead of creating these agents, or bots, to beat real-time strategy games, this work instead focuses on creating machine learning-driven agents to work with human players. However, due to the advancements in the deep learning field, machine learning models have become harder to understand, and are called ‘black-box models’ due to the inability to see the inner workings of such algorithms. To remedy this, we describe how we take human-autonomy teaming and explainable artificial intelligence techniques to shed light and provide inside for players to understand recommendations by the system.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    https://deepmind.com/blog/article/alphastar-mastering-real-time-strategy-game-starcraft-ii.

  2. 2.

    https://sites.google.com/site/micrortsaicompetition/home.

  3. 3.

    https://bot-bowl.com/.

  4. 4.

    Regulation (EU) 2016/679.

  5. 5.

    https://www.acm.org/articles/bulletins/2017/january/usacm-statement-algorithmic-accountability.

  6. 6.

    https://palletsprojects.com/p/flask.

  7. 7.

    https://humansystems.arc.nasa.gov/groups/tlx/downloads/TLXScale.pdf.

References

  1. Shively, R.J., Lachter, J., Brandt, S.L., Matessa, M., Battiste, V., Johnson, W.W.: Why human-autonomy teaming? In: Baldwin, C. (ed.) AHFE 2017. AISC, vol. 586, pp. 3–11. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-60642-2_1

    Chapter  Google Scholar 

  2. Ezer, N., Bruni, S., Cai, Y., Hepenstal, S.J., Miller, C.A., Schmorrow, D.D.: Trust engineering for Human-AI teams. In: Proceedings of the Human Factors and Ergonomics Society Annual Meeting, vol. 63, no. 1, pp. 322–362 (2019)

    Google Scholar 

  3. Reddy, S., Dragan, A.D., Levine, S.: Shared autonomy via deep reinforcement learning. arXiv prepreint arXiv:1802.01744 (2018)

  4. Explainable Artificial Intelligence (XAI): DARPA Broad Agency Announcement, DARPA-BAA-16-53 (2016)

    Google Scholar 

  5. Rudin, C.: Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Mach. Intell. 1, 206–215 (2019)

    Article  Google Scholar 

  6. Goodman, B., Flaxman, S.: European union regulations on algorithmic decision-making and a ‘Right to Explain’. AI Magazine 38(3), 50–57 (2017)

    Article  Google Scholar 

  7. Ribeiro, M.T., Singh, S., Guestrin, C.: Why should i trust you?: explaining the predictions of any classifier. CoRR, abs/1602.04938 (2016)

    Google Scholar 

  8. Shrikumar, A., Greenside, P., Kundaje, A.: Learning important features through propagating activation differences. CoRR, abs/1704.02685 (2017)

    Google Scholar 

  9. Zeiler, M.D., Fergus, R.: Visualizing and understanding convolutional networks. CoRR, abs/1311.2901 (2013)

    Google Scholar 

  10. Selvaraju, R.R., Das, A., Vedantam, R., Cogswell, M., Parikh, D., Batra, D.: Grad-CAM: why did you say that? visual explanations from deep networks via gradiant-based localization. CoRR, abs/1610.02391 (2016)

    Google Scholar 

  11. Vinyals, O., et al.: StarCraft II: a new challenge for reinforcement learning. arXiv preprint:1708.04782

    Google Scholar 

  12. Lee, D., Tang, H., Zhang, J.O., Xu, H., Darrell, T., Abbeel, P.: Modular architecture for StarCraft II with deep reinforcement learning. In: AAAI Fourteenth Artificial Intelligence and Interactive Digital Entertainment Conference (2018)

    Google Scholar 

  13. Justesen, N., Risi, S.: Learning macromanagement in StarCraft from replays using deep learning. In: IEEE Conference on Computational Intelligence in Games, pp. 162–169 (2017)

    Google Scholar 

  14. Reed, A.E., Mikels, J.A., Lockenhoff, C.E.: Choosing with confidence: self-efficacy and preferences for choice. Judgm. Decis. Mak. 7(2), 173–180 (2012)

    Google Scholar 

  15. Synnaeve, G. et al.: Forward modeling for partial observation strategy games - a StarCraft defogger. In: Proceedings of the 32nd International Conference on Neural Information Processing Systems, pp. 10761–10771 (2018)

    Google Scholar 

  16. Sun, Y., Vu, K.-P.L.: Population stereotypes for color associations. In: Harris, D. (ed.) EPCE 2018. LNCS (LNAI), vol. 10906, pp. 480–489. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-91122-9_39

    Chapter  Google Scholar 

  17. Endsley, M.R.: Situation awareness global assessment technique (SAGAT). In: Proceedings of the IEEE 1988 National Aerospace and Electronics Conference (1988)

    Google Scholar 

  18. Kulesza, T., Burnett, M., Wong, W.-K., Stumph, S.: Principles of explanatory debugging to personalize interactive machine learning. In: Proceedings of the 20th International Conference on Intelligent User Interfaces, pp. 126–137 (2015)

    Google Scholar 

  19. Anderson, A., et al: Explaining reinforcement learning to mere mortals: an empirical study. CoRR, abs/1903.09708 (2019)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Crisrael Lucero .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 This is a U.S. government work and not under copyright protection in the U.S.; foreign copyright protection may apply

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Lucero, C., Izumigawa, C., Frederiksen, K., Nans, L., Iden, R., Lange, D.S. (2020). Human-Autonomy Teaming and Explainable AI Capabilities in RTS Games. In: Harris, D., Li, WC. (eds) Engineering Psychology and Cognitive Ergonomics. Cognition and Design. HCII 2020. Lecture Notes in Computer Science(), vol 12187. Springer, Cham. https://doi.org/10.1007/978-3-030-49183-3_13

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-49183-3_13

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-49182-6

  • Online ISBN: 978-3-030-49183-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics