skip to main content
10.1145/3512576.3512596acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicitConference Proceedingsconference-collections
research-article

RTS Game AI Robots Winner Prediction Based on Replay Data by using Deep Learning

Published: 11 April 2022 Publication History

Abstract

The success of the AlphaGo series of artificial intelligence (AI) programs in the game of Go has injected new vitality into AI research. The research of real-time strategy (RTS) games using deep learning algorithms provides an ideal method for AI planning, state evaluations, and human behavior research. Our work focused on winner prediction for RTS game AI robots based on states and actions information of replay dataset generated by μRTS simulator by using deep learning. Since previous winner prediction methods only encoded states information without actions information, the game information is incomplete and the accuracy of prediction is low. In the present work, we used one-hot encoding to encode both “states” elements and “actions” elements in the attributes of each sampled time point and corresponding “winner” attribute, which can achieve higher winner prediction performance. We used the encoded datasets to train (validate) and test five deep learning algorithms: CNN, MSCNN, CNP, LSTM, and BNN. The experimental results showed that all five deep learning methods can achieve an accuracy greater than 77% (often greater than 80%) using this new encoding on a task comparing with around 60% accuracy of previous algorithms. The Area Under the Receiver Operating Characteristic (ROC) Curve (AUC) value can achieve greater than 0.85 (often greater than 0.9). Finally, we sampled 20 time points from the beginning to the end of each match's replay data to illustrate the feasibility of our methods.

References

[1]
Chijindu EVC. 2012. Search in Artificial Intelligence Problem Solving. African Journal of Computing & ICT 5(5), 37–42
[2]
Lecun Y., Bengio Y., Hinton G. 2015. Deep learning. Nature 521(7553), 436–444
[3]
Krizhevsky A., Sutskever I., Hinton G. 2012. ImageNet classification with deep convolutional neural networks. In: Pereira F., Burges C. J. C., Bottou L., Weinberger K. Q.Adv. Neural(eds.) Inf. Process. Syst., Vol. 25, pp. 1097–1105
[4]
He K., Zhang X., Ren S., 2016. Deep residual learning for image recognition. IEEE Conf. Comput. Vis. Pattern Recognit, pp. 770–778.
[5]
Silver D., Huang A., Maddison C.J., 2016. Mastering the game of Go with deep neural networks and tree search. Nature 529(7587), 484–489.
[6]
Silver D., Schrittwieser J., Simonyan K., 2017. Mastering the game of Go without human knowledge. Nature 550(7676), 354–359.
[7]
Vinyals O., Babuschkin I., Czarnecki W.M., 2019. Grandmaster level in StarCraft II using multi-agent reinforcement learning. Nature 575(7782), 350–355.
[8]
Huang J., Yang W.L. 2018. A multi-size convolution neural network for RTS games winner prediction. MATEC Web of Conferences, pp. 257–261.
[9]
Ontan˜o´n S., Synnaeve G., Uriarte A., 2013. A Survey of Real-Time Strategy Game AI Research and Competition in StarCraft. IEEE Transactions on Computational Intelligence and AI in Games 5(4), 293–311.
[10]
Buro M. 2003. Real-time strategy games: a new AI research challenge. Intl Joint Conf. Artificial Intelligence, pp. 1534—1535.
[11]
Stanescu M., Barriga N.A., Hess A., 2016. Evaluating Real-Time Strategy Game States Using Convolutional Neural Networks. IEEE Conference on Computational Intelligence and Games (CIG), pp. 1–7.
[12]
Ravari Y.N., Bakkes S., Spronck P. 2016. StarCraft Winner Prediction, Artificial Intelligence in Adversarial Games: Papers from the AIIDE Workshop. AAAI Technical Report, WS–16–21.
[13]
Yang Z., Ontan˜o´n S. 2018. Learning Map-Independent Evaluation Functions for RealTime Strategy Games. IEEE Conference on Computational Intelligence and Games (CIG), pp. 301–307.
[14]
Yang W.L., Huang J., Hu Y. 2019. A Modified multi-size convolution neural network for winner prediction based on time serial datasets. International Conference on Mathematics and Artificial Intelligence (ICMAI 2019), Chengdu, China, ACM Digital Library, ISBN: 978–1–4503–6258–0, pp. 110–114.
[15]
Shleyfman A., Komenda A., Domshlak C. 2014. On Combinatorial Actions and CMABs with Linear Side Information. Frontiers in Artificial Intelligence and Applications, pp. 825–830.
[16]
Lin M., Wang T., LI X.B., 2019. An uncertainty-incorporated approach to predict the winner in StarCraft II using neural processes. IEEE Access 7, 101609– 101619.
[17]
Lecun Y., Bottou L., Bengio Y., 1998. Gradient-based learning applied to document recognition. Proceedings of the IEEE 86(11), 2278–2324.
[18]
Callet P.L., Viard-Gaudin C., Barba D. 2006. A Convolutional Neural Network Approach for Objective Video Quality Assessment. IEEE Transaction on neural networks 17(5), 1316–1327(2006).
[19]
Garnelo M., Rosenbaum D., Maddison C.J., 2018. Conditional Neural Processes. International Conference on Machine Learning, arXiv:1807.01613.
[20]
Kononenko I. 1989. Bayesian Neural Networks. Biological Cybernetics 61, 361– 370.
[21]
Roth W., Pernkopf F. 2020. Bayesian Neural Networks with Weight Sharing Using Dirichlet Processes. IEEE Transactions on Pattern Analysis and Machine Intelligence 42(1), 246–252.
[22]
Sepp H., Ju¨rgen S. 1997. Long Short-Term Memory. Neural Computation 9(8), 1735–1780.
[23]
Yang W., Zhang Q., Peng Y. 2018. A Dynamic Hierarchical Evaluating Network for Real-Time Strategy Games. MATEC Web of Conferences 208(4), ICMIE, pp. 1– 6.
[24]
Ontan˜o´n S., Buro M. 2015. Adversarial Hierarchical-Task Network Planning for Complex Real-Time Games. International Joint Conference on Artificial Intelligence (IJCAI), pp. 1652–1658.
[25]
Erickson G.K.S., Buro M. 2014. Global State Evaluation in StarCraft. Annual AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment (AIIDE), pp. 112–118.
[26]
Blundell C., Cornebise J., Kavukcuoglu K., 2015. Weight uncertainty in neural networks. JMLR Workshop Conf Proc 37, pp. 1613-–1622.
[27]
Shridhar K., Laumann F., Maurin A.L., 2018. Bayesian convolutional neural networks with variational inference. arXiv:1806.05978v6 [cs.LG] 14, pp. 1–12.

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Other conferences
ICIT '21: Proceedings of the 2021 9th International Conference on Information Technology: IoT and Smart City
December 2021
584 pages
ISBN:9781450384971
DOI:10.1145/3512576
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 11 April 2022

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. deep learning
  2. game AI robots
  3. real-time strategy
  4. replay data
  5. winner prediction

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

ICIT 2021
ICIT 2021: IoT and Smart City
December 22 - 25, 2021
Guangzhou, China

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 53
    Total Downloads
  • Downloads (Last 12 months)12
  • Downloads (Last 6 weeks)1
Reflects downloads up to 22 Jan 2025

Other Metrics

Citations

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format.

HTML Format

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media