Skip to main content

Deep Reinforcement Learning Methods in Match-3 Game

  • Conference paper
  • First Online:

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 11832))

Abstract

A large number of methods are being developed in the deep reinforcement learning area recently, but the scope of their application is limited. The number of environments does not always allow for a comprehensive assessment of a new agent training algorithm. The main purpose of this article is to present another environment for Match-3 game that could be expanded, which would have a connection with the real business. The results for the most popular deep reinforcement learning algorithms are presented as a baseline.

I. Makarov—The work was supported by the Russian Science Foundation under grant 17-11-01294 and performed at National Research University Higher School of Economics, Russia.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Notes

  1. 1.

    Detailed implementation can be found at https://github.com/kamildar/gym-match3.

References

  1. Brockman, G., et al.: OpenAI Gym (2016)

    Google Scholar 

  2. Browne, C., et al.: A survey of Monte Carlo tree search methods. IEEE Trans. Comput. Intell. AI Games 4(1), 1–43 (2012). https://doi.org/10.1109/TCIAIG.2012.2186810. http://www.incompleteideas.net/609dropbox/otherreadingsandresources/MCTS-survey.pdf

    Article  Google Scholar 

  3. Chollet, F.: Xception: deep learning with depthwise separable convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1251–1258. IEEE (2017)

    Google Scholar 

  4. Eisen, P., Gudmundsson, S.F., Dowling, J.: Simulating human game play for level difficulty estimation with convolutional neural networks. Technical report, KTH, School of Information and Communication Technology (ICT) (2017). http://kth.diva-portal.org/smash/get/diva2:1149021/FULLTEXT01.pdf

  5. Kocsis, L., Szepesvári, C.: Bandit based Monte-Carlo planning. In: Fürnkranz, J., Scheffer, T., Spiliopoulou, M. (eds.) ECML 2006. LNCS (LNAI), vol. 4212, pp. 282–293. Springer, Heidelberg (2006). https://doi.org/10.1007/11871842_29

    Chapter  Google Scholar 

  6. Liang, E., et al.: RLlib: abstractions for distributed reinforcement learning. arXiv preprint arXiv:1712.09381, December 2017. http://arxiv.org/abs/1712.09381

  7. Mnih, V., et al.: Asynchronous methods for deep reinforcement learning. In: International Conference on Machine Learning, pp. 1928–1937 (2016)

    Google Scholar 

  8. Mnih, V., et al.: Playing Atari with deep reinforcement learning. CoRR (2013). http://arxiv.org/abs/1312.5602

  9. Mnih, V., et al.: Human-level control through deep reinforcement learning. Nature 518(7540), 529–533 (2015). https://doi.org/10.1038/nature14236. http://www.nature.com/articles/nature14236

    Article  Google Scholar 

  10. Poromaa, E.R.: Crushing candy crush: predicting human success rate in a mobile game using Monte-Carlo tree search. Technical report, KTH, School of Computer Science and Communication (CSC) (2017). http://kth.diva-portal.org/smash/get/diva2:1093469/FULLTEXT01.pdf

  11. Purmonen, S.: Predicting game level difficulty using deep neural networks. Technical report, KTH, School of Computer Science and Communication (CSC) (2017). http://kth.diva-portal.org/smash/get/diva2:1154062/FULLTEXT01.pdf

  12. Schaul, T., Quan, J., Antonoglou, I., Silver, D.: Prioritized experience replay (2015)

    Google Scholar 

  13. Schulman, J., Levine, S., Abbeel, P., Jordan, M., Moritz, P.: Trust region policy optimization. In: International Conference on Machine Learning, pp. 1889–1897 (2015)

    Google Scholar 

  14. Schulman, J., Wolski, F., Dhariwal, P., Radford, A., Klimov, O.: Proximal policy optimization algorithms, July 2017. http://arxiv.org/abs/1707.06347

  15. Silver, D., et al.: Mastering Chess and Shogi by self-play with a general reinforcement learning algorithm, December 2017. http://arxiv.org/abs/1712.01815

  16. Sutton, R.S., Barto, A.G., et al.: Introduction to Reinforcement Learning, vol. 135. MIT Press, Cambridge (1998)

    MATH  Google Scholar 

  17. Van Hasselt, H., Guez, A., Silver, D.: Deep reinforcement learning with double q-learning. In: Thirtieth AAAI Conference on Artificial Intelligence (2016)

    Google Scholar 

  18. Wang, Z., Schaul, T., Hessel, M., van Hasselt, H., Lanctot, M., de Freitas, N.: Dueling network architectures for deep reinforcement learning, November 2015. http://arxiv.org/abs/1511.06581

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ilya Makarov .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2019 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Kamaldinov, I., Makarov, I. (2019). Deep Reinforcement Learning Methods in Match-3 Game. In: van der Aalst, W., et al. Analysis of Images, Social Networks and Texts. AIST 2019. Lecture Notes in Computer Science(), vol 11832. Springer, Cham. https://doi.org/10.1007/978-3-030-37334-4_5

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-37334-4_5

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-37333-7

  • Online ISBN: 978-3-030-37334-4

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics