Skip to main content

Enhance Performance of Action Evaluation Functions with Stochastic Optimization Algorithms

  • Conference paper
  • First Online:
Context-Aware Systems and Applications (ICCASA 2016)

Abstract

In this paper, we describe how to optimize the weights of board cells from data set of game records, the weights of board cells are applied in the action evaluation function which usually uses to enhance Monte Carlo Tree Search programs. The general optimization process is introduced and discussed, and one specific method is implemented. We use Othello as a testing environment, and experiment results is better if the action evaluation function is better.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Araki, N., Yoshida, K., Tsuruoka, Y., Tsujii, J.: Move prediction in Go with the maximum entropy method. In: Proceedings of the IEEE Symposium on Computational Intelligence and Games (2007)

    Google Scholar 

  2. Stern, D., Herbrich, R., Graepel, T.: Bayesian pattern ranking for move prediction in the game of Go. In: Proceedings of the 23rd international conference on Machine learning, Pittsburgh, pp. 873–880 (2006)

    Google Scholar 

  3. Coulom, R.: Computing Elo ratings of move patterns in the game of Go. In: Computer Games Workshop, Amsterdam, Netherlands (2007)

    Google Scholar 

  4. http://skatgame.net/mburo/ggs/game-archive/Othello (2012)

  5. Werf, E., Uiterwijk, J.W.H.M., Postma, E., Herik, J.: Local move prediction in Go. In: Schaeffer, J., Müller, M., Björnsson, Y. (eds.) CG 2002. LNCS, vol. 2883, pp. 393–412. Springer, Heidelberg (2003). doi:10.1007/978-3-540-40031-8_26

    Chapter  Google Scholar 

  6. Cazenave, T.: Automatic acquisition of tactical go rules. In: 3rd Game Programming Workshop in Hakone, Japan, pp. 10–19 (1996)

    Google Scholar 

  7. Chaslot, G., Bakkes, S., Szita, I., Spronck, P.: Monte-Carlo tree search: a new framework for game AI. AIIDE 2008

    Google Scholar 

  8. http://www.apld.co.uk/riscworld/volume3/issue5/agrm/chap09.htm

  9. Buro, M.: From simple features to sophisticated evaluation functions. In: Herik, H.J., Iida, H. (eds.) CG 1998. LNCS, vol. 1558, pp. 126–145. Springer, Heidelberg (1999). doi:10.1007/3-540-48957-6_8

    Chapter  Google Scholar 

  10. Ikeda, K., Viennot, S.: Efficiency of static knowledge bias in monte-carlo tree search. Computers and Games 2013 (2013)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Nguyen Quoc Huy .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2017 ICST Institute for Computer Sciences, Social Informatics and Telecommunications Engineering

About this paper

Cite this paper

Huy, N.Q., Nam, D.D., Quoc, D.C. (2017). Enhance Performance of Action Evaluation Functions with Stochastic Optimization Algorithms. In: Cong Vinh, P., Tuan Anh, L., Loan, N., Vongdoiwang Siricharoen, W. (eds) Context-Aware Systems and Applications. ICCASA 2016. Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering, vol 193. Springer, Cham. https://doi.org/10.1007/978-3-319-56357-2_19

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-56357-2_19

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-56356-5

  • Online ISBN: 978-3-319-56357-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics