Abstract. Incorporating a priori knowledge, such as expert knowledge, metaheuristics, human preferences, and most importantly domain knowledge discovered during evolutionary search, into evolutionary algorithms has gained increasing interest in recent years. In this chapter, we present a method for systematically inserting expert knowledge into evolutionary board game framework at the opening, middle, and endgame stages. In the opening stage, openings defined by the experts are used. In this work, we use speciation techniques to search for diverse strategies that embody different styles of game play and combine them using voting for higher performance. This idea comes from the common knowledge that the combination of diverse well-playing strategies can defeat the best one because they can complement each other for higher performance. Finally, we use an endgame database. Experimental results on checkers and Othello games show that the proposed method is promising to evolve better strategies.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Schaeffer, J.: One Jump Ahead: Challenging Human Supremacy in Checkers. Springer (1997)
Fogel, D.B.: Blondie 24: Playing at the edge of AI. Morgan Kaufmann (2001)
Kim, K.-J. and Cho, S.-B.: Systematically incorporating domain-specific knowl- edge into evolutionary speciated checkers players. IEEE Transactions on Evolu- tionary Computation, 9(6) (2005) 615-627
Kim, K.-J. and Cho, S.-B.: Evolutionary Othello players boosted by opening knowledge. World Congress on Computational Intelligence (2006)
Fogel, D.B., Hays, T.J., Hahn, S. and Quon, J.: A self-learning evolutionary chess program. Proc. of the IEEE 92(12) (2004) 1947-1954
Chernev, I.: The Compleat Draughts Player. Oxford University Press, (1981)
Schaeffer, J., Culberson, J., Treloar, N., Knight, B., Lu, P. and Szafron, D.: A world championship caliber checkers program. Artificial Intelligence 53(2-3) (1992)273-290
Schaeffer, J., Lake, R., Lu, P. and Bryant, M.: Chinook: The man-machine world checkers champion. AI Magazine 17(1) (1996) 21-29
R. Lake, J. Schaeffer and P. Lu, “Solving large retrograde analysis problems using a network of workstations,” Proc. of Advances in Computer Chess VII, pp. 135-162, 1994.
Fogel, D.B.: Evolutionary entertainment with intelligent agents. IEEE Com- puter, 36(6) (2003) 106-108
Chellapilla, K. and Fogel, D.B.: Evolving neural networks to play checkers with- out relying on expert knowledge. IEEE Trans. on Neural Networks 10(6) (1999) 1382-1391
Chellapilla, K. and Fogel, D.B.: Evolving an expert checkers playing program without using human expertise. IEEE Trans. on Evolutionary Computation 5(4) (2001)422-428
Fogel, D.B.: Evolving a checkers player without relying on human experience. ACM Intelligence, 11(2) (2000) 20-27
Chellapilla, K. and Fogel, D.B.: Anaconda defeats Hoyle 6-0: A case study com- peting an evolved checkers program against commercially available software. Proc. of the 2000 Congress on Evolutionary Computation 2 (2000) 857-863
Fogel, D.B. and Chellapilla, K.: Verifying Anaconda’s expert rating by com- peting against Chinook: Experiments in co-evolving a neural checkers player. Neurocomputing 42(1-4) (2002) 69-86
Chong, S.Y., Tan, M.K. and White, J.D.: Observing the evolution of neural net- works learning to play the game of Othello. IEEE Transactions on Evolutionary Computation 9(3) (2005) 240-251
Sun, C.-T. and Wu, M.-D.: Multi-stage genetic algorithm learning in game play- ing. NAFIPS/IFIS/NASA ’94 (1994) 223-227
Sun, C.-T. and Wu, M.-D.: Self-adaptive genetic algorithm learning in game playing. IEEE International Conference on Evolutionary Computation 2 (1995) 814-818
Moriarty, D.E. and Miikkulainen, R.: Discovering complex Othello strategies through evolutionary neural networks. Connection Science 7 (1995) 195-209
Moriarty, D. E. and Miikkulainen, R.: Evolving neural networks to focus mini- max search. Proc. of the 12th National Conf. on Artificial Intelligence (AAAI-94) (1994)1371-1377
Moriarty, D.E. and Miikkulainen, R.: Improving game-tree search with evolu- tionary neural networks. Proc. of the First IEEE Conf. on Evolutionary Computation, 1 (1994) 496-501
Richards, N., Moriarty, D. and Miikkulainen, R.: Evolving neural networks to play go. Applied Intelligence 8 (1998) 85-96
Stanley, K.O. and Miikkulainen, R.: Evolving a roving eye for go. Proc. of the Genetic and Evolutionary Computation Conference (GECCO-2004) (2004) 1226-1238
Lubberts, A. and Miikkulainen, R.: Co-evolving a go-playing neural networks. Proc. of 2001 Genetic and Evolutionary Computation Conference Workshop Program (GECCO-2001), (2001) 14-19
Perez-Bergquist, A.S.: Applying ESP and region specialists to neuro-evolution for go. Technical Report CSTR01-24, Department of Computer Science, The University of Texas at Austin, May (2001)
Barone, L. and While, L.:An adaptive learning model for simplified poker using evolutionary algorithms. Proc. of the 1999 Congress on Evolutionary Computation 1 (1999) 153-160
Pollack, J.B. and Blair, A.D.: Co-evolution in the successful learning of backgammon strategy. Machine Learning 32(3) (1998) 225-240
Kendall, G. and Whitwell, G.: An evolutionary approach for the tuning of a chess evaluation function using population dynamics. Proc. of the 2001 Congress on Evolutionary Computation 2 (2001) 995-1002
Kendall, G. and Smith, C.: The evolution of blackjack strategies. Proc. of the 2003 Congress on Evolutionary Computation 4 (2003) 2474-2481
Fogel, D.B.: Evolving strategies in blackjack. Proc. of the 2004 Congress on Evolutionary Computation (2004) 1427-1434
Ono, W. and Lim, Y.-J.: An investigation on piece differential information in co-evolution on games using Kalah. Proc. of Congress on Evolutionary Compu- tation 3 (2003) 1632-1638
Lucas, S.M. and Kendall, G.: Evolutionary computation and Games. IEEE Com- putational Intelligence Magazine (2006) 10-18
Chong, S.Y., Ku, D.C., Lim, H.S., Tan, M.K. and White, J.D.: Evolved neural networks learning Othello strategies. The 2003 Congress on Evolutionary Computation 3 (2003) 2222-2229
Sun, C.-T., Liao, Y.-H., Lu, J.-Y. and Zheng, F.-M.: Genetic algorithm learning in game playing with multiple coaches. IEEE World Congress on Computational Intelligence 1 (1994) 239-243
Alliot, J. and Durand, N.: A genetic algorithm to improve an Othello program. Artificial Evolution (1995) 307-319
Eskin, E. and Siegel, E.: Genetic programming applied to Othello: Introducing students to machine learning research. Proceedings of the Thirtieth SIGCSE Technical Symposium on Computer Science Education, 31(1) (1999) 242-246
Bateman, R.: Training a multi-layer feedforward neural network to play Othello using the backpropagation algorithm and reinforcement learning. Journal of Computing Sciences in Colleges 19(5) (2004) 296-297
Smith, R.E. and Gray, B.: Co-adaptive genetic algorithms: An example in Oth- ello Strategy. TCGA Report No. 94002, The University of Alabama (1994)
Leuski, A. and Utgoff, P.E.: What a neural network can learn about Othello. Technical Report TR96-10, Department of Computer Science, University of Massachusetts, Amherst (1996)
Leuski, A.: Learning of position evaluation in the game of Othello. Technical Report TR 95-23, Department of Computer Science, University of Massa- chusetts, Amherst (1995)
Moriarty, D.E., Schultz, A.C. and Grefenstette, J.J.: Evolutionary algorithms for reinforcement learning. Journal of Artificial Intelligence Research 11 (1999) 241-276
Hansen, L.K., and Salamon, P.: Neural network ensembles. IEEE Trans. on Pattern Analysis and Machine Intelligence 12(10) (1990) 993-1001
Kim, K.-J. and Cho, S.-B.: Evolving speciated checker players with crowding algorithm. Congress on Evolutionary Computation (2002) 407-412
CEC 2006 Othello Competition, http://algoval.essex.ac.uk:8080/othello/html/Othello.html.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2007 Springer-Verlag Berlin Heidelberg
About this chapter
Cite this chapter
Kim, KJ., Cho, SB. (2007). Evolutionary Algorithms for Board Game Players with Domain Knowledge. In: Baba, N., Jain, L.C., Handa, H. (eds) Advanced Intelligent Paradigms in Computer Games. Studies in Computational Intelligence, vol 71. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-72705-7_4
Download citation
DOI: https://doi.org/10.1007/978-3-540-72705-7_4
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-72704-0
Online ISBN: 978-3-540-72705-7
eBook Packages: EngineeringEngineering (R0)