Abstract
This paper presents the current state of the online game Iconoscope and analyzes the data collected from almost 45 months of continuous operation. Iconoscope is a freeform creation game which aims to foster the creativity of its users through diagrammatic lateral thinking, as users are required to depict abstract concepts as icons which may be misinterpreted by other users as different abstract concepts. From users’ responses collected from an online gallery of all icons drawn with Iconoscope, we collect a corpus of over 500 icons which contain annotations of visual appeal. Several machine learning algorithms are tested for their ability to predict the appeal of an icon from its visual appearance and other properties. Findings show the impact of the representation on the model’s accuracy and highlight how such a predictive model of quality can be applied to evaluate new icons (human-authored or generated).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Ballard, D.H., Hinton, G.E., Sejnowski, T.J.: Parallel visual computation. Nature 306(5938), 21 (1983)
Cachia, R., et al.: Creativity in schools in Europe: a survey of teachers (2009). http://ipts.jrc.ec.europa.eu/publications/pub.cfm?id=2940. Accessed Nov 2016
Clevert, D.A., Unterthiner, T., Hochreiter, S.: Fast and accurate deep network learning by exponential linear units (ELUs). arXiv preprint arXiv:1511.07289 (2015)
De Bono, E.: Lateral Thinking: Creativity Step by Step. Harper Collins (2010)
Goodfellow, I., Bengio, Y., Courville, A.: Deep Learning. MIT press (2016)
Ioffe, S., Szegedy, C.: Batch normalization: accelerating deep network training by reducing internal covariate shift. arXiv preprint arXiv:1502.03167 (2015)
Jeffrey, B., Craft, A.: Teaching creatively and teaching for creativity: distinctions and relationships. Educ. Stud. 30(1), 77–87 (2004)
Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. In: Advances in Neural Information Processing Systems (2012)
LeCun, Y., Bengio, Y., et al.: Convolutional networks for images, speech, and time series. Handb. Brain Theory Neural Netw. 3361(10), 1995 (1995)
Correia, J., Ciesielski, V., Liapis, A. (eds.): EvoMUSART 2017. LNCS, vol. 10198. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-55750-2
Liapis, A., Hoover, A.K., Yannakakis, G.N., Alexopoulos, C., Dimaraki, E.V.: Motivating visual interpretations in iconoscope: designing a game for fostering creativity. In: Proceedings of the Conference on the Foundations of Digital Games (2015)
Liapis, A., Yannakakis, G.N., Alexopoulos, C., Lopes, P.: Can computers foster human users’ creativity? theory and praxis of mixed-initiative co-creativity. Digit. Cult. Educ. (DCE) 8(2), 136–152 (2016)
Makantasis, K., Doulamis, A., Doulamis, N., Psychas, K.: Deep learning based human behavior recognition in industrial workflows. In: Proceedings of International Conference on Image Processing, pp. 1609–1613. IEEE (2016)
Park, E., Han, X., Berg, T.L., Berg, A.C.: Combining multiple sources of knowledge in deep CNNS for action recognition. In: Proceedings of the Winter Conference on Applications of Computer Vision (WACV), pp. 1–8. IEEE (2016)
Plato, C.D.: The Collected Dialogues. Princeton University Press (1961)
Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: unified, real-time object detection. In: Procedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 779–788 (2016)
Sawyer, K.: Educating for innovation. Thinking Skills Creativity 1, 41–48 (2006)
Scaltsas, T., Alexopoulos, C.: Creating creativity through emotive thinking. In: Proceedings of the World Congress of Philosophy (2013)
Simonyan, K., Zisserman, A.: Two-stream convolutional networks for action recognition in videos. In: Advances in Neural Information Processing Systems (2014)
Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)
Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15(1), 1929–1958 (2014)
Srivastava, N., Salakhutdinov, R.R.: Multimodal learning with deep boltzmann machines. In: Advances in Neural Information Processing Systems (2012)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Liapis, A., Gravina, D., Kastbjerg, E., Yannakakis, G.N. (2019). Modelling the Quality of Visual Creations in Iconoscope. In: Liapis, A., Yannakakis, G., Gentile, M., Ninaus, M. (eds) Games and Learning Alliance. GALA 2019. Lecture Notes in Computer Science(), vol 11899. Springer, Cham. https://doi.org/10.1007/978-3-030-34350-7_13
Download citation
DOI: https://doi.org/10.1007/978-3-030-34350-7_13
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-34349-1
Online ISBN: 978-3-030-34350-7
eBook Packages: Computer ScienceComputer Science (R0)