Abstract
This paper studies to what extent agent development changes one’s own strategy. While this question has many general implications it is of special interest to the study of peer designed agents (PDAs), which are computer agents developed by non-experts. This latter emerging technology, has been widely advocated in recent literature for the purpose of replacing people in simulations and investigating human behavior. Its main premise is that strategies programmed into these agents reliably reflect, to some extent, the behavior used by their programmers in real life. We show that PDA development has an important side effect that has not been addressed to date—the process that merely attempts to capture one’s strategy is also likely to affect the developer’s strategy. This result has many implications concerning the appropriate design of PDA-based simulations as well as the validity of using PDAs for studying individual decision making. The phenomenon is demonstrated experimentally, using two very different application domains and several performance measures. Our findings suggest that the effects on one’s strategy arise both in situations where it is potentially possible for people to reason about the optimal strategy (in which case PDA development will enhance the use of an optimal strategy) and in those where calculating the optimal strategy is computationally challenging (in which case PDA development will push people to use more effective strategies, on average). Since in our experiments PDA development actually improved the developer’s strategy, PDA development can be suggested as a means for improving people’s problem solving skills. Finally, we show that the improvement achieved in people’s strategies through agent development is not attributed to the expressive aspect of agent development per-se but rather there is a crucial additional gain in the process of designing and programming ones strategy into an agent.
Similar content being viewed by others
Notes
The inefficiency improvement measures the decrease (in percentages) in the difference between the average result achieved and the theoretical bound (6 in the case of outcome performance), as the difference between the two represents the strategy’s inefficiency.
This kind of analysis for the outcome performance measure is futile, since this measure, when taken individually, highly depends on chance.
The idea was to equalize conditions, so that people would have the chance to thoroughly think through their stated strategy, much like how PDA development takes some time. people will have the chance to thoroughly think through their stated strategy.
For a comparison between AMT and other recruitment methods see [14].
Note that better performance of the adversary (the human players) infers worse performance of the patrolling strategy.
References
Azaria, A., Rabinovich, Z., Kraus, S., & Goldman, C. V. (2013). Strategic information disclosure to people with multiple alternatives. Transactions on intelligent systems and technology (TIST) ACM journal.
Azaria, A., Richardson, A., & Kraus, S. (2014). An Agent for the Prospect Presentation Problem. AAMAS.
Elmalech, A., & Sarne, D. (2012). Evaluating the applicability of peer-designed agents in mechanisms evaluation. In Proceedings of IAT (pp. 374–381). New York: MIT Press.
Lin, R., Kraus, S., Agmon, N., Barrett, S., & Stone, P. (2011). Comparing agents’ success against people in security domains. In Proceedings of AAAI conference on artificial intelligence, San Francisco.
Chalamish, M., Sarne, D., & Lin, R. (2012). Enhancing parking simulations using peer-designed agents. ITS, 14(1), 1–7.
Grosz, B., Kraus, S., Talman, S., Stossel, B., & Havlin, M. (2004). The influence of social dependencies on decision-making: Initial investigations with a new game. In Proceedings of AAMAS (pp. 782–789). New York,. NY: ACM Press.
Azaria, A., Richardson, A., Elmalech, A., & Rosenfeld, A. (2014). Automated agents’ behavior in the trust-revenge game in comparison to other cultures. In Proceedings of the 2014 international conference on AAMAS, Saint Paul.
Shin, J., & Ariely, D. (2004). Keeping doors open: The effect of unavailability on incentives to keep options viable. Management Science, 50(5), 575–586.
Agmon, N., Kraus, S., Kaminka, G., & Sadov, V. (2009). Adversarial uncertainty in multi-robot patrol. In Proceedings of IJCAI, Milan.
Auer, P., Cesa-Bianchi, N., Freund, Y., & Schapire, R.E. (1995). Gambling in a rigged casino: The adversarial multi-armed bandit problem. In Proceedings of the 36th annual symposium on foundations of computer science (pp. 322–331). Los Alamitos: IEEE Computer Society Press.
Elmalech, A., Sarne, D., & Grosz, B. J. (2015). Problem restructuring for better decision making in recurring decision situations. Autonomous Agents and Multi-Agent Systems, 29(1), 1–39.
MOSS a system for detecting software plagiarism. (2015). Accessed May 6, 2015 from https://theory.stanford.edu/~aiken/moss/.
AMT: Amazon mechanical turk. (2010). http://www.mturk.com.
Paolacci, G., Chandler, J., & Ipeirotis, P. (2010). Running experiments on amazon mechanical turk. Judgment and Decision Making, 5(5), 411–419.
Agmon, N., Kraus, S., & Kaminka, G. (2011). Multi-robot adversarial patrolling: Facing a full-knowledge opponent. JAIR, 42, 887–916.
Agmon, N., Sadov, V., Kaminka, G., & Kraus, S. (2008). The impact of adversarial knowledge on adversarial planning in perimeter patrol. In Proceedings of AAMAS (pp. 55–62), Estoril.
Bosse, T., Gerritsen, C., Hoogendoorn, M., Jaffry, S., & Treur, J. (2011). Agent-based vs. population-based simulation of displacement of crime: A comparative study. WIAS, 9(2), 147–160.
Zhang, Y., Biggers, K., He, L., Reddy, S., Sepulvado, D., Yen, J., & Ioerger, T. (2001). A distributed intelligent agent architecture for simulating aggregate-level behavior and interactions on the battlefield. In Proceedings of Fifth World Multi-Conference on SCI (pp. 58–63), Orlando.
Takahashi, T., Tadokoro, S., Ohta, M., & Ito, N. (2002). Agent based approach in disaster rescue simulation—from test-bed of multiagent system to practical application. In Proceedings of RoboCup 2001: Robot Soccer World Cup V (pp. 102–111). Berlin, Heidelberg: Springer.
Massaguer, D., Balasubramanian, V., Mehrotra, S., & Venkatasubramanian, N. (2006). Multi-agent simulation of disaster response. In Proceedings of agent technology for disaster management (pp. 124–130), Hakodate.
Musse, S., & Thalmann, D. (2001). Hierarchical model for real time simulation of virtual human crowds. Transactions on Visualization and Computer Graphics, 7(2), 152–164.
Shao, W., & Terzopoulos, D. (2007). Autonomous pedestrians. Graphical Models, 69(5–6), 246–274.
Terzopoulos, D., Tu, X., & Grzeszczuk, R. (1994). Artificial fishes: Autonomous locomotion, perception, behavior, and learning in a simulated physical world. Artificial Life, 1(4), 327–351.
Ulicny, B., & Thalmann, D. (2002). Towards interactive real-time crowd behavior simulation. Computer Graphics Forum, 21(4), 767–775.
Gode, D., & Sunder, S. (1993). Allocative efficiency of markets with zero-intelligence traders: Market as a partial substitute for individual rationality. Journal of Political Economy, 101, 119–137.
Brewer, P., Huang, M., Nelson, B., & Plott, C. (2002). On the behavioral foundations of the law of supply and demand: Human convergence and robot randomness. Experimental Economics, 5(3), 179–208.
Van Boening, M., & Wilcox, N. (1996). Avoidable cost: Ride a double auction roller coaster. The American Economic Review, 86, 461–477.
Chavez, A., & Maes, P. (1996). Kasbah: An agent marketplace for buying and selling goods. In Proceedings of intelligent agents and multi-agent technology (pp. 8–12). Blackpool: The Practical Application Company Ltd.
Selten, R., Mitzkewitz, M., & Uhlich, G. (1997). Duopoly strategies programmed by experienced players. Econometrica, 65(3), 517–555.
Cason, T., & Mui, V. (1998). Social influence in the sequential dictator game. Journal of Mathematical Psychology, 42(2), 248–265.
Bosch, A., & Silvestre, J. (1999). Does risk aversion or attraction depend on income? An experiment. Economics Letters, 65(3), 265–273.
Bosch, A., & Silvestre, J. (2006). Risk aversion and embedding bias. Materia (s), 1, 01–2006.
Sonnemans, J. (2000). Decisions and strategies in a sequential search experiment. Journal of Economic Psychology, 21(1), 91–102.
Casari, M., & Cason, T. (2009). The strategy method lowers measured trustworthy behavior. Economics Letters, 103(3), 157–159.
Brandts, J., & Charness, G. (2003). Truth or consequences: An experiment. Management Science, 49(1), 116–130.
Murphy, R., Rapoport, A., & Parco, J. (2007). Credible signaling in real-time trust dilemmas. Tucson: University of Arizona, Mimeo.
Cheng, K., Zuckerman, I., Nau, D., & Golbeck, J. (2011). The life game: Cognitive strategies for repeated stochastic games. In Proceedings of social computing (socialcom) (pp. 95–102), Boston.
Harries, C., Evans, J., & Dennis, I. (2000). Measuring doctors’ self-insight into their treatment decisions. Applied Cognitive Psychology, 14, 455–477.
Bertrand, M., & Mullainathan, S. (2001). Do people mean what they say? Implications for subjective survey data. American Economic Review, 91(2), 67–72.
Kempen, G., Van Heuvelen, M., Van den Brink, R., Kooijman, A., Klein, M., Houx, P., et al. (1996). Factors affecting contrasting results between self-reported and performance-based levels of physical limitations. Age and Ageing, 25(6), 458–464.
Clements, D. H., & Gullo, D. F. (1984). Effects of computer programming on young children’s cognition. Journal of Educational Psychology, 76(6), 1051.
Feurzeig, W., Papert, S. A., & Lawler, B. (2011). Programming-languages as a conceptual framework for teaching mathematics. Interactive Learning Environments, 19(5), 487–501.
Jeffries, R., Turner, A. A., Polson, P. G., & Atwood, M. E. (1981). The processes involved in designing software. Cognitive skills and their acquisition (pp. 255–283). Hillsdale: Lawrence Erlbaum.
Nickerson, R. (1982). Computer programming as a vehicle for teaching thinking skills. Thinking: The Journal of Philosophy for Children, 4(3/4), 42–48.
Resnick, M., Maloney, J., Rosenbaum, E., Silver, J., Silverman, B., et al. (2009). Scratch: Programming for all. Communications of the ACM, 52(11), 60–67.
Elmalech, A., Sarne, D., & Agmon, N. (2014). Can agent development affect developers strategy? In Proceedings of AAAI (pp. 923–929), Loannina.
Elmalech, A., Sarne, D., & Agmon, N. (2014). Peer designed agents: Just reflect or also affect? In Proceedings of AAMAS (pp. 1429–1430).
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Elmalech, A., Sarne, D. & Agmon, N. Agent development as a strategy shaper. Auton Agent Multi-Agent Syst 30, 506–525 (2016). https://doi.org/10.1007/s10458-015-9299-z
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10458-015-9299-z