Skip to main content
Log in

Agent development as a strategy shaper

  • Published:
Autonomous Agents and Multi-Agent Systems Aims and scope Submit manuscript

Abstract

This paper studies to what extent agent development changes one’s own strategy. While this question has many general implications it is of special interest to the study of peer designed agents (PDAs), which are computer agents developed by non-experts. This latter emerging technology, has been widely advocated in recent literature for the purpose of replacing people in simulations and investigating human behavior. Its main premise is that strategies programmed into these agents reliably reflect, to some extent, the behavior used by their programmers in real life. We show that PDA development has an important side effect that has not been addressed to date—the process that merely attempts to capture one’s strategy is also likely to affect the developer’s strategy. This result has many implications concerning the appropriate design of PDA-based simulations as well as the validity of using PDAs for studying individual decision making. The phenomenon is demonstrated experimentally, using two very different application domains and several performance measures. Our findings suggest that the effects on one’s strategy arise both in situations where it is potentially possible for people to reason about the optimal strategy (in which case PDA development will enhance the use of an optimal strategy) and in those where calculating the optimal strategy is computationally challenging (in which case PDA development will push people to use more effective strategies, on average). Since in our experiments PDA development actually improved the developer’s strategy, PDA development can be suggested as a means for improving people’s problem solving skills. Finally, we show that the improvement achieved in people’s strategies through agent development is not attributed to the expressive aspect of agent development per-se but rather there is a crucial additional gain in the process of designing and programming ones strategy into an agent.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15
Fig. 16

Similar content being viewed by others

Notes

  1. The inefficiency improvement measures the decrease (in percentages) in the difference between the average result achieved and the theoretical bound (6 in the case of outcome performance), as the difference between the two represents the strategy’s inefficiency.

  2. This kind of analysis for the outcome performance measure is futile, since this measure, when taken individually, highly depends on chance.

  3. The idea was to equalize conditions, so that people would have the chance to thoroughly think through their stated strategy, much like how PDA development takes some time. people will have the chance to thoroughly think through their stated strategy.

  4. For a comparison between AMT and other recruitment methods see [14].

  5. Note that better performance of the adversary (the human players) infers worse performance of the patrolling strategy.

References

  1. Azaria, A., Rabinovich, Z., Kraus, S., & Goldman, C. V. (2013). Strategic information disclosure to people with multiple alternatives. Transactions on intelligent systems and technology (TIST) ACM journal.

  2. Azaria, A., Richardson, A., & Kraus, S. (2014). An Agent for the Prospect Presentation Problem. AAMAS.

  3. Elmalech, A., & Sarne, D. (2012). Evaluating the applicability of peer-designed agents in mechanisms evaluation. In Proceedings of IAT (pp. 374–381). New York: MIT Press.

  4. Lin, R., Kraus, S., Agmon, N., Barrett, S., & Stone, P. (2011). Comparing agents’ success against people in security domains. In Proceedings of AAAI conference on artificial intelligence, San Francisco.

  5. Chalamish, M., Sarne, D., & Lin, R. (2012). Enhancing parking simulations using peer-designed agents. ITS, 14(1), 1–7.

    Google Scholar 

  6. Grosz, B., Kraus, S., Talman, S., Stossel, B., & Havlin, M. (2004). The influence of social dependencies on decision-making: Initial investigations with a new game. In Proceedings of AAMAS (pp. 782–789). New York,. NY: ACM Press.

  7. Azaria, A., Richardson, A., Elmalech, A., & Rosenfeld, A. (2014). Automated agents’ behavior in the trust-revenge game in comparison to other cultures. In Proceedings of the 2014 international conference on AAMAS, Saint Paul.

  8. Shin, J., & Ariely, D. (2004). Keeping doors open: The effect of unavailability on incentives to keep options viable. Management Science, 50(5), 575–586.

    Article  Google Scholar 

  9. Agmon, N., Kraus, S., Kaminka, G., & Sadov, V. (2009). Adversarial uncertainty in multi-robot patrol. In Proceedings of IJCAI, Milan.

  10. Auer, P., Cesa-Bianchi, N., Freund, Y., & Schapire, R.E. (1995). Gambling in a rigged casino: The adversarial multi-armed bandit problem. In Proceedings of the 36th annual symposium on foundations of computer science (pp. 322–331). Los Alamitos: IEEE Computer Society Press.

  11. Elmalech, A., Sarne, D., & Grosz, B. J. (2015). Problem restructuring for better decision making in recurring decision situations. Autonomous Agents and Multi-Agent Systems, 29(1), 1–39.

    Article  Google Scholar 

  12. MOSS a system for detecting software plagiarism. (2015). Accessed May 6, 2015 from https://theory.stanford.edu/~aiken/moss/.

  13. AMT: Amazon mechanical turk. (2010). http://www.mturk.com.

  14. Paolacci, G., Chandler, J., & Ipeirotis, P. (2010). Running experiments on amazon mechanical turk. Judgment and Decision Making, 5(5), 411–419.

    Google Scholar 

  15. Agmon, N., Kraus, S., & Kaminka, G. (2011). Multi-robot adversarial patrolling: Facing a full-knowledge opponent. JAIR, 42, 887–916.

    MathSciNet  MATH  Google Scholar 

  16. Agmon, N., Sadov, V., Kaminka, G., & Kraus, S. (2008). The impact of adversarial knowledge on adversarial planning in perimeter patrol. In Proceedings of AAMAS (pp. 55–62), Estoril.

  17. Bosse, T., Gerritsen, C., Hoogendoorn, M., Jaffry, S., & Treur, J. (2011). Agent-based vs. population-based simulation of displacement of crime: A comparative study. WIAS, 9(2), 147–160.

    Google Scholar 

  18. Zhang, Y., Biggers, K., He, L., Reddy, S., Sepulvado, D., Yen, J., & Ioerger, T. (2001). A distributed intelligent agent architecture for simulating aggregate-level behavior and interactions on the battlefield. In Proceedings of Fifth World Multi-Conference on SCI (pp. 58–63), Orlando.

  19. Takahashi, T., Tadokoro, S., Ohta, M., & Ito, N. (2002). Agent based approach in disaster rescue simulation—from test-bed of multiagent system to practical application. In Proceedings of RoboCup 2001: Robot Soccer World Cup V (pp. 102–111). Berlin, Heidelberg: Springer.

  20. Massaguer, D., Balasubramanian, V., Mehrotra, S., & Venkatasubramanian, N. (2006). Multi-agent simulation of disaster response. In Proceedings of agent technology for disaster management (pp. 124–130), Hakodate.

  21. Musse, S., & Thalmann, D. (2001). Hierarchical model for real time simulation of virtual human crowds. Transactions on Visualization and Computer Graphics, 7(2), 152–164.

    Article  Google Scholar 

  22. Shao, W., & Terzopoulos, D. (2007). Autonomous pedestrians. Graphical Models, 69(5–6), 246–274.

    Article  Google Scholar 

  23. Terzopoulos, D., Tu, X., & Grzeszczuk, R. (1994). Artificial fishes: Autonomous locomotion, perception, behavior, and learning in a simulated physical world. Artificial Life, 1(4), 327–351.

    Article  Google Scholar 

  24. Ulicny, B., & Thalmann, D. (2002). Towards interactive real-time crowd behavior simulation. Computer Graphics Forum, 21(4), 767–775.

    Article  Google Scholar 

  25. Gode, D., & Sunder, S. (1993). Allocative efficiency of markets with zero-intelligence traders: Market as a partial substitute for individual rationality. Journal of Political Economy, 101, 119–137.

    Article  Google Scholar 

  26. Brewer, P., Huang, M., Nelson, B., & Plott, C. (2002). On the behavioral foundations of the law of supply and demand: Human convergence and robot randomness. Experimental Economics, 5(3), 179–208.

    Article  MATH  Google Scholar 

  27. Van Boening, M., & Wilcox, N. (1996). Avoidable cost: Ride a double auction roller coaster. The American Economic Review, 86, 461–477.

    Google Scholar 

  28. Chavez, A., & Maes, P. (1996). Kasbah: An agent marketplace for buying and selling goods. In Proceedings of intelligent agents and multi-agent technology (pp. 8–12). Blackpool: The Practical Application Company Ltd.

  29. Selten, R., Mitzkewitz, M., & Uhlich, G. (1997). Duopoly strategies programmed by experienced players. Econometrica, 65(3), 517–555.

    Article  MathSciNet  MATH  Google Scholar 

  30. Cason, T., & Mui, V. (1998). Social influence in the sequential dictator game. Journal of Mathematical Psychology, 42(2), 248–265.

    Article  MATH  Google Scholar 

  31. Bosch, A., & Silvestre, J. (1999). Does risk aversion or attraction depend on income? An experiment. Economics Letters, 65(3), 265–273.

    Article  MATH  Google Scholar 

  32. Bosch, A., & Silvestre, J. (2006). Risk aversion and embedding bias. Materia (s), 1, 01–2006.

    Google Scholar 

  33. Sonnemans, J. (2000). Decisions and strategies in a sequential search experiment. Journal of Economic Psychology, 21(1), 91–102.

    Article  Google Scholar 

  34. Casari, M., & Cason, T. (2009). The strategy method lowers measured trustworthy behavior. Economics Letters, 103(3), 157–159.

    Article  Google Scholar 

  35. Brandts, J., & Charness, G. (2003). Truth or consequences: An experiment. Management Science, 49(1), 116–130.

    Article  MATH  Google Scholar 

  36. Murphy, R., Rapoport, A., & Parco, J. (2007). Credible signaling in real-time trust dilemmas. Tucson: University of Arizona, Mimeo.

    Google Scholar 

  37. Cheng, K., Zuckerman, I., Nau, D., & Golbeck, J. (2011). The life game: Cognitive strategies for repeated stochastic games. In Proceedings of social computing (socialcom) (pp. 95–102), Boston.

  38. Harries, C., Evans, J., & Dennis, I. (2000). Measuring doctors’ self-insight into their treatment decisions. Applied Cognitive Psychology, 14, 455–477.

    Article  Google Scholar 

  39. Bertrand, M., & Mullainathan, S. (2001). Do people mean what they say? Implications for subjective survey data. American Economic Review, 91(2), 67–72.

    Article  Google Scholar 

  40. Kempen, G., Van Heuvelen, M., Van den Brink, R., Kooijman, A., Klein, M., Houx, P., et al. (1996). Factors affecting contrasting results between self-reported and performance-based levels of physical limitations. Age and Ageing, 25(6), 458–464.

    Article  Google Scholar 

  41. Clements, D. H., & Gullo, D. F. (1984). Effects of computer programming on young children’s cognition. Journal of Educational Psychology, 76(6), 1051.

    Article  Google Scholar 

  42. Feurzeig, W., Papert, S. A., & Lawler, B. (2011). Programming-languages as a conceptual framework for teaching mathematics. Interactive Learning Environments, 19(5), 487–501.

    Article  Google Scholar 

  43. Jeffries, R., Turner, A. A., Polson, P. G., & Atwood, M. E. (1981). The processes involved in designing software. Cognitive skills and their acquisition (pp. 255–283). Hillsdale: Lawrence Erlbaum.

    Google Scholar 

  44. Nickerson, R. (1982). Computer programming as a vehicle for teaching thinking skills. Thinking: The Journal of Philosophy for Children, 4(3/4), 42–48.

    Google Scholar 

  45. Resnick, M., Maloney, J., Rosenbaum, E., Silver, J., Silverman, B., et al. (2009). Scratch: Programming for all. Communications of the ACM, 52(11), 60–67.

    Article  Google Scholar 

  46. Elmalech, A., Sarne, D., & Agmon, N. (2014). Can agent development affect developers strategy? In Proceedings of AAAI (pp. 923–929), Loannina.

  47. Elmalech, A., Sarne, D., & Agmon, N. (2014). Peer designed agents: Just reflect or also affect? In Proceedings of AAMAS (pp. 1429–1430).

Download references

Acknowledgments

We would like to thank Amos Azaria for helping us to collect the data for the no-PDA experiment. A preliminary version of this paper was presented at HAIDM14, and appears in the proceedings of AAAI 2014 [46] and AAMAS 2014 [47].

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Avshalom Elmalech.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Elmalech, A., Sarne, D. & Agmon, N. Agent development as a strategy shaper. Auton Agent Multi-Agent Syst 30, 506–525 (2016). https://doi.org/10.1007/s10458-015-9299-z

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10458-015-9299-z

Keywords

Navigation