Abstract
Anxiety is one of the most basic emotional states and also the most common disorder. AI agents however are typically focused on maximising performance, concentrating on expected values and disregarding the degree of exposure to uncertainty. This paper introduces a formalism derived from Partially Observable Markov Decision Processes (POMDPs) to give the first model based on cognitive psychology of the anxiety induced by epistemic uncertainty (i.e. the lack of precision of knowledge about the current state of the world). An algorithm to generate policies balancing reward maximisation and anxiety reduction is given. It is then used on a classical example to demonstrate how this can lead in some cases to a dramatic reduction of epistemic uncertainty for nearly no cost and thus a more human-friendly reward optimisation. The empirical validation shows results reminiscent of behaviours that cognitive psychology identifies as coping mechanisms to anxiety.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
References
Araya, M., Buffet, O., Thomas, V., Charpillet, F.: A pomdp extension with belief-dependent rewards. In: Advances in Neural Information Processing Systems, vol. 23 (2010)
Ayer, T., Alagoz, O., Stout, N.K.: Or forum-a POMDP approach to personalize mammography screening decisions. Oper. Res. 60(5), 1019–1034 (2012)
Bach, D.R., Dolan, R.J.: Knowing how much you don’t know: a neural organization of uncertainty estimates. Nat. Rev. Neurosci. 13, 572–586 (2012)
Balaban, E., Arnon, T., Shirley, M.H., Brisson, S.F., Gao, A.: A system health aware pomdp framework for planetary rover traverse evaluation and refinement. In: 2018 AIAA Information Systems-AIAA Infotech @ Aerospace (2018)
Barlow, D.: Anxiety and Its Disorders: The Nature and Treatment of Anxiety and Panic. Anxiety and Its Disorders: The Nature and Treatment of Anxiety and Panic, Guilford Publications (2004)
Beck, A.T., Clark, D.A.: An information processing model of anxiety: automatic and strategic processes. Behav. Res. Ther. 35(1), 49–58 (1997)
Bravo, R.Z.B., Leiras, A., Cyrino Oliveira, F.L.: The use of UAVs in humanitarian relief: an application of POMDP-based methodology for finding victims. Prod. Oper. Manage. 28(2), 421–440 (2019)
Broekens, J., Jacobs, E., Jonker, C.M.: A reinforcement learning model of joy, distress, hope and fear. Connect. Sci. 27(3), 215–233 (2015)
Cassandra, A.R.: A survey of POMDP applications. In: Working Notes of AAAI Fall Symposium on Planning with Partially Observable Markov Decision Processes (1998)
Cassandra, A.R., Kaelbling, L.P., Littman, M.L.: Acting optimally in partially observable stochastic domains. In: Proceedings of the Twelfth AAAI National Conference on Artificial Intelligence, AAAI 1994, pp. 1023–1028. AAAI Press (1994)
Çelik, M., Ergun, O., Keskinocak, P.: The post-disaster debris clearance problem under incomplete information. Oper. Res. 63(1), 65–85 (2015)
Corotis, R., Ellis, H., Jiang, M.: Modeling of risk-based inspection, maintenance and life-cycle cost with partially observable markov decision processes. Struct. Infrastruct. Eng. 1, 75–84 (2005)
Duke, K.E., Goldsmith, K., Amir, O.: Is the preference for certainty always so certain? J. Assoc. Consum. Res. 3(1), 63–80 (2018)
Institute of Health Metrics and Evaluation, Global Health Data Exchange. www.vizhub.healthdata.org/gbd-results/, Accessed 14 Mai 2022
Ibekwe, H., Kamrani, A.: Navigation for autonomous robots in partially observable facilities. In: 2014 World Automation Congress, pp. 1–5 (2014)
Kaelbling, L.P., Littman, M.L., Cassandra, A.R.: Planning and acting in partially observable stochastic domains. Artif. Intell. 101(1–2), 99–134 (1998)
Mafi, N., Abtahi, F., Fasel, I.: Information theoretic reward shaping for curiosity driven learning in POMDPs. In: 2011 IEEE International Conference on Development and Learning (ICDL), vol. 2, pp. 1–7 (2011)
Marecki, J., Varakantham, P.: Risk-sensitive planning in partially observable environments. In: Proceedings of the 9th International Conference on Autonomous Agents and Multiagent Systems, AAMAS. 2, pp. 1357–1368 (2010)
Miceli, M., Castelfranchi, C.: Anxiety as an “epistemic” emotion: an uncertainty theory of anxiety. Anxiety Stress Coping 18, 291–319 (2005)
Niroui, F., Sprenger, B., Nejat, G.: Robot exploration in unknown cluttered environments when dealing with uncertainty. In: 2017 IEEE International Symposium on Robotics and Intelligent Sensors (IRIS), pp. 224–229. IEEE (2017)
Geneva: World Health Organization: Mental health and covid-19: Early evidence of the pandemic’s impact (2022)
Osmanağaoğlu, N., Creswell, C., Dodd, H.F.: Intolerance of uncertainty, anxiety, and worry in children and adolescents: a meta-analysis. J. Affect. Disord. 225, 80–90 (2018)
Papakonstantinou, K., Shinozuka, M.: Planning structural inspection and maintenance policies via dynamic programming and Markov processes. Part II: POMDP implementation. Reliab. Eng. Syst. Saf. 130, 214–224 (2014)
pomdp.org: Examples. www.pomdp.org/examples/, Accessed 26 Jun 2022
Roy, N., Gordon, G., Thrun, S.: Finding approximate POMDP solutions through belief compression. J. Artif. Intell. Res. 23, 1–40 (2005)
Silver, D., Veness, J.: Monte-carlo planning in large pomdps. In: Lafferty, J., Williams, C., Shawe-Taylor, J., Zemel, R., Culotta, A. (eds.) Advances in Neural Information Processing Systems, vol. 23. Curran Associates, Inc. (2010)
Thomas, V., Hutin, G., Buffet, O.: Monte Carlo information- oriented planning (2021). www.arxiv.org/abs/2103.11345
Vanhée, L., Jeanpierre, L., Mouaddib, A.I.: Anxiety-sensitive planning: from formal foundations to algorithms and applications. In: Proceedings of the International Conference on Automated Planning and Scheduling, vol. 32, pp. 730–740 (2022)
Wang, Q.A.: Probability distribution and entropy as a measure of uncertainty. J. Phys. A: Math. Theor. 41(6), 065004 (2008)
Wu, F., Ramchurn, S.D., Chen, X.: Coordinating human-UAV teams in disaster response. In: Proceedings of the 25th International Joint Conference on Artificial Intelligence (IJCAI), pp. 524–530 (2016)
Zhang, W., Wang, H.: Diagnostic policies optimization for chronic diseases based on POMDP model. In: Healthcare (2022)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Gutsche, L., Vanhée, L. (2024). The Value of Knowledge: Joining Reward and Epistemic Certainty Optimisation for Anxiety-Sensitive Planning. In: Amigoni, F., Sinha, A. (eds) Autonomous Agents and Multiagent Systems. Best and Visionary Papers. AAMAS 2023. Lecture Notes in Computer Science(), vol 14456. Springer, Cham. https://doi.org/10.1007/978-3-031-56255-6_2
Download citation
DOI: https://doi.org/10.1007/978-3-031-56255-6_2
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-56254-9
Online ISBN: 978-3-031-56255-6
eBook Packages: Computer ScienceComputer Science (R0)