Abstract
In the last years, we have seen a rise in the use of assistants that are becoming more and more natural in their interactions with people. An emerging characteristic is the proactivity in the assistant interaction. The areas of use for these types of assistants range from health, education, to general assistance for tasks, among others., and the proactivity is usually a means to an end, usually to improve user engagement. Given the growing popularity, we have taken the opportunity in this paper to perform a systematic literature review which focuses on agents with a primary focus on them being proactive. During this, we have observed several interesting patterns, such as the main form of interaction for these agents is through verbal interaction, or the fact that they are usually robots. Many of these papers study user response and feelings to different levels of proactivity, with some defining a time-based proactive response, and other focusing on user involvement when defining proactivity levels. All these findings regarding proactivity make it possible to propose, a model based on the proactivity level and the agent’s ability to learn from each interaction, which is what we are presenting in this paper.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
Sinclair J., 2001, “The Story of Clippy: Microsoft’s Much-Hated Office Assitant”, in https://edition.cnn.com/2001/TECH/ptech/11/26/clippy.obit.idg/index.html. Last accessed: 2023-05-20.
References
Almada, A., Yu, Q., Patel, P.: Proactive chatbot framework based on the ps2clh model: An ai-deep learning chatbot assistant for students, vol. 542. LNNS, pp. 751–770 (2023). https://doi.org/10.1007/978-3-031-16072-1_54
Bremner, P., Dennis, L., Fisher, M., Winfield, A.: On proactive, transparent, and verifiable ethical reasoning for robots, vol. 107(3), pp. 541–561 (2019). https://doi.org/10.1109/JPROC.2019.2898267
Buyukgoz, S., Grosinger, J., Chetouani, M., Saffiotti, A.: Two ways to make your robot proactive: reasoning about human intentions or reasoning about possible futures, vol. 9 (2022). https://doi.org/10.3389/frobt.2022.929267
Chien, S.Y., Lin, Y.L., Chang, B.F.: The effects of intimacy and proactivity on trust in human-humanoid robot interaction (2022). https://doi.org/10.1007/s10796-022-10324-y
Falk, J., Poulakos, S., Kapadia, M., Sumner, R.: Pica: proactive intelligent conversational agent for interactive narratives, pp. 141–146 (2018). https://doi.org/10.1145/3267851.3267892
Faria, G., Silva, T., Abreu, J.: Active and healthy aging: The role of a proactive information assistant embedded on tv, vol. 1597. CCIS, pp. 70–84 (2022). https://doi.org/10.1007/978-3-031-22210-8_5
Garrell, A., Villamizar, M., Moreno-Noguer, F., Sanfeliu, A.: Teaching robot’s proactive behavior using human assistance 9(2), 231–249 (2017). https://doi.org/10.1007/s12369-016-0389-0
Grosinger, J., Pecora, F., Saffiotti, A.: Making robots proactive through equilibrium maintenance. vol. 2016, pp. 3375–3381 (2016)
He, T., Jazizadeh, F., Arpan, L.: Ai-powered virtual assistants nudging occupants for energy saving: proactive smart speakers for hvac control 50(4), 394–409 (2022). https://doi.org/10.1080/09613218.2021.2012119
Iio, T., Satake, S., Kanda, T., Hayashi, K., Ferreri, F., Hagita, N.: Human-like guide robot that proactively explains exhibits 12(2), 549–566 (2020). https://doi.org/10.1007/s12369-019-00587-y
Johnson, E., González, I., Mondéjar, T., Cabañero-Gómez, L., Fontecha, J., Hervás, R.: An affective and cognitive toy to support mood disorders. In: Informatics, vol. 7, p. 48. MDPI (2020)
Ko, J.K., Koo, D., Kim, M.: A novel affinity enhancing method for human robot interaction-preliminary study with proactive docent avatar, vol. 2021, pp. 1007–1011 (2021). https://doi.org/10.23919/ICCAS52745.2021.9649834
Kraus, M., et al.: Was that successful? on integrating proactive meta-dialogue in a diy-assistant using multimodal cues, pp. 585–594 (2020). https://doi.org/10.1145/3382507.3418818
Kraus, M., Wagner, N., Callejas, Z., Minker, W.: The role of trust in proactive conversational assistants 9, 112821–112836 (2021). https://doi.org/10.1109/ACCESS.2021.3103893
Kraus, M., et al.: Kurt: a household assistance robot capable of proactive dialogue, vol. 2022, pp. 855–859 (2022). https://doi.org/10.1109/HRI53351.2022.9889357
Kraus, M., Wagner, N., Untereiner, N., Minker, W.: Including social expectations for trustworthy proactive human-robot dialogue, pp. 23–33 (2022). https://doi.org/10.1145/3503252.3531294
Li, D., Liu, C., Xie, L.: How do consumers engage with proactive service robots? the roles of interaction orientation and corporate reputation 34(11), 3962–3981 (2022). https://doi.org/10.1108/IJCHM-10-2021-1284
Liao, L., Yang, G., Shah, C.: Proactive conversational agents, pp. 1244–1247 (2023). https://doi.org/10.1145/3539597.3572724
Liao, Q., Davis, M., Geyer, W., Muller, M., Shami, N.: What can you do? studying social-agent orientation and agent proactive interactions with an agent for employees, pp. 264–275 (2016). https://doi.org/10.1145/2901790.2901842
Liu, P., Glas, D., Kanda, T., Ishiguro, H.: Learning proactive behavior for interactive social robots 42(5), 1067–1085 (2018). https://doi.org/10.1007/s10514-017-9671-8
Moulin-Frier, C., et al.: Dac-h3: a proactive robot cognitive architecture to acquire and express knowledge about the world and the self 10(4), 1005–1022 (2018). https://doi.org/10.1109/TCDS.2017.2754143
Page, M.J., et al.: The prisma 2020 statement: an updated guideline for reporting systematic reviews. BMJ 372 (2021). https://doi.org/10.1136/bmj.n71, https://www.bmj.com/content/372/bmj.n71
Patel, M.: Longitudinal proactive robot assistance, pp. 775–777 (2023). https://doi.org/10.1145/3568294.3579982
Peng, Z., Kwon, Y., Lu, J., Wu, Z., Ma, X.: Design and evaluation of service robot’s proactivity in decision-making support process (2019). https://doi.org/10.1145/3290605.3300328
Picard, R.W.: Affective computing-mit media laboratory perceptual computing section technical report no. 321. Cambridge, MA 2139, 92 (1995)
Ring, L., Barry, B., Totzke, K., Bickmore, T.: Addressing loneliness and isolation in older adults: Proactive affective agents provide better support, pp. 61–66 (2013). https://doi.org/10.1109/ACII.2013.17
Shvo, M., Hari, R., O’Reilly, Z., Abolore, S., Wang, S.Y., McIlraith, S.: Proactive robotic assistance via theory of mind. vol. 2022, pp. 9148–9155 (2022). https://doi.org/10.1109/IROS47612.2022.9981627
Sirithunge, C., Jayasekara, A., Chandima, D.: Proactive robots with the perception of nonverbal human behavior: A review 7, 77308–77327 (2019). https://doi.org/10.1109/ACCESS.2019.2921986
Tan, H., et al.: Relationship between social robot proactive behavior and the human perception of anthropomorphic attributes 34(20), 1324–1336 (2020). https://doi.org/10.1080/01691864.2020.1831699
Ujjwal, K., Chodorowski, J.: A case study of adding proactivity in indoor social robots using belief-desire-intention (bdi) model, vol. 4(4) (2019). https://doi.org/10.3390/biomimetics4040074
Villa, L., Hervás, R., Cruz-Sandoval, D., Favela, J.: Design and evaluation of proactive behavior in conversational assistants: approach with the eva companion robot, vol. 594. LNNS, pp. 234–245 (2023). https://doi.org/10.1007/978-3-031-21333-5_23
Villa, L., Hervás, R., Cruz-Sandoval, D., Favela, J.: Design and evaluation of proactive behavior in conversational assistants: Approach with the eva companion robot. In: International Conference on Ubiquitous Computing and Ambient Intelligence, pp. 234–245. Springer (2022). https://doi.org/10.1007/978-3-031-21333-5_23
Xie, L., Liu, C., Li, D.: Proactivity or passivity? an investigation of the effect of service robots’ proactive behaviour on customer co-creation intention 106 (2022). https://doi.org/10.1016/j.ijhm.2022.103271
Zargham, N., et al.: Understanding circumstances for desirable proactive behaviour of voice assistants: The proactivity dilemma (2022). https://doi.org/10.1145/3543829.3543834
Zhou, L., Gao, J., Li, D., Shum, H.Y.: The design and implementation of xiaoice, an empathetic social chatbot. Comput. Linguist. 46(1), 53–93 (2020)
Zhu, L., Williams, T.: Effects of proactive explanations by robots on human-robot trust, vol. 12483. LNAI, pp. 85–95 (2020). https://doi.org/10.1007/978-3-030-62056-1_8
Acknowledgements
This work was supported by SHARA3 project, funded by the Junta de Comunidades de Castilla - La Mancha (SBPLY/21/180501/000160).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Johnson, E., Villa, L., Mondéjar, T., Hervás, R. (2023). Proactivity in Conversational Assistants: The mPLiCA Model Based on a Systematic Literature Review. In: Bravo, J., Urzáiz, G. (eds) Proceedings of the 15th International Conference on Ubiquitous Computing & Ambient Intelligence (UCAmI 2023). UCAmI 2023. Lecture Notes in Networks and Systems, vol 835. Springer, Cham. https://doi.org/10.1007/978-3-031-48306-6_28
Download citation
DOI: https://doi.org/10.1007/978-3-031-48306-6_28
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-48305-9
Online ISBN: 978-3-031-48306-6
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)