Abstract
Through advances in their conversational abilities, chatbots have started to request and process an increasing variety of sensitive personal information. The accurate disclosure of sensitive information is essential where it is used to provide advice and support to users in the healthcare and finance sectors. In this study, we explore users’ concerns regarding factors associated with the use of sensitive data by chatbot providers. We surveyed a representative sample of 491 British citizens. Our results show that the user concerns focus on deleting personal information and concerns about their data’s inappropriate use. We also identified that individuals were concerned about losing control over their data after a conversation with conversational agents. We found no effect from a user’s gender or education but did find an effect from the user’s age, with those over 45 being more concerned than those under 45. We also considered the factors that engender trust in a chatbot. Our respondents’ primary focus was on the chatbot’s technical elements, with factors such as the response quality being identified as the most critical factor. We again found no effect from the user’s gender or education level; however, when we considered some social factors (e.g. avatars or perceived ‘friendliness’), we found those under 45 years old rated these as more important than those over 45. The paper concludes with a discussion of these results within the context of designing inclusive, digital systems that support a wide range of users.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
- 2.
Typically taken at 15 years of age.
- 3.
A subject-based qualification between typically forming the period from leaving compulsory education to pre-university education.
References
Fan, H., Poole, M.S.: What is personalization? perspectives on the design and implementation of personalization in information systems. J. Organ. Comput. Electron. Comm. 16(3–4), 179–202 (2006)
Følstad, A., Nordheim, C.B., Bjørkli, C.A.: What makes users trust a chatbot for customer service? an exploratory interview study. In: Bodrunova, S.S. (ed.) INSCI 2018. LNCS, vol. 11193, pp. 194–208. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01437-7_16
Ghosh, C., Eastin, M.S.: Understanding users’ relationship with voice assistants and how it affects privacy concerns and information disclosure behavior. In: Moallem, A. (ed.) HCII 2020. LNCS, vol. 12210, pp. 381–392. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-50309-3_25
Ischen, C., Araujo, T., Voorveld, H., van Noort, G., Smit, E.: Privacy concerns in chatbot interactions. In: Følstad, A., et al. (eds.) CONVERSATIONS 2019. LNCS, vol. 11970, pp. 34–48. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-39540-7_3
Lee, H., Park, H., Kim, J.: Why do people share their context information on social network services? a qualitative study and an experimental study on users’ behavior of balancing perceived benefit and risk. Int. J. Hum.-Comput. Studi. 71(9), 862–877 (2013)
McCullagh, P.: Regression models for ordinal data. J. R. Stat. Soc. Ser. B (Meth.) 42(2), 109–142 (1980). http://www.jstor.org/stable/2984952
Sağlam, R.B., Nurse, J.R.C.: Is your chatbot GDPR compliant? open issues in agent design. In: Proceedings of the 2nd Conference on Conversational User Interfaces, pp. 1–3 (2020)
Suganuma, Y., Narita, J., Nishigaki, M., Ohki, T.: Understanding the impact of service trials on privacy disclosure. In: Stephanidis, C., Antona, M. (eds.) HCII 2020. CCIS, vol. 1226, pp. 605–612. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-50732-9_78
Treiblmaier, H., Chong, S.: Trust and perceived risk of personal information as antecedents of online information disclosure: results from three countries. In: Global Diffusion and Adoption of Technologies for Knowledge and Information Sharing, pp. 341–361. IGI Global (2013)
Williams, M., Nurse, J.R.C., Creese, S.: Smartwatch games: encouraging privacy-protective behaviour in a longitudinal study. Comput. Hum. Behav. 99, 38–54 (2019)
Acknowledgements
This work is funded by the UK EPSRC ‘A Platform for Responsive Conversational Agents to Enhance Engagement and Disclosure (PRoCEED)’ project (EP/S027211/1 and EP/S027297/1).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Belen Saglam, R., Nurse, J.R.C., Hodges, D. (2021). Privacy Concerns in Chatbot Interactions: When to Trust and When to Worry. In: Stephanidis, C., Antona, M., Ntoa, S. (eds) HCI International 2021 - Posters. HCII 2021. Communications in Computer and Information Science, vol 1420. Springer, Cham. https://doi.org/10.1007/978-3-030-78642-7_53
Download citation
DOI: https://doi.org/10.1007/978-3-030-78642-7_53
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-78641-0
Online ISBN: 978-3-030-78642-7
eBook Packages: Computer ScienceComputer Science (R0)