Skip to main content

A Study on Trust Building in AI Systems Through User Commitment

  • Conference paper
  • First Online:
Human Interface and the Management of Information (HCII 2023)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 14015))

Included in the following conference series:

  • 843 Accesses

Abstract

There is always a risk of misjudgment in machine-learning based AI system that might cause monetary or safety problems to users. For this reason, it is an essential human-centered AI issue to let AI system users recognize the risk rightly and provide an opportunity to evaluate of AI system quality including the risk mitigation capability. If a user finds that an AI system is beneficial enough and the misjudgment risk is manageable, he can build trust in the system and use it. However, such user commitment has not been focused in discussion of AI trust or AI quality evaluation. This paper proposes a framework of AI system evaluation by users. The framework allows them to evaluate AI system’s quality-in-use components including risk related attributes in comprehensive manner. Based on the framework we conducted a questionnaire survey about the evaluation of personal AI systems and confirmed that users’ feedback could prioritize the influence of the quality-in-use components to their intention to use, and that could help the system provider how to implement the risk mitigation capability.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Shima. S., Ogawa, R., Sagawa, Y., Takemura, T.: Quantitative measurement of the decline in quality in use due to AI misjudgment. In: Proceedings of Computer Security Symposium 2022, pp. 759–766 (2022) (in Japanese)

    Google Scholar 

  2. Shima. S., Ogawa, R., Sagawa, Y., Takemura, T.: Impact analysis of the influence of AI misjudgment risk on intention to use AI systems. In: Proceedings of 2023 Symposium on Cryptography and Security (SCIS2023), 4E-2 (2023) (in Japanese)

    Google Scholar 

  3. Fukuzumi, S., Hirasawa, N., Wada, N., Komiyama, T., Azuma, M.: Proposal of quality in use in software quality. In: Kurosu, M. (ed.) HCII 2020. LNCS, vol. 12181, pp. 431–438. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-49059-1_31

    Chapter  Google Scholar 

  4. ISO/IEC 25010:2011 Systems and Software Engineering – Systems and Software Quality Requirements and Evaluation (SQuaRE) System and Software Quality Models. https://www.iso.org/standard/35733.html. Accessed 05 Feb 2023

  5. ISO/IEC DIS 25059 Software Engineering – Systems and Software Quality Requirements And Evaluation (SQuaRE) – Quality Model for AI Systems. https://www.iso.org/standard/80655.html. Accessed 05 Feb 2023

  6. Gezici, B., Tarhan, A.K.: Systematic literature review on software quality for AI-based software. Empir. Softw. Eng. 27(3), 1–65 (2022). https://doi.org/10.1007/s10664-021-10105-2

    Article  Google Scholar 

  7. Fujii, G., et al.: Guidelines for quality assurance of machine learning-based artificial intelligence. Int. J. Softw. Eng. Knowl. Eng. 30(11–12), 1589–1606 (2020)

    Article  Google Scholar 

  8. Nakamichi, K., et al.: Requirements-driven method to determine quality characteristics and measurements for machine learning software and its evaluation. In: IEEE 28th International Requirements Engineering Conference (RE), pp. 260–270 (2020)

    Google Scholar 

  9. Heck, O.V.: A Quality Model for Trustworthy AI Systems. https://fontysblogt.nl/a-quality-model-for-trustworthy-ai-systems. Accessed 05 Feb 2023

  10. NIST, Trust and Artificial Intelligence. https://tsapps.nist.gov/publication/get_pdf.cfm?pub_id=931087. Accessed 05 Feb 2023

  11. Davis, F.D.: Perceived usefulness, perceived ease of use, and user acceptance of information technology. MIS Q. 13, 319–340 (1989)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ryuichi Ogawa .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Ogawa, R., Shima, S., Takemura, T., Fukuzumi, Si. (2023). A Study on Trust Building in AI Systems Through User Commitment. In: Mori, H., Asahi, Y. (eds) Human Interface and the Management of Information. HCII 2023. Lecture Notes in Computer Science, vol 14015. Springer, Cham. https://doi.org/10.1007/978-3-031-35132-7_42

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-35132-7_42

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-35131-0

  • Online ISBN: 978-3-031-35132-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics