skip to main content
10.1145/3530800.3534529acmconferencesArticle/Chapter ViewAbstractPublication PagestappConference Proceedingsconference-collections
short-paper

Provenance-based explanations: are they useful?

Published: 12 June 2022 Publication History

Abstract

As machine-learning models and data-driven decisions continue to be on the rise, interpreting and explaining their results has recently gained increased interest as well. In this context, provenance has often been advertised as metadata that can be used to provide explanations about some data processing. Advantages of such explanations are possibly many: they may help data engineers get a better sense of their data processing pipelines, offer transparency to auditors or end users, and ultimately, improve the trust one has in the data processing. While today, provenance can be efficiently captured for a large variety of data processing pipelines, systematic research on how useful explanations based on provenance actually is scarce. This paper advocates that there is an urgent need to study in which settings what type of provenance-based explanation is useful, given the somewhat mitigated results that have become available for explanations of machine learning results. This is underlined by a first study that shows that provenance-based explanations of record linkage decisions may not have the desired effect.

References

[1]
James Cheney, Laura Chiticariu, and Wang Chiew Tan. 2009. Provenance in Databases: Why, How, and Where. Found. Trends Databases 1, 4 (2009), 379--474.
[2]
Juliana Freire, David Koop, Emanuele Santos, and Cláudio T. Silva. 2008. Provenance for Computational Tasks: A Survey. Comput. Sci. Eng. 10, 3 (2008), 11--21.
[3]
Melanie Herschel, Ralf Diestelkämper, and Houssem Ben Lahmar. 2017. A survey on provenance: What for? What form? What from? The VLDB Journal 26, 6 (2017), 881--906.
[4]
Seokki Lee, Bertram Ludäscher, and Boris Glavic. 2020. Approximate Summaries for Why and Why-not Provenance. Proc. VLDB Endow. 13, 6 (2020), 912--924.
[5]
Scott M. Lundberg and Su-In Lee. 2017. A Unified Approach to Interpreting Model Predictions. In Proceedings of the 31st International Conference on Neural Information Processing Systems (Long Beach, California, USA) (NIPS'17). Curran Associates Inc., Red Hook, NY, USA, 4768--4777.
[6]
Sarah Oppold and Melanie Herschel. 2018. Provenance for Entity Resolution. In Provenance and Annotation of Data and Processes - 7th International Provenance and Annotation Workshop, IPAW 2018, London, UK, July 9--10, 2018, Proceedings. 226--230.
[7]
Mauricio Reyes, Raphael Meier, Särgio Pereira, Carlos A Silva, Fried-Michael Dahlweid, Hendrik von Tengg-Kobligk, Ronald M Summers, and Roland Wiest. 2020. On the Interpretability of Artificial Intelligence in Radiology: Challenges and Opportunities. Radiology: Artificial Intelligence 2, 3 (2020), 1--12.
[8]
Marco Tulio Ribeiro, Sameer Singh, and Carlos Guestrin. 2016. "Why Should I Trust You?": Explaining the Predictions of Any Classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 1135--1144.
[9]
Sudeepa Roy, Laurel J. Orr, and Dan Suciu. 2015. Explaining Query Answers with Explanation-Ready Databases. Proc. VLDB Endow. 9, 4 (2015), 348--359.
[10]
Xiaolan Wang and Alexandra Meliou. 2019. Explain3D: Explaining Disagreements in Disjoint Datasets. Proc. VLDB Endow. 12, 7 (2019), 779--792.
[11]
Hilde J. P. Weerts, Werner van Ipenburg, and Mykola Pechenizkiy. 2019. A Human-Grounded Evaluation of SHAP for Alert Processing.
[12]
Melanie Weis, Felix Naumann, Ulrich Jehle, Jens Lufter, and Holger Schuster. 2008. Industry-scale duplicate detection. Proc. VLDB Endow. 1, 2 (2008), 1253--1264.
[13]
Yunfeng Zhang, Q. Vera Liao, and Rachel K. E. Bellamy. 2020. Effect of Confidence and Explanation on Accuracy and Trust Calibration in AI-Assisted Decision Making. In Proceedings of the International Conference on Fairness, Accountability, and Transparency. Association for Computing Machinery, New York, NY, USA, 295--305.

Cited By

View all
  • (2022)Towards Observability for Production Machine Learning PipelinesProceedings of the VLDB Endowment10.14778/3565838.356585315:13(4015-4022)Online publication date: 1-Sep-2022

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
TaPP '22: Proceedings of the 14th International Workshop on the Theory and Practice of Provenance
June 2022
67 pages
ISBN:9781450393492
DOI:10.1145/3530800
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 12 June 2022

Permissions

Request permissions for this article.

Check for updates

Qualifiers

  • Short-paper

Funding Sources

  • Deutsche Forschungsgesellschaft

Conference

SIGMOD/PODS '22
Sponsor:

Acceptance Rates

TaPP '22 Paper Acceptance Rate 10 of 17 submissions, 59%;
Overall Acceptance Rate 10 of 17 submissions, 59%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)31
  • Downloads (Last 6 weeks)3
Reflects downloads up to 28 Feb 2025

Other Metrics

Citations

Cited By

View all
  • (2022)Towards Observability for Production Machine Learning PipelinesProceedings of the VLDB Endowment10.14778/3565838.356585315:13(4015-4022)Online publication date: 1-Sep-2022

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media