Do Not Trust Additive Explanations

From MaRDI portal
Publication:126735

DOI10.48550/ARXIV.1903.11420arXiv1903.11420MaRDI QIDQ126735FDOQ126735


Authors: Przemysław Biecek, Alicja Gosiewska Edit this on Wikidata


Publication date: 27 March 2019

Abstract: Explainable Artificial Intelligence (XAI)has received a great deal of attention recently. Explainability is being presented as a remedy for the distrust of complex and opaque models. Model agnostic methods such as LIME, SHAP, or Break Down promise instance-level interpretability for any complex machine learning model. But how faithful are these additive explanations? Can we rely on additive explanations for non-additive models? In this paper, we (1) examine the behavior of the most popular instance-level explanations under the presence of interactions, (2) introduce a new method that detects interactions for instance-level explanations, (3) perform a large scale benchmark to see how frequently additive explanations may be misleading.








Cited In (2)





This page was built for publication: Do Not Trust Additive Explanations

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q126735)