On the failings of Shapley values for explainability

Explainable Artificial Intelligence (XAI) is widely considered to be critical for building trust into the deployment of systems that integrate the use of machine learning (ML) models. For more than two decades Shapley values have been used as the theoretical underpinning for some methods of XAI, bei...

Celý popis

Uloženo v:
Podrobná bibliografie
Vydáno v:International journal of approximate reasoning Ročník 171; s. 109112
Hlavní autoři: Huang, Xuanxiang, Marques-Silva, Joao
Médium: Journal Article
Jazyk:angličtina
Vydáno: Elsevier Inc 01.08.2024
Témata:
ISSN:0888-613X, 1873-4731
On-line přístup:Získat plný text
Tagy: Přidat tag
Žádné tagy, Buďte první, kdo vytvoří štítek k tomuto záznamu!
Popis
Shrnutí:Explainable Artificial Intelligence (XAI) is widely considered to be critical for building trust into the deployment of systems that integrate the use of machine learning (ML) models. For more than two decades Shapley values have been used as the theoretical underpinning for some methods of XAI, being commonly referred to as SHAP scores. Some of these methods of XAI now rank among the most widely used, including in high-risk domains. This paper proves that the existing definitions of SHAP scores will necessarily yield misleading information about the relative importance of features for predictions. The paper identifies a number of ways in which misleading information can be conveyed to human decision makers, and proves that there exist classifiers which will yield such misleading information. Furthermore, the paper offers empirical evidence that such theoretical limitations of SHAP scores are routinely observed in ML classifiers.
ISSN:0888-613X
1873-4731
DOI:10.1016/j.ijar.2023.109112