Nothing Special   »   [go: up one dir, main page]

skip to main content
10.5555/3295222.3295230guideproceedingsArticle/Chapter ViewAbstractPublication PagesnipsConference Proceedingsconference-collections
Article
Free access

A unified approach to interpreting model predictions

Published: 04 December 2017 Publication History

Abstract

Understanding why a model makes a certain prediction can be as crucial as the prediction's accuracy in many applications. However, the highest accuracy for large modern datasets is often achieved by complex models that even experts struggle to interpret, such as ensemble or deep learning models, creating a tension between accuracy and interpretability. In response, various methods have recently been proposed to help users interpret the predictions of complex models, but it is often unclear how these methods are related and when one method is preferable over another. To address this problem, we present a unified framework for interpreting predictions, SHAP (SHapley Additive exPlanations). SHAP assigns each feature an importance value for a particular prediction. Its novel components include: (1) the identification of a new class of additive feature importance measures, and (2) theoretical results showing there is a unique solution in this class with a set of desirable properties. The new class unifies six existing methods, notable because several recent methods in the class lack the proposed desirable properties. Based on insights from this unification, we present new methods that show improved computational performance and/or better consistency with human intuition than previous approaches.

References

[1]
Sebastian Bach et al. "On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation". In: PloS One 10.7 (2015), e0130140.
[2]
A Charnes et al. "Extremal principle solutions of games in characteristic function form: core, Chebychev and Shapley value generalizations". In: Econometrics of Planning and Efficiency 11 (1988), pp. 123-133.
[3]
Anupam Datta, Shayak Sen, and Yair Zick. "Algorithmic transparency via quantitative input influence: Theory and experiments with learning systems". In: Security and Privacy (SP), 2016 IEEE Symposium on. IEEE. 2016, pp. 598-617.
[4]
Stan Lipovetsky and Michael Conklin. "Analysis of regression in game theory approach". In: Applied Stochastic Models in Business and Industry 17.4 (2001), pp. 319-330.
[5]
Marco Tulio Ribeiro, Sameer Singh, and Carlos Guestrin. "Why should i trust you?: Explaining the predictions of any classifier". In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. ACM. 2016, pp. 1135-1144.
[6]
Lloyd S Shapley. "A value for n-person games". In: Contributions to the Theory of Games 2.28 (1953), pp. 307-317.
[7]
Avanti Shrikumar, Peyton Greenside, and Anshul Kundaje. "Learning Important Features Through Propagating Activation Differences". In: arXiv preprint arXiv:1704.02685 (2017).
[8]
Avanti Shrikumar et al. "Not Just a Black Box: Learning Important Features Through Propagating Activation Differences". In: arXiv preprint arXiv:1605.01713 (2016).
[9]
Erik Štrumbelj and Igor Kononenko. "Explaining prediction models and individual predictions with feature contributions". In: Knowledge and information systems 41.3 (2014), pp. 647-665.
[10]
H Peyton Young. "Monotonic solutions of cooperative games". In: International Journal of Game Theory 14.2 (1985), pp. 65-72.

Cited By

View all
  • (2025)XAI Reveals the Causes of Attention Deficit Hyperactivity Disorder (ADHD) Bias in Student Performance PredictionProceedings of the 15th International Learning Analytics and Knowledge Conference10.1145/3706468.3706521(418-428)Online publication date: 3-Mar-2025
  • (2025)X-DFS: Explainable Artificial Intelligence Guided Design-for-Security Solution Space ExplorationIEEE Transactions on Information Forensics and Security10.1109/TIFS.2024.351585520(753-766)Online publication date: 1-Jan-2025
  • (2025)Using AI explainable models and handwriting/drawing tasks for psychological well-beingInformation Systems10.1016/j.is.2024.102465127:COnline publication date: 1-Jan-2025
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image Guide Proceedings
NIPS'17: Proceedings of the 31st International Conference on Neural Information Processing Systems
December 2017
7104 pages

Publisher

Curran Associates Inc.

Red Hook, NY, United States

Publication History

Published: 04 December 2017

Qualifiers

  • Article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)8,955
  • Downloads (Last 6 weeks)1,433
Reflects downloads up to 05 Mar 2025

Other Metrics

Citations

Cited By

View all
  • (2025)XAI Reveals the Causes of Attention Deficit Hyperactivity Disorder (ADHD) Bias in Student Performance PredictionProceedings of the 15th International Learning Analytics and Knowledge Conference10.1145/3706468.3706521(418-428)Online publication date: 3-Mar-2025
  • (2025)X-DFS: Explainable Artificial Intelligence Guided Design-for-Security Solution Space ExplorationIEEE Transactions on Information Forensics and Security10.1109/TIFS.2024.351585520(753-766)Online publication date: 1-Jan-2025
  • (2025)Using AI explainable models and handwriting/drawing tasks for psychological well-beingInformation Systems10.1016/j.is.2024.102465127:COnline publication date: 1-Jan-2025
  • (2025)A hybrid ensemble model to detect Bitcoin fraudulent transactionsEngineering Applications of Artificial Intelligence10.1016/j.engappai.2024.109810141:COnline publication date: 1-Feb-2025
  • (2025)Interpretation with baseline shapley value for feature groups on tree modelsFrontiers of Computer Science: Selected Publications from Chinese Universities10.1007/s11704-024-40117-219:5Online publication date: 1-May-2025
  • (2025)PerfoRank: cluster-based performance ranking for improved performance evaluation and estimation in professional cyclingMachine Language10.1007/s10994-024-06716-7114:1Online publication date: 16-Jan-2025
  • (2024)XplainProceedings of the 33rd USENIX Conference on Security Symposium10.5555/3698900.3699065(2937-2953)Online publication date: 14-Aug-2024
  • (2024)Benchmarking deletion metrics with the principled explanationsProceedings of the 41st International Conference on Machine Learning10.5555/3692070.3694186(51569-51595)Online publication date: 21-Jul-2024
  • (2024)Feature importance disparities for data bias investigationsProceedings of the 41st International Conference on Machine Learning10.5555/3692070.3692309(6170-6201)Online publication date: 21-Jul-2024
  • (2024)Cutting the black boxProceedings of the Thirty-Third International Joint Conference on Artificial Intelligence10.24963/ijcai.2024/406(3669-3678)Online publication date: 3-Aug-2024
  • Show More Cited By

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Login options

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media