arXiv Analytics

Sign in

arXiv:2106.12228 [stat.ML]AbstractReferencesReviewsResources

groupShapley: Efficient prediction explanation with Shapley values for feature groups

Martin Jullum, Annabelle Redelmeier, Kjersti Aas

Published 2021-06-23Version 1

Shapley values has established itself as one of the most appropriate and theoretically sound frameworks for explaining predictions from complex machine learning models. The popularity of Shapley values in the explanation setting is probably due to its unique theoretical properties. The main drawback with Shapley values, however, is that its computational complexity grows exponentially in the number of input features, making it unfeasible in many real world situations where there could be hundreds or thousands of features. Furthermore, with many (dependent) features, presenting/visualizing and interpreting the computed Shapley values also becomes challenging. The present paper introduces groupShapley: a conceptually simple approach for dealing with the aforementioned bottlenecks. The idea is to group the features, for example by type or dependence, and then compute and present Shapley values for these groups instead of for all individual features. Reducing hundreds or thousands of features to half a dozen or so, makes precise computations practically feasible and the presentation and knowledge extraction greatly simplified. We prove that under certain conditions, groupShapley is equivalent to summing the feature-wise Shapley values within each feature group. Moreover, we provide a simulation study exemplifying the differences when these conditions are not met. We illustrate the usability of the approach in a real world car insurance example, where groupShapley is used to provide simple and intuitive explanations.

Related articles: Most relevant | Search more
arXiv:2205.13662 [stat.ML] (Published 2022-05-26)
Explaining Preferences with Shapley Values
arXiv:1909.03495 [stat.ML] (Published 2019-09-08)
Shapley Values of Reconstruction Errors of PCA for Explaining Anomaly Detection
arXiv:1903.10464 [stat.ML] (Published 2019-03-25)
Explaining individual predictions when features are dependent: More accurate approximations to Shapley values