arXiv Analytics

Sign in

arXiv:2106.12543 [cs.LG]AbstractReferencesReviewsResources

Synthetic Benchmarks for Scientific Research in Explainable Machine Learning

Yang Liu, Sujay Khandagale, Colin White, Willie Neiswanger

Published 2021-06-23Version 1

As machine learning models grow more complex and their applications become more high-stakes, tools for explaining model predictions have become increasingly important. Despite the widespread use of explainability techniques, evaluating and comparing different feature attribution methods remains challenging: evaluations ideally require human studies, and empirical evaluation metrics are often computationally prohibitive on real-world datasets. In this work, we address this issue by releasing XAI-Bench: a suite of synthetic datasets along with a library for benchmarking feature attribution algorithms. Unlike real-world datasets, synthetic datasets allow the efficient computation of conditional expected values that are needed to evaluate ground-truth Shapley values and other metrics. The synthetic datasets we release offer a wide variety of parameters that can be configured to simulate real-world data. We demonstrate the power of our library by benchmarking popular explainability techniques across several evaluation metrics and identifying failure modes for popular explainers. The efficiency of our library will help bring new explainability methods from development to deployment.

Related articles: Most relevant | Search more
arXiv:2009.11698 [cs.LG] (Published 2020-09-18)
Principles and Practice of Explainable Machine Learning
arXiv:1909.06342 [cs.LG] (Published 2019-09-13)
Explainable Machine Learning in Deployment
Umang Bhatt et al.
arXiv:2203.04768 [cs.LG] (Published 2022-03-09)
Explainable Machine Learning for Predicting Homicide Clearance in the United States