arXiv Analytics

Sign in

arXiv:2405.20664 [cs.LG]AbstractReferencesReviewsResources

Weak Robust Compatibility Between Learning Algorithms and Counterfactual Explanation Generation Algorithms

Ao Xu, Tieru Wu

Published 2024-05-31Version 1

Counterfactual explanation generation is a powerful method for Explainable Artificial Intelligence. It can help users understand why machine learning models make specific decisions, and how to change those decisions. Evaluating the robustness of counterfactual explanation algorithms is therefore crucial. Previous literature has widely studied the robustness based on the perturbation of input instances. However, the robustness defined from the perspective of perturbed instances is sometimes biased, because this definition ignores the impact of learning algorithms on robustness. In this paper, we propose a more reasonable definition, Weak Robust Compatibility, based on the perspective of explanation strength. In practice, we propose WRC-Test to help us generate more robust counterfactuals. Meanwhile, we designed experiments to verify the effectiveness of WRC-Test. Theoretically, we introduce the concepts of PAC learning theory and define the concept of PAC WRC-Approximability. Based on reasonable assumptions, we establish oracle inequalities about weak robustness, which gives a sufficient condition for PAC WRC-Approximability.

Related articles: Most relevant | Search more
arXiv:1005.2243 [cs.LG] (Published 2010-05-13)
Robustness and Generalization
arXiv:2309.13087 [cs.LG] (Published 2023-09-22)
Learning algorithms for identification of whisky using portable Raman spectroscopy
arXiv:1812.07385 [cs.LG] (Published 2018-12-15)
Perturbation Analysis of Learning Algorithms: A Unifying Perspective on Generation of Adversarial Examples