Don't Explain Noise: Robust Counterfactuals for Randomized Ensembles. Forel, A., Parmentier, A., & Vidal, T. Technical Report arXiv:2205.14116, 2022.
Don't Explain Noise: Robust Counterfactuals for Randomized Ensembles [link]Paper  abstract   bibtex   
Counterfactual explanations describe how to modify a feature vector in order to flip the outcome of a trained classifier. Several heuristic and optimal methods have been proposed to generate these explanations. However, the robustness of counterfactual explanations when the classifier is re-trained has yet to be studied. Our goal is to obtain counterfactual explanations for random forests that are robust to algorithmic uncertainty. We study the link between the robustness of ensemble models and the robustness of base learners and frame the generation of robust counterfactual explanations as a chance-constrained optimization problem. We develop a practical method with good empirical performance and provide finite-sample and asymptotic guarantees for simple random forests of stumps. We show that existing methods give surprisingly low robustness: the validity of naive counterfactuals is below $}50\{%}{$ on most data sets and can fall to $}20\{%}{$ on large problem instances with many features. Even with high plausibility, counterfactual explanations often exhibit low robustness to algorithmic uncertainty. In contrast, our method achieves high robustness with only a small increase in the distance from counterfactual explanations to their initial observations. Furthermore, we highlight the connection between the robustness of counterfactual explanations and the predictive importance of features.
@techreport{Forel2022,
abstract = {Counterfactual explanations describe how to modify a feature vector in order to flip the outcome of a trained classifier. Several heuristic and optimal methods have been proposed to generate these explanations. However, the robustness of counterfactual explanations when the classifier is re-trained has yet to be studied. Our goal is to obtain counterfactual explanations for random forests that are robust to algorithmic uncertainty. We study the link between the robustness of ensemble models and the robustness of base learners and frame the generation of robust counterfactual explanations as a chance-constrained optimization problem. We develop a practical method with good empirical performance and provide finite-sample and asymptotic guarantees for simple random forests of stumps. We show that existing methods give surprisingly low robustness: the validity of naive counterfactuals is below {\$}50\backslash{\%}{\$} on most data sets and can fall to {\$}20\backslash{\%}{\$} on large problem instances with many features. Even with high plausibility, counterfactual explanations often exhibit low robustness to algorithmic uncertainty. In contrast, our method achieves high robustness with only a small increase in the distance from counterfactual explanations to their initial observations. Furthermore, we highlight the connection between the robustness of counterfactual explanations and the predictive importance of features.},
archivePrefix = {arXiv},
arxivId = {2205.14116},
author = {Forel, A. and Parmentier, A. and Vidal, T.},
eprint = {2205.14116},
file = {:C$\backslash$:/Users/Thibaut/Documents/Mendeley-Articles/Forel, Parmentier, Vidal/Forel, Parmentier, Vidal - 2022 - Don't Explain Noise Robust Counterfactuals for Randomized Ensembles(2).pdf:pdf},
institution = {arXiv:2205.14116},
title = {{Don't Explain Noise: Robust Counterfactuals for Randomized Ensembles}},
url = {http://arxiv.org/abs/2205.14116},
year = {2022}
}

Downloads: 0