As a popular approach in Explainable AI (XAI), an increasing number of counterfactual explanation algorithms have been proposed in the context of making machine learning classifiers more trustworthy and transparent. This paper reports our evaluations of algorithms that can output diverse counterfactuals for one instance. We first evaluate the performance of DiCE-Random, DiCE-KDTree, DiCE-Genetic and Alibi-CFRL, taking XGBoost as the machine learning model for binary classification problems. Then, we compare their suggested feature changes with feature importance by SHAP. Moreover, our study highlights that synthetic counterfactuals, drawn from the input domain but not necessarily the training data, outperform native counter-factuals from the training data regarding data privacy and validity. This research aims to guide practitioners in choosing the most suitable algorithm for generating diverse counterfactual explanations.
Rodriguez, PauCaccia, MassimoLacoste, AlexandreZamparo, LeeLaradji, IssamCharlin, LaurentVazquez, David
Pau RodríguezM. CacciaAlexandre LacosteLee ZamparoIssam LaradjiLaurent CharlinDavid Vázquez
Susanne DandlChristoph MolnarMartin BinderBernd Bischl