Author:
Mertes Silvan,Huber Tobias,Weitz Katharina,Heimerl Alexander,André Elisabeth
Abstract
With the ongoing rise of machine learning, the need for methods for explaining decisions made by artificial intelligence systems is becoming a more and more important topic. Especially for image classification tasks, many state-of-the-art tools to explain such classifiers rely on visual highlighting of important areas of the input data. Contrary, counterfactual explanation systems try to enable a counterfactual reasoning by modifying the input image in a way such that the classifier would have made a different prediction. By doing so, the users of counterfactual explanation systems are equipped with a completely different kind of explanatory information. However, methods for generating realistic counterfactual explanations for image classifiers are still rare. Especially in medical contexts, where relevant information often consists of textural and structural information, high-quality counterfactual images have the potential to give meaningful insights into decision processes. In this work, we present GANterfactual, an approach to generate such counterfactual image explanations based on adversarial image-to-image translation techniques. Additionally, we conduct a user study to evaluate our approach in an exemplary medical use case. Our results show that, in the chosen medical use-case, counterfactual explanations lead to significantly better results regarding mental models, explanation satisfaction, trust, emotions, and self-efficacy than two state-of-the art systems that work with saliency maps, namely LIME and LRP.
Funder
Deutsche Forschungsgemeinschaft
Bayerisches Staatsministerium für Wissenschaft, Forschung und Kunst
Reference46 articles.
1. AhsanM. M.
GuptaK. D.
IslamM.
SenS.
RahmanM. L.
HossainM.
Study of different deep learning approach with explainable AI for screening patients with COVID-19 symptoms: using CT scan and chest x-ray image dataset. 2020
2. Evaluating saliency map explanations for convolutional neural networks: a user study;Alqaraawi,2020
3. Explaining reinforcement learning to mere mortals: an empirical study;Anderson,2019
4. Explainable artificial intelligence (XAI): concepts, taxonomies, opportunities and challenges toward responsible AI;Arrieta;Inform. Fus,2020
5. On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation;Bach;PLoS ONE,2015
Cited by
35 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献