Skip to main content
Article
Counterfactual explanation of Bayesian model uncertainty
Neural Computing and Applications
  • Gohar Ali, COMSATS University Islamabad
  • Feras Al-Obeidat, Zayed University
  • Abdallah Tubaishat, Zayed University
  • Tehseen Zia, COMSATS University Islamabad
  • Muhammad Ilyas, University of Sargodha
  • Alvaro Rocha, Instituto Superior de Economia e Gestão, Universidade de Lisboa
ORCID Identifiers

0000-0001-8176-3373

Document Type
Article
Publication Date
1-1-2021
Abstract

Artificial intelligence systems are becoming ubiquitous in everyday life as well as in high-risk environments, such as autonomous driving, medical treatment, and medicine. The opaque nature of the deep neural network raises concerns about its adoption in high-risk environments. It is important for researchers to explain how these models reach their decisions. Most of the existing methods rely on softmax to explain model decisions. However, softmax is shown to be often misleading, particularly giving unjustified high confidence even for samples far from the training data. To overcome this shortcoming, we propose Bayesian model uncertainty for producing counterfactual explanations. In this paper, we compare the counterfactual explanation of models based on Bayesian uncertainty and softmax score. This work predictively produces minimal important features, which maximally change classifier output to explain the decision-making process of the Bayesian model. We used MNIST and Caltech Bird 2011 datasets for experiments. The results show that the Bayesian model outperforms the softmax model and produces more concise and human-understandable counterfactuals.

Publisher
Springer Science and Business Media LLC
Disciplines
Keywords
  • Bayesian model uncertainty,
  • Counterfactual explanation,
  • Deep learning
Scopus ID
85115614824
Indexed in Scopus
Yes
Open Access
No
https://doi.org/10.1007/s00521-021-06528-z
Citation Information
Gohar Ali, Feras Al-Obeidat, Abdallah Tubaishat, Tehseen Zia, et al.. "Counterfactual explanation of Bayesian model uncertainty" Neural Computing and Applications (2021) ISSN: <a href="https://v2.sherpa.ac.uk/id/publication/issn/0941-0643" target="_blank">0941-0643</a>
Available at: http://works.bepress.com/feras-al-obeidat/54/