kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Explainable Artificial Intelligence for Human Decision Support System in the Medical Domain
Aalto Univ, Dept Comp Sci, Konemiehentie 2, Espoo 02150, Finland.;Umeå Univ, Dept Comp Sci, S-90187 Umeå, Sweden..
Aalto Univ, Dept Comp Sci, Konemiehentie 2, Espoo 02150, Finland.;Bournemouth Univ, Dept Comp & Informat, Poole BH12 5BB, Dorset, England..
KTH, School of Electrical Engineering and Computer Science (EECS). Aalto Univ, Dept Comp Sci, Konemiehentie 2, Espoo 02150, Finland..ORCID iD: 0000-0001-7553-4291
Aalto Univ, Dept Comp Sci, Konemiehentie 2, Espoo 02150, Finland.;Umeå Univ, Dept Comp Sci, S-90187 Umeå, Sweden..
2021 (English)In: Machine Learning and Knowledge Extraction, ISSN 2504-4990, Vol. 3, no 3, p. 740-770Article in journal (Refereed) Published
Abstract [en]

In this paper, we present the potential of Explainable Artificial Intelligence methods for decision support in medical image analysis scenarios. Using three types of explainable methods applied to the same medical image data set, we aimed to improve the comprehensibility of the decisions provided by the Convolutional Neural Network (CNN). In vivo gastral images obtained by a video capsule endoscopy (VCE) were the subject of visual explanations, with the goal of increasing health professionals' trust in black-box predictions. We implemented two post hoc interpretable machine learning methods, called Local Interpretable Model-Agnostic Explanations (LIME) and SHapley Additive exPlanations (SHAP), and an alternative explanation approach, the Contextual Importance and Utility (CIU) method. The produced explanations were assessed by human evaluation. We conducted three user studies based on explanations provided by LIME, SHAP and CIU. Users from different non-medical backgrounds carried out a series of tests in a web-based survey setting and stated their experience and understanding of the given explanations. Three user groups (n = 20, 20, 20) with three distinct forms of explanations were quantitatively analyzed. We found that, as hypothesized, the CIU-explainable method performed better than both LIME and SHAP methods in terms of improving support for human decision-making and being more transparent and thus understandable to users. Additionally, CIU outperformed LIME and SHAP by generating explanations more rapidly. Our findings suggest that there are notable differences in human decision-making between various explanation support settings. In line with that, we present three potential explainable methods that, with future improvements in implementation, can be generalized to different medical data sets and can provide effective decision support to medical experts.

Place, publisher, year, edition, pages
MDPI AG , 2021. Vol. 3, no 3, p. 740-770
Keywords [en]
explainable artificial intelligence, human decision support, image recognition, medical image analyses
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:kth:diva-303546DOI: 10.3390/make3030037ISI: 000703047500001Scopus ID: 2-s2.0-85142788243OAI: oai:DiVA.org:kth-303546DiVA, id: diva2:1604406
Note

QC 20211019

Available from: 2021-10-19 Created: 2021-10-19 Last updated: 2023-12-05Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Saluja, Rohit

Search in DiVA

By author/editor
Saluja, Rohit
By organisation
School of Electrical Engineering and Computer Science (EECS)
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 38 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf