Umeå University's logo

umu.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Context-based image explanations for deep neural networks
Umeå University, Faculty of Science and Technology, Department of Computing Science.ORCID iD: 0000-0002-1232-346x
Department of Computer Science, University of Oxford, United Kingdom.
Umeå University, Faculty of Science and Technology, Department of Computing Science.
2021 (English)In: Image and Vision Computing, ISSN 0262-8856, E-ISSN 1872-8138, Vol. 116, article id 104310Article in journal (Refereed) Published
Abstract [en]

With the increased use of machine learning in decision-making scenarios, there has been a growing interest in explaining and understanding the outcomes of machine learning models. Despite this growing interest, existing works on interpretability and explanations have been mostly intended for expert users. Explanations for general users have been neglected in many usable and practical applications (e.g., image tagging, caption generation). It is important for non-technical users to understand features and how they affect an instance-specific prediction to satisfy the need for justification. In this paper, we propose a model-agnostic method for generating context-based explanations aiming for general users. We implement partial masking on segmented components to identify the contextual importance of each segment in scene classification tasks. We then generate explanations based on feature importance. We present visual and text-based explanations: (i) saliency map presents the pertinent components with a descriptive textual justification, (ii) visual map with a color bar graph showing the relative importance of each feature for a prediction. Evaluating the explanations using a user study (N = 50), we observed that our proposed explanation method visually outperformed existing gradient and occlusion based methods. Hence, our proposed explanation method could be deployed to explain models’ decisions to non-expert users in real-world applications.

Place, publisher, year, edition, pages
Elsevier, 2021. Vol. 116, article id 104310
Keywords [en]
Contextual importance, DNNs, Explainable AI, Visual explanations
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:umu:diva-191025DOI: 10.1016/j.imavis.2021.104310ISI: 000710706800003Scopus ID: 2-s2.0-85115994867OAI: oai:DiVA.org:umu-191025DiVA, id: diva2:1625810
Available from: 2022-01-10 Created: 2022-01-10 Last updated: 2023-09-05Bibliographically approved
In thesis
1. Context-based explanations for machine learning predictions
Open this publication in new window or tab >>Context-based explanations for machine learning predictions
2022 (English)Doctoral thesis, comprehensive summary (Other academic)
Alternative title[sv]
Kontextbaserade förklaringar för maskininlärningsförutsägelser
Abstract [en]

In recent years, growing concern regarding trust in algorithmic decision-making has drawn attention to more transparent and interpretable models. Laws and regulations are moving towards requiring this functionality from information systems to prevent unintended side effects. Such as the European Union's General Data Protection Regulations (GDPR) set out the right to be informed regarding machine-generated decisions. Individuals affected by these decisions can question, confront and challenge the inferences automatically produced by machine learning models. Consequently, such matters necessitate AI systems to be transparent and explainable for various practical applications.

Furthermore, explanations help evaluate these systems' strengths and limitations, thereby fostering trustworthiness. As important as it is, existing studies mainly focus on creating mathematically interpretable models or explaining black-box algorithms with intrinsically interpretable surrogate models. In general, these explanations are intended for technical users to evaluate the correctness of a model and are often hard to interpret by general users.  

Given a critical need for methods that consider end-user requirements, this thesis focuses on generating intelligible explanations for predictions made by machine learning algorithms. As a starting point, we present the outcome of a systematic literature review of the existing research on generating and communicating explanations in goal-driven eXplainable AI (XAI), such as agents and robots. These are known for their ability to communicate their decisions in human understandable terms. Influenced by that, we discuss the design and evaluation of our proposed explanation methods for black-box algorithms in different machine learning applications, including image recognition, scene classification, and disease prediction.

Taken together, the methods and tools presented in this thesis could be used to explain machine learning predictions or as a baseline to compare to other explanation techniques, enabling interpretation indicators for experts and non-technical users. The findings would also be of interest to domains using machine learning models for high-stake decision-making to investigate the practical utility of proposed explanation methods.

Place, publisher, year, edition, pages
Umeå: Umeå University, 2022. p. 48
Series
Report / UMINF, ISSN 0348-0542
Keywords
Explainable AI, explainability, interpretability, black-box models, deep learning, neural networks, contextual importance
National Category
Computer Systems
Research subject
Computer Science
Identifiers
urn:nbn:se:umu:diva-198943 (URN)978-91-7855-859-9 (ISBN)978-91-7855-860-5 (ISBN)
Public defence
2022-09-26, NAT.D.320, Naturvetarhuset, Umeå, 08:30 (English)
Opponent
Supervisors
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Available from: 2022-09-05 Created: 2022-08-29 Last updated: 2022-08-30Bibliographically approved

Open Access in DiVA

fulltext(2967 kB)572 downloads
File information
File name FULLTEXT01.pdfFile size 2967 kBChecksum SHA-512
81a2b3f1e5da4d60572fea1029814bc217de3f85e78d14a2f419d31eb991fa36997d6713408d3fcc84976c8dbd2cf68d41ceb2624ae39083e0313b1c61b5c412
Type fulltextMimetype application/pdf

Other links

Publisher's full textScopus

Authority records

Anjomshoae, SuleJiang, Lili

Search in DiVA

By author/editor
Anjomshoae, SuleJiang, Lili
By organisation
Department of Computing Science
In the same journal
Image and Vision Computing
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar
Total: 573 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 439 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf