Umeå University's logo

umu.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Multimodal explainability via latent shift applied to COVID-19 stratification
Unit of Computer Systems and Bioinformatics, Department of Engineering, University Campus Bio-Medico of Rome, Rome, Italy.
Unit of Computer Systems and Bioinformatics, Department of Engineering, University Campus Bio-Medico of Rome, Rome, Italy.
IRCCS Istituto Ortopedico Galeazzi, Milan, Italy; Department of Biomedical, Surgical and Dental Sciences, Università degli Studi di Milano, Milan, Italy.
Department of Radiology, Sant'Anna Hospital, Como, San Fermo della Battaglia, Italy.
Show others and affiliations
2024 (English)In: Pattern Recognition, ISSN 0031-3203, E-ISSN 1873-5142, Vol. 156, article id 110825Article in journal (Refereed) Published
Abstract [en]

We are witnessing a widespread adoption of artificial intelligence in healthcare. However, most of the advancements in deep learning in this area consider only unimodal data, neglecting other modalities. Their multimodal interpretation necessary for supporting diagnosis, prognosis and treatment decisions. In this work we present a deep architecture, which jointly learns modality reconstructions and sample classifications using tabular and imaging data. The explanation of the decision taken is computed by applying a latent shift that, simulates a counterfactual prediction revealing the features of each modality that contribute the most to the decision and a quantitative score indicating the modality importance. We validate our approach in the context of COVID-19 pandemic using the AIforCOVID dataset, which contains multimodal data for the early identification of patients at risk of severe outcome. The results show that the proposed method provides meaningful explanations without degrading the classification performance.

Place, publisher, year, edition, pages
Elsevier, 2024. Vol. 156, article id 110825
Keywords [en]
Classification, COVID-19, Joint fusion, Multimodal deep learning, XAI
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:umu:diva-228190DOI: 10.1016/j.patcog.2024.110825ISI: 001284879500001Scopus ID: 2-s2.0-85199889813OAI: oai:DiVA.org:umu-228190DiVA, id: diva2:1887305
Funder
Swedish Research Council, 2018-05973Available from: 2024-08-07 Created: 2024-08-07 Last updated: 2025-04-24Bibliographically approved

Open Access in DiVA

fulltext(1368 kB)75 downloads
File information
File name FULLTEXT01.pdfFile size 1368 kBChecksum SHA-512
64492496443353badc3faf2bac7d955bba3f2d6a58cf24f50786de5e819a44f6c3cec5740d656def0c212a2ee272db64348e4132d9071a90984769876cbb6f45
Type fulltextMimetype application/pdf

Other links

Publisher's full textScopus

Authority records

Soda, Paolo

Search in DiVA

By author/editor
Soda, Paolo
By organisation
Department of Diagnostics and InterventionRadiation Physics
In the same journal
Pattern Recognition
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar
Total: 75 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 208 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf