Multimodal explainability via latent shift applied to COVID-19 stratificationShow others and affiliations
2024 (English)In: Pattern Recognition, ISSN 0031-3203, E-ISSN 1873-5142, Vol. 156, article id 110825Article in journal (Refereed) Published
Abstract [en]
We are witnessing a widespread adoption of artificial intelligence in healthcare. However, most of the advancements in deep learning in this area consider only unimodal data, neglecting other modalities. Their multimodal interpretation necessary for supporting diagnosis, prognosis and treatment decisions. In this work we present a deep architecture, which jointly learns modality reconstructions and sample classifications using tabular and imaging data. The explanation of the decision taken is computed by applying a latent shift that, simulates a counterfactual prediction revealing the features of each modality that contribute the most to the decision and a quantitative score indicating the modality importance. We validate our approach in the context of COVID-19 pandemic using the AIforCOVID dataset, which contains multimodal data for the early identification of patients at risk of severe outcome. The results show that the proposed method provides meaningful explanations without degrading the classification performance.
Place, publisher, year, edition, pages
Elsevier, 2024. Vol. 156, article id 110825
Keywords [en]
Classification, COVID-19, Joint fusion, Multimodal deep learning, XAI
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:umu:diva-228190DOI: 10.1016/j.patcog.2024.110825ISI: 001284879500001Scopus ID: 2-s2.0-85199889813OAI: oai:DiVA.org:umu-228190DiVA, id: diva2:1887305
Funder
Swedish Research Council, 2018-059732024-08-072024-08-072025-04-24Bibliographically approved