Umeå University's logo

umu.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
CN-waterfall: a deep convolutional neural network for multimodal physiological affect detection
Umeå University, Faculty of Science and Technology, Department of Computing Science.ORCID iD: 0000-0002-9009-0999
Centre for Applied Autonomous Sensor Systems (AASS), Örebro University, Örebro, Sweden.
Umeå University, Faculty of Science and Technology, Department of Computing Science. School of Science and Technology, Aalto University, Espoo, Finland.ORCID iD: 0000-0002-8078-5172
2022 (English)In: Neural Computing & Applications, ISSN 0941-0643, E-ISSN 1433-3058, Vol. 34, no 3, p. 2157-2176Article in journal (Refereed) Published
Abstract [en]

Affective computing solutions, in the literature, mainly rely on machine learning methods designed to accurately detect human affective states. Nevertheless, many of the proposed methods are based on handcrafted features, requiring sufficient expert knowledge in the realm of signal processing. With the advent of deep learning methods, attention has turned toward reduced feature engineering and more end-to-end machine learning. However, most of the proposed models rely on late fusion in a multimodal context. Meanwhile, addressing interrelations between modalities for intermediate-level data representation has been largely neglected. In this paper, we propose a novel deep convolutional neural network, called CN-Waterfall, consisting of two modules: Base and General. While the Base module focuses on the low-level representation of data from each single modality, the General module provides further information, indicating relations between modalities in the intermediate- and high-level data representations. The latter module has been designed based on theoretically grounded concepts in the Explainable AI (XAI) domain, consisting of four different fusions. These fusions are mainly tailored to correlation- and non-correlation-based modalities. To validate our model, we conduct an exhaustive experiment on WESAD and MAHNOB-HCI, two publicly and academically available datasets in the context of multimodal affective computing. We demonstrate that our proposed model significantly improves the performance of physiological-based multimodal affect detection.

Place, publisher, year, edition, pages
Springer, 2022. Vol. 34, no 3, p. 2157-2176
Keywords [en]
Multimodal affect detection, Deep convolutional neural network, Physiological-based sensors, Data fusion
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:umu:diva-187972DOI: 10.1007/s00521-021-06516-3ISI: 000698886400003Scopus ID: 2-s2.0-85115620535OAI: oai:DiVA.org:umu-187972DiVA, id: diva2:1598420
Funder
Knut and Alice Wallenberg FoundationAvailable from: 2021-09-28 Created: 2021-09-28 Last updated: 2024-04-29Bibliographically approved
In thesis
1. Affect detection with explainable AI for time series
Open this publication in new window or tab >>Affect detection with explainable AI for time series
2024 (English)Doctoral thesis, comprehensive summary (Other academic)
Alternative title[sv]
Detektering av känslomässiga reaktioner med förklarande AI för tidsserier
Abstract [en]

The exponential growth in the utilization of machine learning (ML) models has facilitated the development of decision-making systems, particularly in tasks such as affect detection. Affect-aware systems, which discern human affective states by blending emotion theories with practical engineering, have garnered significant attention. Implementing such systems entails various approaches. We focus on leveraging ML techniques to elucidate relations between affective states and multiple physiological indicators including sensory time series data. However, a notable technical challenge arises from the expensive design of existing models, particularly problematic in knowledge constrained environments. This thesis endeavors to address this challenge by proposing a meticulously crafted end-to-end deep learning model, drawing inspiration from the principles of decision explainability in affect detectors.

Explainable artificial intelligence (XAI) seeks to demystify the decision-making process of ML models, mitigating the "black box" effect stemming from their complex, non-linear structures. Enhanced transparency fosters trust among end-users and mitigates the risks associated with biased outcomes. Despite rapid advancements in XAI, particularly in visionary tasks, the methods employed are not readily applicable to time-series data, especially in affect detection tasks. This thesis thus aims to pioneer the fusion of XAI techniques with affect detection, with a dual objective: firstly, to render the decisions of affect detectors transparent through the introduction of a valid explainable model; and secondly, to assess the state of explainability in affect detection time series data by presenting a range of objective metrics. 

In summary, this research carries significant practical implications, benefiting society at large. The proposed affect detector can not only be served as a benchmark in the field, but also perceived as a priori for related tasks such as depression detection. Our work further facilitates a full integration of the detector into real-world settings when coupled with the accompanying explainability tools. These tools can indeed be utilized in any decision-making domains where ML techniques are practiced on time series data. The findings of this research also spread awareness to scholars about carefully designing transparent systems. 

Place, publisher, year, edition, pages
Umeå University, 2024. p. 57
Series
UMINF, ISSN 0348-0542 ; 24.06
Keywords
Explainable AI, Affect Detection, Time Series, Deep Convolutional Neural Network, Machine Learning
National Category
Computer Sciences
Identifiers
urn:nbn:se:umu:diva-223857 (URN)978-91-8070-407-6 (ISBN)978-91-8070-408-3 (ISBN)
Public defence
2024-05-24, Hörsal MIT.A.121, 13:15 (English)
Opponent
Supervisors
Available from: 2024-05-03 Created: 2024-04-29 Last updated: 2024-04-30Bibliographically approved

Open Access in DiVA

fulltext(2656 kB)491 downloads
File information
File name FULLTEXT02.pdfFile size 2656 kBChecksum SHA-512
4f7e834321b082746bb51fe7fa1a75d37411f50226f50b85f38d12ff5d6465da3ddb0abe337c00d78ade5ebeb1c0f7b72de1345119f640e7a1c1d335c25c6fcb
Type fulltextMimetype application/pdf

Other links

Publisher's full textScopus

Authority records

Fouladgar, NazaninFrämling, Kary

Search in DiVA

By author/editor
Fouladgar, NazaninFrämling, Kary
By organisation
Department of Computing Science
In the same journal
Neural Computing & Applications
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar
Total: 591 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 702 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf