Umeå University's logo

umu.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
EFNet-CSM: EfficientNet with a modified attention mechanism for effective fire detection
Sejong University, Seoul, South Korea.
School of Engineering and Computing, University of Lancashire, Lancashire, United Kingdom.
Umeå University, Faculty of Science and Technology, Department of Computing Science.
Sejong University, Seoul, South Korea.
Show others and affiliations
2025 (English)In: Knowledge-Based Systems, ISSN 0950-7051, E-ISSN 1872-7409, Vol. 329, article id 114353Article in journal (Refereed) Published
Abstract [en]

Fire is considered one of the major threats to life, property, ecosystems, global warming, and the economy. Recent advancements in convolution neural networks have shown potential for vision-based fire detection; however, several challenges are associated with these techniques, such as limited model performance and high computational complexity. To address these issues, we present an efficient CNN-based model in which EfficientNetV2B0 is employed as a backbone feature extractor and is integrated with a channel and modified spatial attention mechanism to extract deeper spatial details, thereby weighting important features appropriately. The spatial attention is modified by introducing two depth-separable convolution layers to control computational complexity without affecting the performance. The proposed model is assessed on four benchmark datasets in the domain of remote sensing and CCTV-based systems for effective fire detection. Experimental analysis reveals that our model outperforms existing methods in terms of higher accuracy and inference speed, with lower model size and computational burden, indicating its suitability for deployment on resource-constrained devices in real time. To explain the predictions made by the proposed model, we use explainable artificial intelligence methods called Grad-CAM, guided backpropagation, and guided Grad-CAM to provide visualizations by localizing the most salient regions in the image, as emphasized by the attention mechanism.

Place, publisher, year, edition, pages
Elsevier, 2025. Vol. 329, article id 114353
Keywords [en]
Attention mechanism, Convolution neural networks, Deep learning, Disaster management, Fire detection, Internet of things, Spatial attention, Surveillance systems
National Category
Computer Sciences Bioinformatics (Computational Biology)
Identifiers
URN: urn:nbn:se:umu:diva-245354DOI: 10.1016/j.knosys.2025.114353ISI: 001572118500001Scopus ID: 2-s2.0-105015369220OAI: oai:DiVA.org:umu-245354DiVA, id: diva2:2005527
Available from: 2025-10-10 Created: 2025-10-10 Last updated: 2025-10-10Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Khan, Zulfiqar Ahmad

Search in DiVA

By author/editor
Khan, Zulfiqar Ahmad
By organisation
Department of Computing Science
In the same journal
Knowledge-Based Systems
Computer SciencesBioinformatics (Computational Biology)

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 24 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf