Umeå University's logo

umu.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Evaluation of multislice inputs to convolutional neural networks for medical image segmentation
Umeå University, Faculty of Medicine, Department of Radiation Sciences, Radiation Physics.ORCID iD: 0000-0002-2391-1419
Department of Biomedical Engineering, Eindhoven University of Technology, Eindhoven 5612 AZ, the Netherlands.ORCID iD: 0000-0002-3648-4786
Umeå University, Faculty of Medicine, Department of Radiation Sciences, Radiation Physics.ORCID iD: 0000-0002-8971-9788
Umeå University, Faculty of Medicine, Department of Radiation Sciences, Radiation Physics.ORCID iD: 0000-0001-7119-7646
2020 (English)In: Medical physics (Lancaster), ISSN 0094-2405, Vol. 47, no 12, p. 6216-6231Article in journal (Refereed) Published
Abstract [en]

Purpose: When using convolutional neural networks (CNNs) for segmentation of organs and lesions in medical images, the conventional approach is to work with inputs and outputs either as single slice [two-dimensional (2D)] or whole volumes [three-dimensional (3D)]. One common alternative, in this study denoted as pseudo-3D, is to use a stack of adjacent slices as input and produce a prediction for at least the central slice. This approach gives the network the possibility to capture 3D spatial information, with only a minor additional computational cost.

Methods: In this study, we systematically evaluate the segmentation performance and computational costs of this pseudo-3D approach as a function of the number of input slices, and compare the results to conventional end-to-end 2D and 3D CNNs, and to triplanar orthogonal 2D CNNs. The standard pseudo-3D method regards the neighboring slices as multiple input image channels. We additionally design and evaluate a novel, simple approach where the input stack is a volumetric input that is repeatably convolved in 3D to obtain a 2D feature map. This 2D map is in turn fed into a standard 2D network. We conducted experiments using two different CNN backbone architectures and on eight diverse data sets covering different anatomical regions, imaging modalities, and segmentation tasks.

Results: We found that while both pseudo-3D methods can process a large number of slices at once and still be computationally much more efficient than fully 3D CNNs, a significant improvement over a regular 2D CNN was only observed with two of the eight data sets. triplanar networks had the poorest performance of all the evaluated models. An analysis of the structural properties of the segmentation masks revealed no relations to the segmentation performance with respect to the number of input slices. A post hoc rank sum test which combined all metrics and data sets yielded that only our newly proposed pseudo-3D method with an input size of 13 slices outperformed almost all methods.

Conclusion: In the general case, multislice inputs appear not to improve segmentation results over using 2D or 3D CNNs. For the particular case of 13 input slices, the proposed novel pseudo-3D method does appear to have a slight advantage across all data sets compared to all other methods evaluated in this work.

Place, publisher, year, edition, pages
John Wiley & Sons, 2020. Vol. 47, no 12, p. 6216-6231
Keywords [en]
convolutional neural network, deep learning, medical image segmentation, multislice
National Category
Medical Imaging
Identifiers
URN: urn:nbn:se:umu:diva-177165DOI: 10.1002/mp.14391ISI: 000587714100001PubMedID: 33169365Scopus ID: 2-s2.0-85096775630OAI: oai:DiVA.org:umu-177165DiVA, id: diva2:1507775
Funder
VinnovaAvailable from: 2020-12-08 Created: 2020-12-08 Last updated: 2025-02-09Bibliographically approved
In thesis
1. Resource efficient automatic segmentation of medical images
Open this publication in new window or tab >>Resource efficient automatic segmentation of medical images
2023 (English)Doctoral thesis, comprehensive summary (Other academic)
Alternative title[sv]
Resurseffektiv automatisk medicinsk bildsegmentering
Abstract [en]

Cancer is one of the leading causes of death worldwide. In 2020, there were around 10 million cancer deaths and nearly 20 million new cancer cases in the world. Radiation therapy is essential in cancer treatments because half of the cancer patients receive radiation therapy at some point. During a radiotherapy treatment planning (RTP), an oncologist must manually outline two types of areas of the patient’s body: target, which will be treated, and organs-at-risks (OARs), which are essential to avoid. This step is called delineation. The purpose of the delineation is to generate a sufficient dose plan that can provide adequate radiation dose to a tumor and limit the radiation exposure to healthy tissue. Therefore, accurate delineations are essential to achieve this goal.

Delineation is tedious and demanding for oncologists because it requires hours of concentrating work doing a repeated job. This is a RTP bottleneck which is often time- and resource-intensive. Current software, such as atlasbased techniques, can assist with this procedure by registering the patient’s anatomy to a predetermined anatomical map. However, the atlas-based methods are often slowed down and erroneous for patients with abnormal anatomies.

In recent years, deep learning (DL) methods, particularly convolutional neural networks (CNNs), have led to breakthroughs in numerous medical imaging applications. The core benefits of CNNs are weight sharing and that they can automatically detect important visual features. A typical application of CNNs for medical images is to automatically segment tumors, organs, and structures, which is assumed to save radiation oncologists much time when delineating. This thesis contributes to resource efficient automatic segmentation and covers different aspects of resource efficiency.

In Paper I, we proposed a novel end-to-end cascaded network for semantic segmentation in brain tumors in the multi-modal magnetic resonance imaging challenge in 2019. The proposed method used the hierarchical structure of the tumor sub-regions and was one of the top-ranking teams in the task of quantification of uncertainty in segmentation. A follow-up work to this paper was ranked second in the same task in the same challenge a year later.

We systematically assessed the segmentation performance and computational costs of the technique called pseudo-3D as a function of the number of input slices in Paper II. We compared the results to typical two-dimensional (2D) and three-dimensional (3D) CNNs and a method called triplanar orthogonal 2D. The typical pseudo-3D approach considers adjacent slices to be several image input channels. We discovered that a substantial benefit from employing multiple input slices was apparent for a specific input size.

We introduced a novel loss function in Paper III to address diverse issues, including imbalanced datasets, partially labeled data, and incremental learning. The proposed loss function adjusts to the given data to use all accessible data, even if some lack annotations. We show that the suggested loss function also performs well in an incremental learning context, where an existing model can be modified to incorporate the delineations of newly appearing organs semi-automatically.

In Paper IV, we proposed a novel method for compressing high-dimensional activation maps, which are the primary source of memory use in modern systems. We examined three distinct compression methods for the activation maps to accomplishing this. We demonstrated that the proposed method induces a regularization effect that acts on the layer weight gradients. By employing the proposed technique, we reduced activation map memory usage by up to 95%.

We investigated the use of generative adversarial networks (GANs) to enlarge a small dataset by generating synthetic images in Paper V. We use the real and generated data during training CNNs for the downstream segmentation tasks. Inspired by an existing GAN, we proposed a conditional version to generate high-dimensional and high-quality medical images of different modalities and their corresponding label maps. We evaluated the quality of the generated medical images and the effect of this augmentation on the performance of the segmentation task on six datasets.

Place, publisher, year, edition, pages
Umeå: Umeå University, 2023. p. 101
Series
Umeå University medical dissertations, ISSN 0346-6612 ; 2219
Keywords
radiotherapy, medical imaging, deep learning, convolutional neural network, generative adversarial network, data augmentation, semantic segmentation, classification, activation map compression
National Category
Computer Sciences Radiology, Nuclear Medicine and Medical Imaging Medical Imaging
Research subject
radiation physics
Identifiers
urn:nbn:se:umu:diva-203993 (URN)978-91-7855-956-5 (ISBN)978-91-7855-957-2 (ISBN)
Public defence
2023-02-24, Bergasalen, Byggnad 27, Entré Syd, Norrlands universitetssjukhus (NUS), Umeå, 09:00 (English)
Opponent
Supervisors
Available from: 2023-02-03 Created: 2023-01-24 Last updated: 2025-02-09Bibliographically approved

Open Access in DiVA

fulltext(1779 kB)398 downloads
File information
File name FULLTEXT02.pdfFile size 1779 kBChecksum SHA-512
1c744188dae1d5ff03d9c6b8b2da0558c23eac837eecb52a4ec9de52df8fbaf9c51423865a15940c1c32cc425961858e6b63a3ce64dbaa709808a058cbda81c1
Type fulltextMimetype application/pdf

Other links

Publisher's full textPubMedScopus

Authority records

Vu, Minh H.Nyholm, TufveLöfstedt, Tommy

Search in DiVA

By author/editor
Vu, Minh H.Grimbergen, GuusNyholm, TufveLöfstedt, Tommy
By organisation
Radiation Physics
In the same journal
Medical physics (Lancaster)
Medical Imaging

Search outside of DiVA

GoogleGoogle Scholar
Total: 506 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

doi
pubmed
urn-nbn

Altmetric score

doi
pubmed
urn-nbn
Total: 531 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf