Umeå universitets logga

umu.sePublikationer
Driftmeddelande
För närvarande är det driftstörningar. Felsökning pågår.
Ändra sökning
RefereraExporteraLänk till posten
Permanent länk

Direktlänk
Referera
Referensformat
  • apa
  • ieee
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf
Resource efficient automatic segmentation of medical images
Umeå universitet, Medicinska fakulteten, Institutionen för strålningsvetenskaper, Radiofysik.ORCID-id: 0000-0002-2391-1419
2023 (Engelska)Doktorsavhandling, sammanläggning (Övrigt vetenskapligt)Alternativ titel
Resurseffektiv automatisk medicinsk bildsegmentering (Svenska)
Abstract [en]

Cancer is one of the leading causes of death worldwide. In 2020, there were around 10 million cancer deaths and nearly 20 million new cancer cases in the world. Radiation therapy is essential in cancer treatments because half of the cancer patients receive radiation therapy at some point. During a radiotherapy treatment planning (RTP), an oncologist must manually outline two types of areas of the patient’s body: target, which will be treated, and organs-at-risks (OARs), which are essential to avoid. This step is called delineation. The purpose of the delineation is to generate a sufficient dose plan that can provide adequate radiation dose to a tumor and limit the radiation exposure to healthy tissue. Therefore, accurate delineations are essential to achieve this goal.

Delineation is tedious and demanding for oncologists because it requires hours of concentrating work doing a repeated job. This is a RTP bottleneck which is often time- and resource-intensive. Current software, such as atlasbased techniques, can assist with this procedure by registering the patient’s anatomy to a predetermined anatomical map. However, the atlas-based methods are often slowed down and erroneous for patients with abnormal anatomies.

In recent years, deep learning (DL) methods, particularly convolutional neural networks (CNNs), have led to breakthroughs in numerous medical imaging applications. The core benefits of CNNs are weight sharing and that they can automatically detect important visual features. A typical application of CNNs for medical images is to automatically segment tumors, organs, and structures, which is assumed to save radiation oncologists much time when delineating. This thesis contributes to resource efficient automatic segmentation and covers different aspects of resource efficiency.

In Paper I, we proposed a novel end-to-end cascaded network for semantic segmentation in brain tumors in the multi-modal magnetic resonance imaging challenge in 2019. The proposed method used the hierarchical structure of the tumor sub-regions and was one of the top-ranking teams in the task of quantification of uncertainty in segmentation. A follow-up work to this paper was ranked second in the same task in the same challenge a year later.

We systematically assessed the segmentation performance and computational costs of the technique called pseudo-3D as a function of the number of input slices in Paper II. We compared the results to typical two-dimensional (2D) and three-dimensional (3D) CNNs and a method called triplanar orthogonal 2D. The typical pseudo-3D approach considers adjacent slices to be several image input channels. We discovered that a substantial benefit from employing multiple input slices was apparent for a specific input size.

We introduced a novel loss function in Paper III to address diverse issues, including imbalanced datasets, partially labeled data, and incremental learning. The proposed loss function adjusts to the given data to use all accessible data, even if some lack annotations. We show that the suggested loss function also performs well in an incremental learning context, where an existing model can be modified to incorporate the delineations of newly appearing organs semi-automatically.

In Paper IV, we proposed a novel method for compressing high-dimensional activation maps, which are the primary source of memory use in modern systems. We examined three distinct compression methods for the activation maps to accomplishing this. We demonstrated that the proposed method induces a regularization effect that acts on the layer weight gradients. By employing the proposed technique, we reduced activation map memory usage by up to 95%.

We investigated the use of generative adversarial networks (GANs) to enlarge a small dataset by generating synthetic images in Paper V. We use the real and generated data during training CNNs for the downstream segmentation tasks. Inspired by an existing GAN, we proposed a conditional version to generate high-dimensional and high-quality medical images of different modalities and their corresponding label maps. We evaluated the quality of the generated medical images and the effect of this augmentation on the performance of the segmentation task on six datasets.

Ort, förlag, år, upplaga, sidor
Umeå: Umeå University , 2023. , s. 101
Serie
Umeå University medical dissertations, ISSN 0346-6612 ; 2219
Nyckelord [en]
radiotherapy, medical imaging, deep learning, convolutional neural network, generative adversarial network, data augmentation, semantic segmentation, classification, activation map compression
Nationell ämneskategori
Datavetenskap (datalogi) Radiologi och bildbehandling Medicinsk bildvetenskap
Forskningsämne
radiofysik
Identifikatorer
URN: urn:nbn:se:umu:diva-203993ISBN: 978-91-7855-956-5 (tryckt)ISBN: 978-91-7855-957-2 (digital)OAI: oai:DiVA.org:umu-203993DiVA, id: diva2:1730514
Disputation
2023-02-24, Bergasalen, Byggnad 27, Entré Syd, Norrlands universitetssjukhus (NUS), Umeå, 09:00 (Engelska)
Opponent
Handledare
Tillgänglig från: 2023-02-03 Skapad: 2023-01-24 Senast uppdaterad: 2025-02-09Bibliografiskt granskad
Delarbeten
1. TuNet: End-to-End Hierarchical Brain Tumor Segmentation Using Cascaded Networks
Öppna denna publikation i ny flik eller fönster >>TuNet: End-to-End Hierarchical Brain Tumor Segmentation Using Cascaded Networks
2020 (Engelska)Ingår i: Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries: 5th International Workshop, BrainLes 2019, Held in Conjunction with MICCAI 2019, Shenzhen, China, October 17, 2019, Revised Selected Papers, Part I / [ed] Alessandro Crimi and Spyridon Bakas, Cham: Springer, 2020, s. 174-186Konferensbidrag, Publicerat paper (Refereegranskat)
Abstract [en]

Glioma is one of the most common types of brain tumors; it arises in the glial cells in the human brain and in the spinal cord. In addition to having a high mortality rate, glioma treatment is also very expensive. Hence, automatic and accurate segmentation and measurement from the early stages are critical in order to prolong the survival rates of the patients and to reduce the costs of the treatment. In the present work, we propose a novel end-to-end cascaded network for semantic segmentation in the Brain Tumors in Multimodal Magnetic Resonance Imaging Challenge 2019 that utilizes the hierarchical structure of the tumor sub-regions with ResNet-like blocks and Squeeze-and-Excitation modules after each convolution and concatenation block. By utilizing cross-validation, an average ensemble technique, and a simple post-processing technique, we obtained dice scores of 88.06, 80.84, and 80.29, and Hausdorff Distances (95th percentile) of 6.10, 5.17, and 2.21 for the whole tumor, tumor core, and enhancing tumor, respectively, on the online test set. The proposed method was ranked among the top in the task of Quantification of Uncertainty in Segmentation.

Ort, förlag, år, upplaga, sidor
Cham: Springer, 2020
Serie
Lecture Notes in Computer Science, ISSN 0302-9743 ; 11992
Nationell ämneskategori
Radiologi och bildbehandling Datorgrafik och datorseende
Identifikatorer
urn:nbn:se:umu:diva-177257 (URN)10.1007/978-3-030-46640-4_17 (DOI)000892506100017 ()2-s2.0-85085520133 (Scopus ID)978-3-030-46639-8 (ISBN)978-3-030-46640-4 (ISBN)
Konferens
MICCAI 2019, Shenzhen, China, October 17, 2019
Anmärkning

Also part of the Image Processing, Computer Vision, Pattern Recognition, and Graphics book sub series (LNIP, volume 11992)

Tillgänglig från: 2020-12-03 Skapad: 2020-12-03 Senast uppdaterad: 2025-02-01Bibliografiskt granskad
2. Evaluation of multislice inputs to convolutional neural networks for medical image segmentation
Öppna denna publikation i ny flik eller fönster >>Evaluation of multislice inputs to convolutional neural networks for medical image segmentation
2020 (Engelska)Ingår i: Medical physics (Lancaster), ISSN 0094-2405, Vol. 47, nr 12, s. 6216-6231Artikel i tidskrift (Refereegranskat) Published
Abstract [en]

Purpose: When using convolutional neural networks (CNNs) for segmentation of organs and lesions in medical images, the conventional approach is to work with inputs and outputs either as single slice [two-dimensional (2D)] or whole volumes [three-dimensional (3D)]. One common alternative, in this study denoted as pseudo-3D, is to use a stack of adjacent slices as input and produce a prediction for at least the central slice. This approach gives the network the possibility to capture 3D spatial information, with only a minor additional computational cost.

Methods: In this study, we systematically evaluate the segmentation performance and computational costs of this pseudo-3D approach as a function of the number of input slices, and compare the results to conventional end-to-end 2D and 3D CNNs, and to triplanar orthogonal 2D CNNs. The standard pseudo-3D method regards the neighboring slices as multiple input image channels. We additionally design and evaluate a novel, simple approach where the input stack is a volumetric input that is repeatably convolved in 3D to obtain a 2D feature map. This 2D map is in turn fed into a standard 2D network. We conducted experiments using two different CNN backbone architectures and on eight diverse data sets covering different anatomical regions, imaging modalities, and segmentation tasks.

Results: We found that while both pseudo-3D methods can process a large number of slices at once and still be computationally much more efficient than fully 3D CNNs, a significant improvement over a regular 2D CNN was only observed with two of the eight data sets. triplanar networks had the poorest performance of all the evaluated models. An analysis of the structural properties of the segmentation masks revealed no relations to the segmentation performance with respect to the number of input slices. A post hoc rank sum test which combined all metrics and data sets yielded that only our newly proposed pseudo-3D method with an input size of 13 slices outperformed almost all methods.

Conclusion: In the general case, multislice inputs appear not to improve segmentation results over using 2D or 3D CNNs. For the particular case of 13 input slices, the proposed novel pseudo-3D method does appear to have a slight advantage across all data sets compared to all other methods evaluated in this work.

Ort, förlag, år, upplaga, sidor
John Wiley & Sons, 2020
Nyckelord
convolutional neural network, deep learning, medical image segmentation, multislice
Nationell ämneskategori
Medicinsk bildvetenskap
Identifikatorer
urn:nbn:se:umu:diva-177165 (URN)10.1002/mp.14391 (DOI)000587714100001 ()33169365 (PubMedID)2-s2.0-85096775630 (Scopus ID)
Forskningsfinansiär
Vinnova
Tillgänglig från: 2020-12-08 Skapad: 2020-12-08 Senast uppdaterad: 2025-02-09Bibliografiskt granskad
3. A Data-Adaptive Loss Function for Incomplete Data and Incremental Learning in Semantic Image Segmentation
Öppna denna publikation i ny flik eller fönster >>A Data-Adaptive Loss Function for Incomplete Data and Incremental Learning in Semantic Image Segmentation
2022 (Engelska)Ingår i: IEEE Transactions on Medical Imaging, ISSN 0278-0062, E-ISSN 1558-254X, Vol. 41, nr 6, s. 1320-1330Artikel i tidskrift (Refereegranskat) Published
Abstract [en]

In the last years, deep learning has dramatically improved the performances in a variety of medical image analysis applications. Among different types of deep learning models, convolutional neural networks have been among the most successful and they have been used in many applications in medical imaging.

Training deep convolutional neural networks often requires large amounts of image data to generalize well to new unseen images. It is often time-consuming and expensive to collect large amounts of data in the medical image domain due to expensive imaging systems, and the need for experts to manually make ground truth annotations. A potential problem arises if new structures are added when a decision support system is already deployed and in use. Since the field of radiation therapy is constantly developing, the new structures would also have to be covered by the decision support system.

In the present work, we propose a novel loss function to solve multiple problems: imbalanced datasets, partially-labeled data, and incremental learning. The proposed loss function adapts to the available data in order to utilize all available data, even when some have missing annotations. We demonstrate that the proposed loss function also works well in an incremental learning setting, where an existing model is easily adapted to semi-automatically incorporate delineations of new organs when they appear. Experiments on a large in-house dataset show that the proposed method performs on par with baseline models, while greatly reducing the training time and eliminating the hassle of maintaining multiple models in practice.

Ort, förlag, år, upplaga, sidor
IEEE, 2022
Nyckelord
Adaptation models, Computational modeling, CT, Data models, Image segmentation, Incremental Learning, Medical Imaging, Missing Data, Predictive models, Semantic Image Segmentation, Task analysis, Training
Nationell ämneskategori
Datavetenskap (datalogi) Datorgrafik och datorseende
Identifikatorer
urn:nbn:se:umu:diva-191280 (URN)10.1109/TMI.2021.3139161 (DOI)000804690300003 ()34965206 (PubMedID)2-s2.0-85122295338 (Scopus ID)
Forskningsfinansiär
Vetenskapsrådet, 2018-05973Cancerforskningsfonden i Norrland, AMP 20-1027Cancerforskningsfonden i Norrland, LP 18-2182Region VästerbottenVinnova
Tillgänglig från: 2022-01-13 Skapad: 2022-01-13 Senast uppdaterad: 2025-02-01Bibliografiskt granskad
4. Compressing the activation maps in deep convolutional neural networks and the regularization effect of compression
Öppna denna publikation i ny flik eller fönster >>Compressing the activation maps in deep convolutional neural networks and the regularization effect of compression
(Engelska)Manuskript (preprint) (Övrigt vetenskapligt)
Nationell ämneskategori
Datavetenskap (datalogi)
Identifikatorer
urn:nbn:se:umu:diva-203991 (URN)
Tillgänglig från: 2023-01-24 Skapad: 2023-01-24 Senast uppdaterad: 2024-07-02
5. Using synthetic images to augment small medical image datasets
Öppna denna publikation i ny flik eller fönster >>Using synthetic images to augment small medical image datasets
(Engelska)Manuskript (preprint) (Övrigt vetenskapligt)
Nationell ämneskategori
Datavetenskap (datalogi)
Identifikatorer
urn:nbn:se:umu:diva-203992 (URN)
Tillgänglig från: 2023-01-24 Skapad: 2023-01-24 Senast uppdaterad: 2024-07-02

Open Access i DiVA

fulltext(5694 kB)719 nedladdningar
Filinformation
Filnamn FULLTEXT01.pdfFilstorlek 5694 kBChecksumma SHA-512
a4f09d9737af3935830031b4ad0fc711e0f8619509129e09a1facfae905495582659c52772260e739db2aa21b4a6587542f3c6673b41f3b315a15e9496661417
Typ fulltextMimetyp application/pdf
spikblad(157 kB)120 nedladdningar
Filinformation
Filnamn SPIKBLAD02.pdfFilstorlek 157 kBChecksumma SHA-512
12888c94dbea71e2b53367acca75c3827f478556d060b76f6293fc89ecb01bbf7ce66a6c043e1ec73c4c81df17422e1a0941e0d15686fce356fa114621402795
Typ spikbladMimetyp application/pdf

Person

Vu, Minh Hoang

Sök vidare i DiVA

Av författaren/redaktören
Vu, Minh Hoang
Av organisationen
Radiofysik
Datavetenskap (datalogi)Radiologi och bildbehandlingMedicinsk bildvetenskap

Sök vidare utanför DiVA

GoogleGoogle Scholar
Totalt: 719 nedladdningar
Antalet nedladdningar är summan av nedladdningar för alla fulltexter. Det kan inkludera t.ex tidigare versioner som nu inte längre är tillgängliga.

isbn
urn-nbn

Altmetricpoäng

isbn
urn-nbn
Totalt: 1705 träffar
RefereraExporteraLänk till posten
Permanent länk

Direktlänk
Referera
Referensformat
  • apa
  • ieee
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf