Umeå University's logo

umu.sePublications
Change search
Link to record
Permanent link

Direct link
Alternative names
Publications (10 of 14) Show all publications
Vu, M. H., Edler, D., Wibom, C., Löfstedt, T., Melin, B. S. & Rosvall, M. (2025). A unified framework for tabular generative modeling: loss functions, benchmarks, and improved multi-objective bayesian optimization approaches. Transactions on Machine Learning Research, 12
Open this publication in new window or tab >>A unified framework for tabular generative modeling: loss functions, benchmarks, and improved multi-objective bayesian optimization approaches
Show others...
2025 (English)In: Transactions on Machine Learning Research, E-ISSN 2835-8856, Vol. 12Article in journal (Refereed) Published
Abstract [en]

Deep learning (DL) models require extensive data to achieve strong performance and generalization. Deep generative models (DGMs) offer a solution by synthesizing data. Yet current approaches for tabular data often fail to preserve feature correlations and distributions during training, struggle with multi-metric hyperparameter selection, and lack comprehensive evaluation protocols. We address this gap with a unified framework that integrates training, hyperparameter tuning, and evaluation. First, we introduce a novel correlation- and distribution-aware loss function that regularizes DGMs, enhancing their ability to generate synthetic tabular data that faithfully represents the underlying data distributions. Theoretical analysis establishes stability and consistency guarantees. To enable principled hyper-parameter search via Bayesian optimization (BO), we also propose a new multi-objective aggregation strategy based on iterative objective refinement Bayesian optimization (IORBO), along with a comprehensive statistical testing framework. We validate the proposed approach using a benchmarking framework with twenty real-world datasets and ten established tabular DGM baselines. The correlation-aware loss function significantly improves the synthetic data fidelity and downstream machine learning (ML) performance, while IORBO consistently outperforms standard Bayesian optimization (SBO) in hyper-parameter selection. The unified framework advances tabular generative modeling beyond isolated method improvements. Code is available at: https://github.com/vuhoangminh/TabGen-Framework.

Place, publisher, year, edition, pages
Transactions on Machine Learning Research, 2025
National Category
Artificial Intelligence
Identifiers
urn:nbn:se:umu:diva-249190 (URN)
Available from: 2026-01-29 Created: 2026-01-29 Last updated: 2026-02-02Bibliographically approved
Tronchin, L., Vu, M. H., Soda, P. & Löfstedt, T. (2025). LatentAugment: data augmentation via guided manipulation of GAN's latent space. IEEE Transactions on Pattern Analysis and Machine Intelligence, 47(12), 11519-11533
Open this publication in new window or tab >>LatentAugment: data augmentation via guided manipulation of GAN's latent space
2025 (English)In: IEEE Transactions on Pattern Analysis and Machine Intelligence, ISSN 0162-8828, E-ISSN 1939-3539, Vol. 47, no 12, p. 11519-11533Article in journal (Refereed) Published
Abstract [en]

Data Augmentation (DA) is a technique to increase the quantity and diversity of the training data, and by that alleviate overfitting and improve generalisation. However, standard DA produces synthetic data for augmentation with limited diversity. Generative Adversarial Networks (GANs) may unlock additional information in a dataset by generating synthetic samples having the appearance of real images. However, these models struggle to simultaneously address three key requirements: fidelity and high-quality samples; diversity and mode coverage; and fast sampling. Indeed, GANs generate high-quality samples rapidly, but have poor mode coverage, limiting their adoption in DA applications. We propose LatentAugment, a DA strategy that overcomes the low diversity of GANs, opening up for use in DA applications. Without external supervision, LatentAugment modifies latent vectors and moves them into latent space regions to maximise the synthetic images' diversity and fidelity. It is also agnostic to the dataset and the downstream task. A wide set of experiments shows that LatentAugment improves the generalisation of a deep model translating from MRI-to-CT beating both standard DA as well GAN-based sampling. We further demonstrate its effectiveness when translating from low-energy mammograms to dual-energy subtracted images in contrast-enhanced spectral mammography. Moreover, still in comparison with GAN-based sampling, LatentAugment synthetic samples show superior mode coverage and diversity. Code is available at: https://github.com/ltronchin/LatentAugment.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
Computer vision, generalisation, Generative Adversarial Networks, image synthesis, medical imaging, mode coverage
National Category
Medical Imaging Computer Sciences
Identifiers
urn:nbn:se:umu:diva-244092 (URN)10.1109/TPAMI.2025.3598866 (DOI)40889306 (PubMedID)2-s2.0-105014973123 (Scopus ID)
Funder
Lions Cancerforskningsfond i Norr, LP 18- 2182Lions Cancerforskningsfond i Norr, LP 22-2319National Academic Infrastructure for Supercomputing in Sweden (NAISS)Swedish National Infrastructure for Computing (SNIC)
Available from: 2025-09-22 Created: 2025-09-22 Last updated: 2025-11-21Bibliographically approved
Vu, M. H., Garpebring, A., Nyholm, T. & Löfstedt, T. (2024). Compressing the Activation Maps in Deep Neural Networks and Its Regularizing Effect. Transactions on Machine Learning Research
Open this publication in new window or tab >>Compressing the Activation Maps in Deep Neural Networks and Its Regularizing Effect
2024 (English)In: Transactions on Machine Learning Research, E-ISSN 2835-8856Article in journal (Refereed) Published
Abstract [en]

Deep learning has dramatically improved performance in various image analysis applications in the last few years. However, recent deep learning architectures can be very large, with up to hundreds of layers and millions or even billions of model parameters that are impossible to fit into commodity graphics processing units. We propose a novel approach for compressing high-dimensional activation maps, the most memory-consuming part when training modern deep learning architectures. The proposed method can be used to compress the feature maps of a single layer, multiple layers, or the entire network according to specific needs. To this end, we also evaluated three different methods to compress the activation maps: Wavelet Transform, Discrete Cosine Transform, and Simple Thresholding. We performed experiments in two classification tasks for natural images and two semantic segmentation tasks for medical images. Using the proposed method, we could reduce the memory usage for activation maps by up to 95%. Additionally, we show that the proposed method induces a regularization effect that acts on the layer weight gradients. Code is available at https://github.com/vuhoangminh/Compressing-the-Activation-Maps-in-DNNs.

National Category
Computer Sciences
Identifiers
urn:nbn:se:umu:diva-236503 (URN)2-s2.0-85219517351 (Scopus ID)
Available from: 2025-03-18 Created: 2025-03-18 Last updated: 2025-03-18Bibliographically approved
Vu, M. H. (2023). Resource efficient automatic segmentation of medical images. (Doctoral dissertation). Umeå: Umeå University
Open this publication in new window or tab >>Resource efficient automatic segmentation of medical images
2023 (English)Doctoral thesis, comprehensive summary (Other academic)
Alternative title[sv]
Resurseffektiv automatisk medicinsk bildsegmentering
Abstract [en]

Cancer is one of the leading causes of death worldwide. In 2020, there were around 10 million cancer deaths and nearly 20 million new cancer cases in the world. Radiation therapy is essential in cancer treatments because half of the cancer patients receive radiation therapy at some point. During a radiotherapy treatment planning (RTP), an oncologist must manually outline two types of areas of the patient’s body: target, which will be treated, and organs-at-risks (OARs), which are essential to avoid. This step is called delineation. The purpose of the delineation is to generate a sufficient dose plan that can provide adequate radiation dose to a tumor and limit the radiation exposure to healthy tissue. Therefore, accurate delineations are essential to achieve this goal.

Delineation is tedious and demanding for oncologists because it requires hours of concentrating work doing a repeated job. This is a RTP bottleneck which is often time- and resource-intensive. Current software, such as atlasbased techniques, can assist with this procedure by registering the patient’s anatomy to a predetermined anatomical map. However, the atlas-based methods are often slowed down and erroneous for patients with abnormal anatomies.

In recent years, deep learning (DL) methods, particularly convolutional neural networks (CNNs), have led to breakthroughs in numerous medical imaging applications. The core benefits of CNNs are weight sharing and that they can automatically detect important visual features. A typical application of CNNs for medical images is to automatically segment tumors, organs, and structures, which is assumed to save radiation oncologists much time when delineating. This thesis contributes to resource efficient automatic segmentation and covers different aspects of resource efficiency.

In Paper I, we proposed a novel end-to-end cascaded network for semantic segmentation in brain tumors in the multi-modal magnetic resonance imaging challenge in 2019. The proposed method used the hierarchical structure of the tumor sub-regions and was one of the top-ranking teams in the task of quantification of uncertainty in segmentation. A follow-up work to this paper was ranked second in the same task in the same challenge a year later.

We systematically assessed the segmentation performance and computational costs of the technique called pseudo-3D as a function of the number of input slices in Paper II. We compared the results to typical two-dimensional (2D) and three-dimensional (3D) CNNs and a method called triplanar orthogonal 2D. The typical pseudo-3D approach considers adjacent slices to be several image input channels. We discovered that a substantial benefit from employing multiple input slices was apparent for a specific input size.

We introduced a novel loss function in Paper III to address diverse issues, including imbalanced datasets, partially labeled data, and incremental learning. The proposed loss function adjusts to the given data to use all accessible data, even if some lack annotations. We show that the suggested loss function also performs well in an incremental learning context, where an existing model can be modified to incorporate the delineations of newly appearing organs semi-automatically.

In Paper IV, we proposed a novel method for compressing high-dimensional activation maps, which are the primary source of memory use in modern systems. We examined three distinct compression methods for the activation maps to accomplishing this. We demonstrated that the proposed method induces a regularization effect that acts on the layer weight gradients. By employing the proposed technique, we reduced activation map memory usage by up to 95%.

We investigated the use of generative adversarial networks (GANs) to enlarge a small dataset by generating synthetic images in Paper V. We use the real and generated data during training CNNs for the downstream segmentation tasks. Inspired by an existing GAN, we proposed a conditional version to generate high-dimensional and high-quality medical images of different modalities and their corresponding label maps. We evaluated the quality of the generated medical images and the effect of this augmentation on the performance of the segmentation task on six datasets.

Place, publisher, year, edition, pages
Umeå: Umeå University, 2023. p. 101
Series
Umeå University medical dissertations, ISSN 0346-6612 ; 2219
Keywords
radiotherapy, medical imaging, deep learning, convolutional neural network, generative adversarial network, data augmentation, semantic segmentation, classification, activation map compression
National Category
Computer Sciences Radiology, Nuclear Medicine and Medical Imaging Medical Imaging
Research subject
radiation physics
Identifiers
urn:nbn:se:umu:diva-203993 (URN)978-91-7855-956-5 (ISBN)978-91-7855-957-2 (ISBN)
Public defence
2023-02-24, Bergasalen, Byggnad 27, Entré Syd, Norrlands universitetssjukhus (NUS), Umeå, 09:00 (English)
Opponent
Supervisors
Available from: 2023-02-03 Created: 2023-01-24 Last updated: 2025-02-09Bibliographically approved
Vu, M. H., Norman, G., Nyholm, T. & Löfstedt, T. (2022). A Data-Adaptive Loss Function for Incomplete Data and Incremental Learning in Semantic Image Segmentation. IEEE Transactions on Medical Imaging, 41(6), 1320-1330
Open this publication in new window or tab >>A Data-Adaptive Loss Function for Incomplete Data and Incremental Learning in Semantic Image Segmentation
2022 (English)In: IEEE Transactions on Medical Imaging, ISSN 0278-0062, E-ISSN 1558-254X, Vol. 41, no 6, p. 1320-1330Article in journal (Refereed) Published
Abstract [en]

In the last years, deep learning has dramatically improved the performances in a variety of medical image analysis applications. Among different types of deep learning models, convolutional neural networks have been among the most successful and they have been used in many applications in medical imaging.

Training deep convolutional neural networks often requires large amounts of image data to generalize well to new unseen images. It is often time-consuming and expensive to collect large amounts of data in the medical image domain due to expensive imaging systems, and the need for experts to manually make ground truth annotations. A potential problem arises if new structures are added when a decision support system is already deployed and in use. Since the field of radiation therapy is constantly developing, the new structures would also have to be covered by the decision support system.

In the present work, we propose a novel loss function to solve multiple problems: imbalanced datasets, partially-labeled data, and incremental learning. The proposed loss function adapts to the available data in order to utilize all available data, even when some have missing annotations. We demonstrate that the proposed loss function also works well in an incremental learning setting, where an existing model is easily adapted to semi-automatically incorporate delineations of new organs when they appear. Experiments on a large in-house dataset show that the proposed method performs on par with baseline models, while greatly reducing the training time and eliminating the hassle of maintaining multiple models in practice.

Place, publisher, year, edition, pages
IEEE, 2022
Keywords
Adaptation models, Computational modeling, CT, Data models, Image segmentation, Incremental Learning, Medical Imaging, Missing Data, Predictive models, Semantic Image Segmentation, Task analysis, Training
National Category
Computer Sciences Computer graphics and computer vision
Identifiers
urn:nbn:se:umu:diva-191280 (URN)10.1109/TMI.2021.3139161 (DOI)000804690300003 ()34965206 (PubMedID)2-s2.0-85122295338 (Scopus ID)
Funder
Swedish Research Council, 2018-05973Cancerforskningsfonden i Norrland, AMP 20-1027Cancerforskningsfonden i Norrland, LP 18-2182Region VästerbottenVinnova
Available from: 2022-01-13 Created: 2022-01-13 Last updated: 2025-02-01Bibliographically approved
Mehta, R., Filos, A., Baid, U., Sako, C., McKinley, R., Rebsamen, M., . . . Arbel, T. (2022). QU-BraTS: MICCAI BraTS 2020 Challenge on QuantifyingUncertainty in Brain Tumor Segmentation - Analysis of Ranking Scores and Benchmarking Results. Journal of Machine Learning for Biomedical Imaging, 1-54, Article ID 026.
Open this publication in new window or tab >>QU-BraTS: MICCAI BraTS 2020 Challenge on QuantifyingUncertainty in Brain Tumor Segmentation - Analysis of Ranking Scores and Benchmarking Results
Show others...
2022 (English)In: Journal of Machine Learning for Biomedical Imaging, ISSN 2766-905X, p. 1-54, article id 026Article in journal (Refereed) Published
Abstract [en]

Deep learning (DL) models have provided the state-of-the-art performance in a wide variety of medical imaging benchmarking challenges, including the Brain Tumor Segmentation (BraTS) challenges. However, the task of focal pathology multi-compartment segmentation (e.g., tumor and lesion sub-regions) is particularly challenging, and potential errors hinder the translation of DL models into clinical workflows. Quantifying the reliability of DL model predictions in the form of uncertainties, could enable clinical review of the most uncertain regions, thereby building trust and paving the way towards clinical translation. Recently, a number of uncertainty estimation methods have been introduced for DL medical image segmentation tasks. Developing scores to evaluate and compare the performance of uncertainty measures will assist the end-user in making more informed decisions. In this study, we explore and evaluate a score developed during the BraTS 2019-2020 task on uncertainty quantification (QU-BraTS), and designed to assess and rank uncertainty estimates for brain tumor multi-compartment segmentation. This score (1) rewards uncertainty estimates that produce high confidence in correct assertions, and those that assign low confidence levels at incorrect assertions, and (2) penalizes uncertainty measures that lead to a higher percentages of under-confident correct assertions. We further benchmark the segmentation uncertainties generated by 14 independent participating teams of QU-BraTS 2020, all of which also participated in the main BraTS segmentation task. Overall, our findings confirm the importance and complementary value that uncertainty estimates provide to segmentation algorithms, and hence highlight the need for uncertainty quantification in medical image analyses. Our evaluation code is made publicly available at https://github.com/RagMeh11/QU-BraTS

Keywords
Uncertainty Quantification, Trustworthiness, Segmentation, Brain Tumors, Deep Learning, Neuro-Oncology, Glioma, Glioblastoma
National Category
Computer graphics and computer vision Other Medical Sciences not elsewhere specified
Identifiers
urn:nbn:se:umu:diva-198857 (URN)
Available from: 2022-08-26 Created: 2022-08-26 Last updated: 2025-02-01Bibliographically approved
Vu, M. H., Nyholm, T. & Löfstedt, T. (2021). Multi-decoder Networks with Multi-denoising Inputs for Tumor Segmentation. In: Alessandro Crimi, Spyridon Bakas (Ed.), Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries: . Paper presented at 6th International MICCAI Brainlesion Workshop, BrainLes 2020 Held in Conjunction with 23rd Medical Image Computing for Computer Assisted Intervention Conference, MICCAI 2020, Lima, Peru, October 4, 2020 (pp. 412-423). Springer
Open this publication in new window or tab >>Multi-decoder Networks with Multi-denoising Inputs for Tumor Segmentation
2021 (English)In: Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries / [ed] Alessandro Crimi, Spyridon Bakas, Springer, 2021, p. 412-423Conference paper, Published paper (Refereed)
Abstract [en]

Automatic segmentation of brain glioma from multimodal MRI scans plays a key role in clinical trials and practice. Unfortunately, manual segmentation is very challenging, time-consuming, costly, and often inaccurate despite human expertise due to the high variance and high uncertainty in the human annotations. In the present work, we develop an end-to-end deep-learning-based segmentation method using a multi-decoder architecture by jointly learning three separate sub-problems using a partly shared encoder. We also propose to apply smoothing methods to the input images to generate denoised versions as additional inputs to the network. The validation performance indicates an improvement when using the proposed method. The proposed method was ranked 2nd in the task of Quantification of Uncertainty in Segmentation in the Brain Tumors in Multimodal Magnetic Resonance Imaging Challenge 2020.

Place, publisher, year, edition, pages
Springer, 2021
Series
Lecture Notes in Computer Science, ISSN 0302-9743, E-ISSN 1611-3349 ; 12658
Keywords
Brain tumor segmentation, Deep learning, Ensemble, Medical imaging, MRI, Uncertainty estimation
National Category
Medical Imaging Computer graphics and computer vision
Identifiers
urn:nbn:se:umu:diva-186225 (URN)10.1007/978-3-030-72084-1_37 (DOI)000892566900037 ()2-s2.0-85106168970 (Scopus ID)9783030720834 (ISBN)
Conference
6th International MICCAI Brainlesion Workshop, BrainLes 2020 Held in Conjunction with 23rd Medical Image Computing for Computer Assisted Intervention Conference, MICCAI 2020, Lima, Peru, October 4, 2020
Funder
Cancerforskningsfonden i NorrlandRegion VästerbottenVinnova
Note

Also part of the Image Processing, Computer Vision, Pattern Recognition, and Graphics book sub series (LNIP, volume 12658)

Available from: 2021-07-16 Created: 2021-07-16 Last updated: 2025-02-09Bibliographically approved
Vu, M. H., Löfstedt, T., Nyholm, T. & Sznitman, R. (2020). A Question-Centric Model for Visual Question Answering in Medical Imaging. IEEE Transactions on Medical Imaging, 39(9), 2856-2868
Open this publication in new window or tab >>A Question-Centric Model for Visual Question Answering in Medical Imaging
2020 (English)In: IEEE Transactions on Medical Imaging, ISSN 0278-0062, E-ISSN 1558-254X, Vol. 39, no 9, p. 2856-2868Article in journal (Refereed) Published
Abstract [en]

Deep learning methods have proven extremely effective at performing a variety of medical image analysis tasks. With their potential use in clinical routine, their lack of transparency has however been one of their few weak points, raising concerns regarding their behavior and failure modes. While most research to infer model behavior has focused on indirect strategies that estimate prediction uncertainties and visualize model support in the input image space, the ability to explicitly query a prediction model regarding its image content offers a more direct way to determine the behavior of trained models. To this end, we present a novel Visual Question Answering approach that allows an image to be queried by means of a written question. Experiments on a variety of medical and natural image datasets show that by fusing image and question features in a novel way, the proposed approach achieves an equal or higher accuracy compared to current methods.

Place, publisher, year, edition, pages
IEEE, 2020
Keywords
Visual question answering, deep learning, medical images, medical questions and answers
National Category
Computer graphics and computer vision Natural Language Processing Other Medical Sciences not elsewhere specified
Research subject
Computer Science; Computerized Image Analysis
Identifiers
urn:nbn:se:umu:diva-174694 (URN)10.1109/TMI.2020.2978284 (DOI)000566339800013 ()2-s2.0-85090173541 (Scopus ID)
Funder
Cancerforskningsfonden i NorrlandRegion VästerbottenVinnova
Available from: 2020-09-02 Created: 2020-09-02 Last updated: 2025-02-01Bibliographically approved
Vu, M. H., Grimbergen, G., Nyholm, T. & Löfstedt, T. (2020). Evaluation of multislice inputs to convolutional neural networks for medical image segmentation. Medical physics (Lancaster), 47(12), 6216-6231
Open this publication in new window or tab >>Evaluation of multislice inputs to convolutional neural networks for medical image segmentation
2020 (English)In: Medical physics (Lancaster), ISSN 0094-2405, Vol. 47, no 12, p. 6216-6231Article in journal (Refereed) Published
Abstract [en]

Purpose: When using convolutional neural networks (CNNs) for segmentation of organs and lesions in medical images, the conventional approach is to work with inputs and outputs either as single slice [two-dimensional (2D)] or whole volumes [three-dimensional (3D)]. One common alternative, in this study denoted as pseudo-3D, is to use a stack of adjacent slices as input and produce a prediction for at least the central slice. This approach gives the network the possibility to capture 3D spatial information, with only a minor additional computational cost.

Methods: In this study, we systematically evaluate the segmentation performance and computational costs of this pseudo-3D approach as a function of the number of input slices, and compare the results to conventional end-to-end 2D and 3D CNNs, and to triplanar orthogonal 2D CNNs. The standard pseudo-3D method regards the neighboring slices as multiple input image channels. We additionally design and evaluate a novel, simple approach where the input stack is a volumetric input that is repeatably convolved in 3D to obtain a 2D feature map. This 2D map is in turn fed into a standard 2D network. We conducted experiments using two different CNN backbone architectures and on eight diverse data sets covering different anatomical regions, imaging modalities, and segmentation tasks.

Results: We found that while both pseudo-3D methods can process a large number of slices at once and still be computationally much more efficient than fully 3D CNNs, a significant improvement over a regular 2D CNN was only observed with two of the eight data sets. triplanar networks had the poorest performance of all the evaluated models. An analysis of the structural properties of the segmentation masks revealed no relations to the segmentation performance with respect to the number of input slices. A post hoc rank sum test which combined all metrics and data sets yielded that only our newly proposed pseudo-3D method with an input size of 13 slices outperformed almost all methods.

Conclusion: In the general case, multislice inputs appear not to improve segmentation results over using 2D or 3D CNNs. For the particular case of 13 input slices, the proposed novel pseudo-3D method does appear to have a slight advantage across all data sets compared to all other methods evaluated in this work.

Place, publisher, year, edition, pages
John Wiley & Sons, 2020
Keywords
convolutional neural network, deep learning, medical image segmentation, multislice
National Category
Medical Imaging
Identifiers
urn:nbn:se:umu:diva-177165 (URN)10.1002/mp.14391 (DOI)000587714100001 ()33169365 (PubMedID)2-s2.0-85096775630 (Scopus ID)
Funder
Vinnova
Available from: 2020-12-08 Created: 2020-12-08 Last updated: 2025-02-09Bibliographically approved
Vu, M. H., Nyholm, T. & Löfstedt, T. (2020). TuNet: End-to-End Hierarchical Brain Tumor Segmentation Using Cascaded Networks. In: Alessandro Crimi and Spyridon Bakas (Ed.), Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries: 5th International Workshop, BrainLes 2019, Held in Conjunction with MICCAI 2019, Shenzhen, China, October 17, 2019, Revised Selected Papers, Part I. Paper presented at MICCAI 2019, Shenzhen, China, October 17, 2019 (pp. 174-186). Cham: Springer
Open this publication in new window or tab >>TuNet: End-to-End Hierarchical Brain Tumor Segmentation Using Cascaded Networks
2020 (English)In: Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries: 5th International Workshop, BrainLes 2019, Held in Conjunction with MICCAI 2019, Shenzhen, China, October 17, 2019, Revised Selected Papers, Part I / [ed] Alessandro Crimi and Spyridon Bakas, Cham: Springer, 2020, p. 174-186Conference paper, Published paper (Refereed)
Abstract [en]

Glioma is one of the most common types of brain tumors; it arises in the glial cells in the human brain and in the spinal cord. In addition to having a high mortality rate, glioma treatment is also very expensive. Hence, automatic and accurate segmentation and measurement from the early stages are critical in order to prolong the survival rates of the patients and to reduce the costs of the treatment. In the present work, we propose a novel end-to-end cascaded network for semantic segmentation in the Brain Tumors in Multimodal Magnetic Resonance Imaging Challenge 2019 that utilizes the hierarchical structure of the tumor sub-regions with ResNet-like blocks and Squeeze-and-Excitation modules after each convolution and concatenation block. By utilizing cross-validation, an average ensemble technique, and a simple post-processing technique, we obtained dice scores of 88.06, 80.84, and 80.29, and Hausdorff Distances (95th percentile) of 6.10, 5.17, and 2.21 for the whole tumor, tumor core, and enhancing tumor, respectively, on the online test set. The proposed method was ranked among the top in the task of Quantification of Uncertainty in Segmentation.

Place, publisher, year, edition, pages
Cham: Springer, 2020
Series
Lecture Notes in Computer Science, ISSN 0302-9743 ; 11992
National Category
Radiology, Nuclear Medicine and Medical Imaging Computer graphics and computer vision
Identifiers
urn:nbn:se:umu:diva-177257 (URN)10.1007/978-3-030-46640-4_17 (DOI)000892506100017 ()2-s2.0-85085520133 (Scopus ID)978-3-030-46639-8 (ISBN)978-3-030-46640-4 (ISBN)
Conference
MICCAI 2019, Shenzhen, China, October 17, 2019
Note

Also part of the Image Processing, Computer Vision, Pattern Recognition, and Graphics book sub series (LNIP, volume 11992)

Available from: 2020-12-03 Created: 2020-12-03 Last updated: 2025-02-01Bibliographically approved
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0002-2391-1419

Search in DiVA

Show all publications