Logo: to the web site of Uppsala University

uu.sePublications from Uppsala University
Change search
Link to record
Permanent link

Direct link
Publications (10 of 34) Show all publications
Ahmad, N., Öfverstedt, J., Tarai, S., Bergström, G., Ahlström, H. & Kullberg, J. (2024). Interpretable Uncertainty-Aware Deep Regression with Cohort Saliency Analysis for Three-Slice CT Imaging Studies. In: Ninon Burgos; Caroline Petitjean; Maria Vakalopoulou; Stergios Christodoulidis; Pierrick Coupe; Hervé Delingette; Carole Lartizien; Diana Mateus (Ed.), Proceedings of The 7nd International Conference on Medical Imaging with Deep Learning: . Paper presented at The 7th International Conference on Medical Imaging with Deep Learning, 3-5 July, 2024, Paris, France (pp. 17-32). MLResearchPress
Open this publication in new window or tab >>Interpretable Uncertainty-Aware Deep Regression with Cohort Saliency Analysis for Three-Slice CT Imaging Studies
Show others...
2024 (English)In: Proceedings of The 7nd International Conference on Medical Imaging with Deep Learning / [ed] Ninon Burgos; Caroline Petitjean; Maria Vakalopoulou; Stergios Christodoulidis; Pierrick Coupe; Hervé Delingette; Carole Lartizien; Diana Mateus, MLResearchPress , 2024, p. 17-32Conference paper, Published paper (Refereed)
Abstract [en]

Obesity is associated with an increased risk of morbidity and mortality. Achieving a healthy body composition, which involves maintaining a balance between fat and muscle mass, is important for metabolic health and preventing chronic diseases. Computed tomography (CT) imaging offers detailed insights into the body’s internal structure, aiding in understanding body composition and its related factors. In this feasibility study, we utilized CT image data from 2,724 subjects from the large metabolic health cohort studies SCAPIS and IGT. We train and evaluate an uncertainty-aware deep regression based ResNet-50 network, which outputs its prediction as mean and variance, for quantification of cross-sectional areas of liver, visceral adipose tissue (VAT), and thigh muscle. This was done using collages of three single-slice CT images from the liver, abdomen, and thigh regions. The model demonstrated promising results with the evaluation metrics – including R-squared (R2) and mean absolute error (MAE) for predictions. Additionally, for interpretability, the model was evaluated with saliency analysis based on Grad-CAM (Gradient-weighted Class Activation Mapping) at stages 2, 3, and 4 of the network. Deformable image registration to a template subject further enabled cohort saliency analysis that provide group-wise visualization of image regions of importance for associations to biomarkers of interest. We found that the networks focus on relevant regions for each target, according to prior knowledge.

Place, publisher, year, edition, pages
MLResearchPress, 2024
Series
Proceedings of Machine Learning Research, PMLR, E-ISSN 2640-3498 ; 250
National Category
Radiology and Medical Imaging Epidemiology
Research subject
Machine learning
Identifiers
urn:nbn:se:uu:diva-554960 (URN)
Conference
The 7th International Conference on Medical Imaging with Deep Learning, 3-5 July, 2024, Paris, France
Funder
Swedish Research Council, 2019-04756EXODIAB - Excellence of Diabetes Research in SwedenSwedish Heart Lung Foundation
Available from: 2025-04-18 Created: 2025-04-18 Last updated: 2025-06-10Bibliographically approved
Tarai, S., Lundström, E., Öfverstedt, J., Jönsson, H., Ahmad, N., Ahlström, H. & Kullberg, J. (2024). Prediction of Total Metabolic Tumor Volume from Tissue-Wise FDG-PET/CT Projections, Interpreted Using Cohort Saliency Analysis. In: Moi Hoon Yap; Connah Kendrick; Ardhendu Behera; Timothy Cootes; Reyer Zwiggelaar (Ed.), Medical Image Understanding and Analysis: 28th Annual Conference, MIUA 2024, Manchester, UK, July 24–26, 2024, Proceedings, Part II. Paper presented at 28th Annual Conference, MIUA 2024, Manchester, UK, July 24–26, 2024 (pp. 242-255). Cham: Springer
Open this publication in new window or tab >>Prediction of Total Metabolic Tumor Volume from Tissue-Wise FDG-PET/CT Projections, Interpreted Using Cohort Saliency Analysis
Show others...
2024 (English)In: Medical Image Understanding and Analysis: 28th Annual Conference, MIUA 2024, Manchester, UK, July 24–26, 2024, Proceedings, Part II / [ed] Moi Hoon Yap; Connah Kendrick; Ardhendu Behera; Timothy Cootes; Reyer Zwiggelaar, Cham: Springer, 2024, p. 242-255Conference paper, Published paper (Refereed)
Abstract [en]

Early and accurate prediction of clinical outcomes holds great potential for patient prognostics and personalized treatment planning. Development of automated methods for estimation of medical image-based clinical parameters (e.g. total metabolic tumor volume, TMTV) could pave the way for predicting advanced clinical outcomes not explicitly available in the images, such as overall survival. We developed an automated framework that extracted tissue-wise multi-channel 2D projections from whole-body FDG-PET/CT volumes, by separating tissues based on CT Hounsfield units, and used a DenseNet-121 to estimate the TMTV from the projections. For transparency and interpretability, an image registration-based cohort saliency analysis was proposed. The network was applied on the autoPET cohort (501 scans representing lymphoma, lung cancer, melanoma) and evaluated using a single channel method (baseline) and a multi-channel method (proposed), for the purpose of comparison. The incorporation of multiple channels demonstrated an advantage in the TMTV prediction, outperforming the baseline model with a ΔMAE = –14.34 ml; ΔR2 = 0.1584; ΔICC = 0.1316 (p-value = 0.0098). The Pearson correlation coefficient (r) was computed between the ground truth (GT) tumor projections and the aggregated saliency maps. Statistical comparison, via bootstrapping, showed that the proposed model consistently outperformed the baseline, with significantly higher r across all cancer types and both sexes, except for melanoma in females. This implied that the aggregated saliency maps generated by the proposed model showed higher correspondence with the GT, compared to the baseline model. Our approach offers a promising and interpretable framework for the automated prediction of TMTV, with further potential to also predict advanced clinical outcomes.

Place, publisher, year, edition, pages
Cham: Springer, 2024
Series
Lecture Notes in Computer Science (LNCS), ISSN 0302-9743, E-ISSN 1611-3349 ; 14860
Keywords
Whole-body PET/CT projections, Deep regression, Cohort saliency analysis
National Category
Cancer and Oncology Radiology, Nuclear Medicine and Medical Imaging
Identifiers
urn:nbn:se:uu:diva-541110 (URN)10.1007/978-3-031-66958-3_18 (DOI)001314340700018 ()978-3-031-66957-6 (ISBN)978-3-031-66958-3 (ISBN)
Conference
28th Annual Conference, MIUA 2024, Manchester, UK, July 24–26, 2024
Funder
Swedish Cancer Society, 201303 PjF 01Stiftelsen för Makarna Gottfrid och Karin Erikssons fond
Available from: 2024-10-28 Created: 2024-10-28 Last updated: 2025-05-13Bibliographically approved
Nordling, L., Öfverstedt, J., Lindblad, J. & Sladoje, N. (2023). Contrastive Learning of Equivariant Image Representations for Multimodal Deformable Registration. In: 2023 IEEE 20th International Symposium on Biomedical Imaging (ISBI): . Paper presented at IEEE 20th International Symposium on Biomedical Imaging, Cartagena, Colombia, 18-21 April, 2023. Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Contrastive Learning of Equivariant Image Representations for Multimodal Deformable Registration
2023 (English)In: 2023 IEEE 20th International Symposium on Biomedical Imaging (ISBI), Institute of Electrical and Electronics Engineers (IEEE), 2023Conference paper, Published paper (Refereed)
Abstract [en]

We propose a method for multimodal deformable image registration which combines a powerful deep learning approach to generate CoMIRs, dense image-like representations of multimodal image pairs, with INSPIRE, a robust framework for monomodal deformable image registration. We introduce new equivariance constraints to improve the consistency of CoMIRs under deformation. We evaluate the method on three publicly available multimodal datasets, one remote sensing, one histological, and one cytological. The proposed method demonstrates general applicability and consistently outperforms state-of-the-art registration tools \elastixname and VoxelMorph. We share source code of the proposed method and complete experimental setup as open-source at: https://github.com/MIDA-group/CoMIR_INSPIRE.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2023
Series
International Symposium on Biomedical Imaging, ISSN 1945-7928, E-ISSN 1945-8452
Keywords
image alignment, correlative imaging, representation learning, equivariance
National Category
Computer Sciences
Identifiers
urn:nbn:se:uu:diva-510266 (URN)10.1109/ISBI53787.2023.10230465 (DOI)001062050500143 ()978-1-6654-7358-3 (ISBN)978-1-6654-7359-0 (ISBN)
Conference
IEEE 20th International Symposium on Biomedical Imaging, Cartagena, Colombia, 18-21 April, 2023
Funder
Vinnova, 2017-02447Vinnova, 2020-03611Vinnova, 2021-01420Swedish Research Council, 2017-04385Swedish National Infrastructure for Computing (SNIC)Swedish Research Council, 2018-05973
Available from: 2023-08-25 Created: 2023-08-25 Last updated: 2023-11-14Bibliographically approved
Öfverstedt, J., Lindblad, J. & Sladoje, N. (2023). INSPIRE: Intensity and Spatial Information-Based Deformable Image Registration. PLOS ONE, 18(3), Article ID e0282432.
Open this publication in new window or tab >>INSPIRE: Intensity and Spatial Information-Based Deformable Image Registration
2023 (English)In: PLOS ONE, E-ISSN 1932-6203, Vol. 18, no 3, article id e0282432Article in journal (Refereed) Published
Abstract [en]

We present INSPIRE, a top-performing general-purpose method for deformable image registration. INSPIRE brings distance measures which combine intensity and spatial information into an elastic B-splines-based transformation model and incorporates an inverse inconsistency penalization supporting symmetric registration performance. We introduce several theoretical and algorithmic solutions which provide high computational efficiency and thereby applicability of the proposed framework in a wide range of real scenarios. We show that INSPIRE delivers highly accurate, as well as stable and robust registration results. We evaluate the method on a 2D dataset created from retinal images, characterized by presence of networks of thin structures. Here INSPIRE exhibits excellent performance, substantially outperforming the widely used reference methods. We also evaluate INSPIRE on the Fundus Image Registration Dataset (FIRE), which consists of 134 pairs of separately acquired retinal images. INSPIRE exhibits excellent performance on the FIRE dataset, substantially outperforming several domain-specific methods. We also evaluate the method on four benchmark datasets of 3D magnetic resonance images of brains, for a total of 2088 pairwise registrations. A comparison with 17 other state-of-the-art methods reveals that INSPIRE provides the best overall performance. Code is available at github.com/MIDA-group/inspire

Place, publisher, year, edition, pages
Public Library of Science (PLoS), 2023
Keywords
Image registration, splines (mathematics), deformable models, set distance, gradient methods, optimization, cost function, iterative methods, fuzzy sets
National Category
Computer graphics and computer vision Medical Imaging
Research subject
Computerized Image Processing
Identifiers
urn:nbn:se:uu:diva-429025 (URN)10.1371/journal.pone.0282432 (DOI)000945993100007 ()36867617 (PubMedID)
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)Vinnova, 2017-02447Vinnova, 2020-03611Vinnova, 2021-01420
Available from: 2020-12-18 Created: 2020-12-18 Last updated: 2025-02-09Bibliographically approved
Nordling, L., Öfverstedt, J., Lindblad, J. & Sladoje, N. (2023). Multimodal deformable image registration using contrastive learning of equivariant image representations. In: : . Paper presented at 40th Swedish Symposium on Image Analysis, Kolmårdens vildmarkshotell, Sweden, 13-15 March, 2023.
Open this publication in new window or tab >>Multimodal deformable image registration using contrastive learning of equivariant image representations
2023 (English)Conference paper, Oral presentation only (Other academic)
Abstract [en]

We propose a method for multimodal deformable image registration which combines a powerful deep learning approach to generate CoMIRs, dense image-like representations of multimodal image pairs, with INSPIRE, a robust framework for monomodal deformable image registration. We introduce new equivariance constraints to improve the consistency of CoMIRs under deformation. We evaluate the method on three publicly available multimodal datasets, one remote sensing, one histological, and one cytological. The proposed method demonstrates general applicability and consistently outperforms state-of-the-art registration tools \elastixname and VoxelMorph. We share source code of the proposed method and complete experimental setup as open-source at: https://github.com/MIDA-group/CoMIR_INSPIRE.

National Category
Computer Sciences
Identifiers
urn:nbn:se:uu:diva-510269 (URN)
Conference
40th Swedish Symposium on Image Analysis, Kolmårdens vildmarkshotell, Sweden, 13-15 March, 2023
Funder
Vinnova, 2017-02447Vinnova, 2020-03611Swedish Research Council, 2017-04385
Available from: 2023-08-25 Created: 2023-08-25 Last updated: 2023-08-30Bibliographically approved
Öfverstedt, J., Lindblad, J. & Sladoje, N. (2022). Cross-Sim-NGF: FFT-Based Global Rigid Multimodal Alignment of Image Volumes using Normalized Gradient Fields. In: : . Paper presented at 10th Workshop on Biomedical Image Registration. Springer Nature Springer Nature
Open this publication in new window or tab >>Cross-Sim-NGF: FFT-Based Global Rigid Multimodal Alignment of Image Volumes using Normalized Gradient Fields
2022 (English)Conference paper, Published paper (Refereed)
Abstract [en]

Multimodal image alignment involves finding spatial correspondences between volumes varying in appearance and structure. Automated alignment methods are often based on local optimization that can be highly sensitive to initialization. We propose a novel efficient algorithm for computing similarity of normalized gradient fields (NGF) in the frequency domain, which we globally optimize to achieve rigid multimodal 3D image alignment. We validate the method experimentally on a dataset comprised of 20 brain volumes acquired in four modalities (T1w, Flair, CT, [18F] FDG PET), synthetically displaced with known transformations. The proposed method exhibits excellent performance on all six possible modality combinations and outperforms the four considered reference methods by a large margin. An important advantage of the method is its speed; global rigid alignment of 3.4 Mvoxel volumes requires approximately 40 seconds of computation, and the proposed algorithm outperforms a direct algorithm for the same task by more than three orders of magnitude. Open-source code is provided.

Place, publisher, year, edition, pages
Springer NatureSpringer Nature, 2022
National Category
Medical Imaging
Research subject
Computerized Image Processing
Identifiers
urn:nbn:se:uu:diva-472285 (URN)10.1007/978-3-031-11203-4_17 (DOI)000883026000017 ()
Conference
10th Workshop on Biomedical Image Registration
Available from: 2022-04-07 Created: 2022-04-07 Last updated: 2025-02-09Bibliographically approved
Öfverstedt, J., Lindblad, J. & Sladoje, N. (2022). Efficient Algorithms for Global Multimodal Image Registration. In: : . Paper presented at Swedish Symposium on Image Analysis (SSBA).
Open this publication in new window or tab >>Efficient Algorithms for Global Multimodal Image Registration
2022 (English)Conference paper, Oral presentation only (Other academic)
Abstract [en]

Multimodal image registration is the process of finding spatial correspondences between images formed by different imaging techniques or under different conditions, to facilitate heterogeneous data fusion and correlative analysis. Two similarity measures widely used in multimodal image registration are mutual information (MI) and similarity of normalized gradient fields (NGF). We propose efficient algorithms for computing MI and similarity of NGF for all discrete axis-aligned shifts in the frequency domain. These fast algorithms enable highly reliable global registration of multimodal images, also for very large displacements,  which we confirm by their performance evaluation on a number of different pairs of modalities.

We consider four datasets, and observe that global maximization of MI is the best choice for two datasets/applications in 2D, while global maximization of similarity of NGF performs best on the remaining two datasets, of which one consists of 2D images, and the other consists of 3D data. This confirms the relevance of both methods; their properties recommend them for application in different scenarios.    

National Category
Medical Imaging
Research subject
Computerized Image Processing
Identifiers
urn:nbn:se:uu:diva-490950 (URN)
Conference
Swedish Symposium on Image Analysis (SSBA)
Available from: 2022-12-16 Created: 2022-12-16 Last updated: 2025-02-09Bibliographically approved
Öfverstedt, J., Lindblad, J. & Sladoje, N. (2022). Fast computation of mutual information in the frequency domain with applications to global multimodal image alignment. Pattern Recognition Letters, 159, 196-203
Open this publication in new window or tab >>Fast computation of mutual information in the frequency domain with applications to global multimodal image alignment
2022 (English)In: Pattern Recognition Letters, ISSN 0167-8655, E-ISSN 1872-7344, Vol. 159, p. 196-203Article in journal (Refereed) Published
Abstract [en]

Multimodal image alignment is the process of finding spatial correspondences between images formed by different imaging techniques or under different conditions, to facilitate heterogeneous data fusion and correlative analysis. The information-theoretic concept of mutual information (MI) is widely used as a similarity measure to guide multimodal alignment processes, where most works have focused on local maximization of MI, which typically works well only for small displacements. This points to a need for global maximization of MI, which has previously been computationally infeasible due to the high run-time complexity of existing algorithms. We propose an efficient algorithm for computing MI for all discrete displacements (formalized as the cross-mutual information function (CMIF)), which is based on cross-correlation computed in the frequency domain. We show that the algorithm is equivalent to a direct method while superior in terms of run-time. Furthermore, we propose a method for multimodal image alignment for transformation models with few degrees of freedom (e.g., rigid) based on the proposed CMIF-algorithm. We evaluate the efficacy of the proposed method on three distinct benchmark datasets, containing remote sensing images, cytological images, and histological images, and we observe excellent success-rates (in recovering known rigid transformations), overall outperforming alternative methods, including local optimization of MI, as well as several recent deep learning-based approaches. We also evaluate the run-times of a GPU implementation of the proposed algorithm and observe speed-ups from 100 to more than 10,000 times for realistic image sizes compared to a GPU implementation of a direct method. Code is shared as open-source at github.com/MIDA-group/globalign.

Place, publisher, year, edition, pages
ElsevierElsevier BV, 2022
Keywords
Mutual information, Image alignment, Global optimization, Multimodal, Entropy
National Category
Medical Imaging Computer graphics and computer vision
Research subject
Computerized Image Processing
Identifiers
urn:nbn:se:uu:diva-447807 (URN)10.1016/j.patrec.2022.05.022 (DOI)000833390400013 ()
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)Vinnova, 2017-02447Swedish Research Council, 2017-04385
Available from: 2021-06-30 Created: 2021-06-30 Last updated: 2025-02-09Bibliographically approved
Lu, J., Öfverstedt, J., Lindblad, J. & Sladoje, N. (2022). Is image-to-image translation the panacea for multimodal image registration?: A comparative study. PLOS ONE, 17(11), Article ID e0276196.
Open this publication in new window or tab >>Is image-to-image translation the panacea for multimodal image registration?: A comparative study
2022 (English)In: PLOS ONE, E-ISSN 1932-6203, Vol. 17, no 11, article id e0276196Article in journal (Refereed) Published
Abstract [en]

Despite current advancement in the field of biomedical image processing, propelled by the deep learning revolution, multimodal image registration, due to its several challenges, is still often performed manually by specialists. The recent success of image-to-image (I2I) translation in computer vision applications and its growing use in biomedical areas provide a tempting possibility of transforming the multimodal registration problem into a, potentially easier, monomodal one. We conduct an empirical study of the applicability of modern I2I translation methods for the task of rigid registration of multimodal biomedical and medical 2D and 3D images. We compare the performance of four Generative Adversarial Network (GAN)-based I2I translation methods and one contrastive representation learning method, subsequently combined with two representative monomodal registration methods, to judge the effectiveness of modality translation for multimodal image registration. We evaluate these method combinations on four publicly available multimodal (2D and 3D) datasets and compare with the performance of registration achieved by several well-known approaches acting directly on multimodal image data. Our results suggest that, although I2I translation may be helpful when the modalities to register are clearly correlated, registration of modalities which express distinctly different properties of the sample are not well handled by the I2I translation approach. The evaluated representation learning method, which aims to find abstract image-like representations of the information shared between the modalities, manages better, and so does the Mutual Information maximisation approach, acting directly on the original multimodal images. We share our complete experimental setup as open-source (https://github.com/MIDA-group/MultiRegEval), including method implementations, evaluation code, and all datasets, for further reproducing and benchmarking.

Place, publisher, year, edition, pages
Public Library of Science (PLoS), 2022
National Category
Medical Imaging Computer graphics and computer vision
Identifiers
urn:nbn:se:uu:diva-490949 (URN)10.1371/journal.pone.0276196 (DOI)000925006300013 ()36441754 (PubMedID)
Funder
Vinnova, 2017-02447Vinnova, 2020-03611Swedish Research Council, 2017-04385
Available from: 2022-12-16 Created: 2022-12-16 Last updated: 2025-02-09Bibliographically approved
Öfverstedt, J. (2022). Methods for Reliable Image Registration: Algorithms, Distance Measures, and Representations. (Doctoral dissertation). Uppsala: Acta Universitatis Upsaliensis
Open this publication in new window or tab >>Methods for Reliable Image Registration: Algorithms, Distance Measures, and Representations
2022 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

Much biomedical and medical research relies on the collection of ever-larger amounts of image data (both 2D images and 3D volumes, as well as time-series) and increasingly from multiple sources. Image registration, the process of finding correspondences between images based on the affinity of features of interest, is often required as a vital step towards the final analysis, which may consist of a comparison of images, measurement of movement, or fusion of complementary information. The contributions in this work are centered around reliable image registration methods for both 2D and 3D images with the aim of wide applicability: similarity and distance measures between images for image registration, algorithms for efficient computation of these, and other commonly used measures for both local and global optimization frameworks, and representations for multimodal image registration where the appearance and structures present in the images may vary dramatically.

The main contributions are: (i) distance measures for affine symmetric intensity image registration, combining intensity and spatial information based on the notion of alpha-cuts from fuzzy set theory; (ii) the extension of the affine registration method to more flexible deformable transformation models, leading to the framework Intensity and Spatial Information-Based Deformable Image Registration (INSPIRE); (iii) two efficient algorithms for computing the proposed distances and their spatial gradients and thereby enabling local gradient-based optimization; (iv) a contrastive representation learning method, Contrastive Multimodal Image Representation for Registration (CoMIR), utilizing deep learning techniques to obtain common representations that can be registered using methods designed for monomodal scenarios; (v) efficient algorithms for global optimization of mutual information and similarities of normalized gradient fields; (vi) a comparative study exploring the applicability of modern image-to-image translation methods to facilitate multimodal registration; (vii) the Stochastic Distance Transform, using the theory of discrete random sets to offer improved noise-insensitivity to distance computations; (viii) extensive evaluation of the proposed image registration methods on a number of different datasets mainly from (bio)medical imaging, where they exhibit excellent performance, and reliability, suggesting wide utility.

Place, publisher, year, edition, pages
Uppsala: Acta Universitatis Upsaliensis, 2022. p. 110
Series
Digital Comprehensive Summaries of Uppsala Dissertations from the Faculty of Science and Technology, ISSN 1651-6214 ; 2107
Keywords
Image registration, alignment, local optimization, global optimization, mutual information, normalized gradient fields, representation learning
National Category
Computer graphics and computer vision Medical Imaging
Research subject
Computerized Image Processing
Identifiers
urn:nbn:se:uu:diva-463393 (URN)978-91-513-1382-5 (ISBN)
Public defence
2022-02-25, 101195, Ångström, Lägerhyddsvägen 1, Uppsala, 13:00 (English)
Opponent
Supervisors
Available from: 2022-02-03 Created: 2022-01-10 Last updated: 2025-02-09
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0003-0253-9037

Search in DiVA

Show all publications