Logo: to the web site of Uppsala University

uu.sePublications from Uppsala University
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Methods for Reliable Image Registration: Algorithms, Distance Measures, and Representations
Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Visual Information and Interaction. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Computerized Image Analysis and Human-Computer Interaction.ORCID iD: 0000-0003-0253-9037
2022 (English)Doctoral thesis, comprehensive summary (Other academic)
Description
Abstract [en]

Much biomedical and medical research relies on the collection of ever-larger amounts of image data (both 2D images and 3D volumes, as well as time-series) and increasingly from multiple sources. Image registration, the process of finding correspondences between images based on the affinity of features of interest, is often required as a vital step towards the final analysis, which may consist of a comparison of images, measurement of movement, or fusion of complementary information. The contributions in this work are centered around reliable image registration methods for both 2D and 3D images with the aim of wide applicability: similarity and distance measures between images for image registration, algorithms for efficient computation of these, and other commonly used measures for both local and global optimization frameworks, and representations for multimodal image registration where the appearance and structures present in the images may vary dramatically.

The main contributions are: (i) distance measures for affine symmetric intensity image registration, combining intensity and spatial information based on the notion of alpha-cuts from fuzzy set theory; (ii) the extension of the affine registration method to more flexible deformable transformation models, leading to the framework Intensity and Spatial Information-Based Deformable Image Registration (INSPIRE); (iii) two efficient algorithms for computing the proposed distances and their spatial gradients and thereby enabling local gradient-based optimization; (iv) a contrastive representation learning method, Contrastive Multimodal Image Representation for Registration (CoMIR), utilizing deep learning techniques to obtain common representations that can be registered using methods designed for monomodal scenarios; (v) efficient algorithms for global optimization of mutual information and similarities of normalized gradient fields; (vi) a comparative study exploring the applicability of modern image-to-image translation methods to facilitate multimodal registration; (vii) the Stochastic Distance Transform, using the theory of discrete random sets to offer improved noise-insensitivity to distance computations; (viii) extensive evaluation of the proposed image registration methods on a number of different datasets mainly from (bio)medical imaging, where they exhibit excellent performance, and reliability, suggesting wide utility.

Place, publisher, year, edition, pages
Uppsala: Acta Universitatis Upsaliensis, 2022. , p. 110
Series
Digital Comprehensive Summaries of Uppsala Dissertations from the Faculty of Science and Technology, ISSN 1651-6214 ; 2107
Keywords [en]
Image registration, alignment, local optimization, global optimization, mutual information, normalized gradient fields, representation learning
National Category
Computer graphics and computer vision Medical Imaging
Research subject
Computerized Image Processing
Identifiers
URN: urn:nbn:se:uu:diva-463393ISBN: 978-91-513-1382-5 (print)OAI: oai:DiVA.org:uu-463393DiVA, id: diva2:1625765
Public defence
2022-02-25, 101195, Ångström, Lägerhyddsvägen 1, Uppsala, 13:00 (English)
Opponent
Supervisors
Available from: 2022-02-03 Created: 2022-01-10 Last updated: 2025-02-09
List of papers
1. Fast and Robust Symmetric Image Registration Based on Distances Combining Intensity and Spatial Information
Open this publication in new window or tab >>Fast and Robust Symmetric Image Registration Based on Distances Combining Intensity and Spatial Information
2019 (English)In: IEEE Transactions on Image Processing, ISSN 1057-7149, E-ISSN 1941-0042, Vol. 28, no 7, p. 3584-3597Article in journal (Refereed) Published
Abstract [en]

Intensity-based image registration approaches rely on similarity measures to guide the search for geometric correspondences with high affinity between images. The properties of the used measure are vital for the robustness and accuracy of the registration. In this study a symmetric, intensity interpolation-free, affine registration framework based on a combination of intensity and spatial information is proposed. The excellent performance of the framework is demonstrated on a combination of synthetic tests, recovering known transformations in the presence of noise, and real applications in biomedical and medical image registration, for both 2D and 3D images. The method exhibits greater robustness and higher accuracy than similarity measures in common use, when inserted into a standard gradientbased registration framework available as part of the open source Insight Segmentation and Registration Toolkit (ITK). The method is also empirically shown to have a low computational cost, making it practical for real applications. Source code is available.

Place, publisher, year, edition, pages
IEEE, 2019
Keywords
Image registration, set distance, gradient methods, optimization, cost function, iterative algorithms, fuzzy sets, magnetic resonance imaging, transmission electron microscopy
National Category
Medical Imaging
Research subject
Computerized Image Processing
Identifiers
urn:nbn:se:uu:diva-377450 (URN)10.1109/TIP.2019.2899947 (DOI)000471067800004 ()
Funder
Vinnova, 2016-02329Swedish Research Council, 2015-05878Swedish Research Council, 2017-04385Vinnova, 2017-02447
Available from: 2019-02-20 Created: 2019-02-20 Last updated: 2025-02-09Bibliographically approved
2. Stochastic Distance Transform: Theory, Algorithms and Applications
Open this publication in new window or tab >>Stochastic Distance Transform: Theory, Algorithms and Applications
2020 (English)In: Journal of Mathematical Imaging and Vision, ISSN 0924-9907, E-ISSN 1573-7683, Vol. 62, p. 751-769Article in journal (Refereed) Published
Abstract [en]

Distance transforms (DTs) are standard tools in image analysis, with applications in image registration and segmentation. The DT is based on extremal (minimal) distance values and is therefore highly sensitive to noise. We present a stochastic distance transform (SDT) based on discrete random sets, in which a model of element-wise probability is utilized and the SDT is computed as the first moment of the distance distribution to the random set. We present two methods for computing the SDT and analyze them w.r.t. accuracy and complexity. Further, we propose a method, utilizing kernel density estimation, for estimating probability functions and associated random sets to use with the SDT. We evaluate the accuracy of the SDT and the proposed framework on images of thin line structures and disks corrupted by salt and pepper noise and observe excellent performance. We also insert the SDT into a segmentation framework and apply it to overlapping objects, where it provides substantially improved performance over previous methods. Finally, we evaluate the SDT and observe very good performance, on simulated images from localization microscopy, a state-of-the-art super-resolution microscopy technique which yields highly spatially localized but noisy point-clouds.

Keywords
Distance transform, Discrete random set, Robust distance, Image segmentation, Deterministic algorithm, Monte-Carl
National Category
Discrete Mathematics Computer graphics and computer vision
Research subject
Computerized Image Processing
Identifiers
urn:nbn:se:uu:diva-413842 (URN)10.1007/s10851-020-00964-7 (DOI)000541189800001 ()
Funder
Vinnova, 2016-02329Vinnova, 2017-02447
Available from: 2020-06-20 Created: 2020-06-20 Last updated: 2025-02-01Bibliographically approved
3. INSPIRE: Intensity and Spatial Information-Based Deformable Image Registration
Open this publication in new window or tab >>INSPIRE: Intensity and Spatial Information-Based Deformable Image Registration
2023 (English)In: PLOS ONE, E-ISSN 1932-6203, Vol. 18, no 3, article id e0282432Article in journal (Refereed) Published
Abstract [en]

We present INSPIRE, a top-performing general-purpose method for deformable image registration. INSPIRE brings distance measures which combine intensity and spatial information into an elastic B-splines-based transformation model and incorporates an inverse inconsistency penalization supporting symmetric registration performance. We introduce several theoretical and algorithmic solutions which provide high computational efficiency and thereby applicability of the proposed framework in a wide range of real scenarios. We show that INSPIRE delivers highly accurate, as well as stable and robust registration results. We evaluate the method on a 2D dataset created from retinal images, characterized by presence of networks of thin structures. Here INSPIRE exhibits excellent performance, substantially outperforming the widely used reference methods. We also evaluate INSPIRE on the Fundus Image Registration Dataset (FIRE), which consists of 134 pairs of separately acquired retinal images. INSPIRE exhibits excellent performance on the FIRE dataset, substantially outperforming several domain-specific methods. We also evaluate the method on four benchmark datasets of 3D magnetic resonance images of brains, for a total of 2088 pairwise registrations. A comparison with 17 other state-of-the-art methods reveals that INSPIRE provides the best overall performance. Code is available at github.com/MIDA-group/inspire

Place, publisher, year, edition, pages
Public Library of Science (PLoS), 2023
Keywords
Image registration, splines (mathematics), deformable models, set distance, gradient methods, optimization, cost function, iterative methods, fuzzy sets
National Category
Computer graphics and computer vision Medical Imaging
Research subject
Computerized Image Processing
Identifiers
urn:nbn:se:uu:diva-429025 (URN)10.1371/journal.pone.0282432 (DOI)000945993100007 ()36867617 (PubMedID)
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)Vinnova, 2017-02447Vinnova, 2020-03611Vinnova, 2021-01420
Available from: 2020-12-18 Created: 2020-12-18 Last updated: 2025-02-09Bibliographically approved
4. Fast computation of mutual information in the frequency domain with applications to global multimodal image alignment
Open this publication in new window or tab >>Fast computation of mutual information in the frequency domain with applications to global multimodal image alignment
2022 (English)In: Pattern Recognition Letters, ISSN 0167-8655, E-ISSN 1872-7344, Vol. 159, p. 196-203Article in journal (Refereed) Published
Abstract [en]

Multimodal image alignment is the process of finding spatial correspondences between images formed by different imaging techniques or under different conditions, to facilitate heterogeneous data fusion and correlative analysis. The information-theoretic concept of mutual information (MI) is widely used as a similarity measure to guide multimodal alignment processes, where most works have focused on local maximization of MI, which typically works well only for small displacements. This points to a need for global maximization of MI, which has previously been computationally infeasible due to the high run-time complexity of existing algorithms. We propose an efficient algorithm for computing MI for all discrete displacements (formalized as the cross-mutual information function (CMIF)), which is based on cross-correlation computed in the frequency domain. We show that the algorithm is equivalent to a direct method while superior in terms of run-time. Furthermore, we propose a method for multimodal image alignment for transformation models with few degrees of freedom (e.g., rigid) based on the proposed CMIF-algorithm. We evaluate the efficacy of the proposed method on three distinct benchmark datasets, containing remote sensing images, cytological images, and histological images, and we observe excellent success-rates (in recovering known rigid transformations), overall outperforming alternative methods, including local optimization of MI, as well as several recent deep learning-based approaches. We also evaluate the run-times of a GPU implementation of the proposed algorithm and observe speed-ups from 100 to more than 10,000 times for realistic image sizes compared to a GPU implementation of a direct method. Code is shared as open-source at github.com/MIDA-group/globalign.

Place, publisher, year, edition, pages
ElsevierElsevier BV, 2022
Keywords
Mutual information, Image alignment, Global optimization, Multimodal, Entropy
National Category
Medical Imaging Computer graphics and computer vision
Research subject
Computerized Image Processing
Identifiers
urn:nbn:se:uu:diva-447807 (URN)10.1016/j.patrec.2022.05.022 (DOI)000833390400013 ()
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)Vinnova, 2017-02447Swedish Research Council, 2017-04385
Available from: 2021-06-30 Created: 2021-06-30 Last updated: 2025-02-09Bibliographically approved
5. Cross-Sim-NGF: FFT-Based Global Rigid Multimodal Alignment of Image Volumes using Normalized Gradient Fields
Open this publication in new window or tab >>Cross-Sim-NGF: FFT-Based Global Rigid Multimodal Alignment of Image Volumes using Normalized Gradient Fields
2021 (English)Manuscript (preprint) (Other academic)
Abstract [en]

Multimodal image alignment involves finding spatial correspondences between volumes varying in appearance and structure. Automated alignment methods are often based on local optimization that can be highly sensitive to their initialization. We propose a global optimization method for rigid multimodal 3D image alignment, based on a novel efficient algorithm for computing similarity of normalized gradient fields (NGF) in the frequency domain. We validate the method experimentally on a dataset comprised of 20 brain volumes acquired in four modalities (T1w, Flair, CT, [18F] FDG PET), synthetically displaced with known transformations. The proposed method exhibits excellent performance on all six possible modality combinations, and outperforms all four reference methods by a large margin. The method is fast; a 3.4Mvoxel global rigid alignment requires approximately 40 seconds of computation, and the proposed algorithm outperforms a direct algorithm for the same task by more than three orders of magnitude. Open-source implementation is provided.

National Category
Medical Imaging
Research subject
Computerized Image Processing
Identifiers
urn:nbn:se:uu:diva-456873 (URN)
Available from: 2021-10-22 Created: 2021-10-22 Last updated: 2025-02-09Bibliographically approved
6. CoMIR: Contrastive Multimodal Image Representation for Registration
Open this publication in new window or tab >>CoMIR: Contrastive Multimodal Image Representation for Registration
Show others...
2020 (English)In: NeurIPS - 34th Conference on Neural Information Processing Systems, 2020Conference paper, Published paper (Refereed)
Abstract [en]

We propose contrastive coding to learn shared, dense image representations, referred to as CoMIRs (Contrastive Multimodal Image Representations). CoMIRs enable the registration of multimodal images where existing registration methods often fail due to a lack of sufficiently similar image structures. CoMIRs reduce the multimodal registration problem to a monomodal one, in which general intensity-based, as well as feature-based, registration algorithms can be applied. The method involves training one neural network per modality on aligned images, using a contrastive loss based on noise-contrastive estimation (InfoNCE). Unlike other contrastive coding methods, used for, e.g., classification, our approach generates image-like representations that contain the information shared between modalities. We introduce a novel, hyperparameter-free modification to InfoNCE, to enforce rotational equivariance of the learnt representations, a property essential to the registration task. We assess the extent of achieved rotational equivariance and the stability of the representations with respect to weight initialization, training set, and hyperparameter settings, on a remote sensing dataset of RGB and near-infrared images. We evaluate the learnt representations through registration of a biomedical dataset of bright-field and second-harmonic generation microscopy images; two modalities with very little apparent correlation. The proposed approach based on CoMIRs significantly outperforms registration of representations created by GAN-based image-to-image translation, as well as a state-of-the-art, application-specific method which takes additional knowledge about the data into account. Code is available at: https://github.com/MIDA-group/CoMIR.

National Category
Medical Imaging Computer graphics and computer vision
Research subject
Computerized Image Processing
Identifiers
urn:nbn:se:uu:diva-430180 (URN)
Conference
NeurIPS - 34th Conference on Neural Information Processing Systems, December, 6-12, 2020,online, Canada
Available from: 2021-01-08 Created: 2021-01-08 Last updated: 2025-02-09Bibliographically approved
7. Is Image-to-Image Translation the Panacea for Multimodal Image Registration? A Comparative Study
Open this publication in new window or tab >>Is Image-to-Image Translation the Panacea for Multimodal Image Registration? A Comparative Study
2021 (English)Manuscript (preprint) (Other academic)
Abstract [en]

Despite current advancement in the field of biomedical image processing, propelled by the deep learning revolution, multimodal image registration, due to its several challenges, is still often performed manually by specialists. The recent success of image-to-image (I2I) translation in computer vision applications and its growing use in biomedical areas provide a tempting possibility of transforming the multimodal registration problem into a, potentially easier, monomodal one. We conduct an empirical study of the applicability of modern I2I translation methods for the task of multimodal biomedical image registration. We compare the performance of four Generative Adversarial Network (GAN)-based methods and one contrastive representation learning method, subsequently combined with two representative monomodal registration methods, to judge the effectiveness of modality translation for multimodal image registration. We evaluate these method combinations on three publicly available multimodal datasets of increasing difficulty, and compare with the performance of registration by Mutual Information maximisation and one modern data-specific multimodal registration method. Our results suggest that, although I2I translation may be helpful when the modalities to register are clearly correlated, registration of modalities which express distinctly different properties of the sample are not well handled by the I2I translation approach. When less information is shared between the modalities, the I2I translation methods struggle to provide good predictions, which impairs the registration performance. The evaluated representation learning method, which aims to find an in-between representation, manages better, and so does the Mutual Information maximisation approach. We share our complete experimental setup as open-source https://github.com/Noodles-321/Registration.

National Category
Medical Imaging Computer graphics and computer vision
Research subject
Computerized Image Processing
Identifiers
urn:nbn:se:uu:diva-456358 (URN)
Available from: 2021-10-18 Created: 2021-10-18 Last updated: 2025-02-09Bibliographically approved

Open Access in DiVA

UUThesis_J-Öfverstedt-2022(3775 kB)8829 downloads
File information
File name FULLTEXT01.pdfFile size 3775 kBChecksum SHA-512
97c82f3b1c60f0e2c67f245d27466071ba1868314b858209914111a4df7722381a4f9500c3d7e0edd79ee6cdd9f614c7ba79a46bba469446fbc57d3aaaa974cf
Type fulltextMimetype application/pdf

Authority records

Öfverstedt, Johan

Search in DiVA

By author/editor
Öfverstedt, Johan
By organisation
Division of Visual Information and InteractionComputerized Image Analysis and Human-Computer Interaction
Computer graphics and computer visionMedical Imaging

Search outside of DiVA

GoogleGoogle Scholar
Total: 8831 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

isbn
urn-nbn

Altmetric score

isbn
urn-nbn
Total: 4930 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf