Page 56 - ARNM-1-2
P. 56

Advances in Radiotherapy
            & Nuclear Medicine                                                    Image fusion’s significance in medicine




            Table 2. Deep learning‑based image fusion methods in NMMI
             Authors             Year   Fusion     Multimodal images          Fusion Techniques         Organ
                                        methods
            Liu et al. [48]      2017  CNN        SPECT-CT, SPECT-MRI, CT-MRI  Siamese convolutional network  Brain
            Xu and Ma [49]       2021  CNN        SPECT-MRI, CT-MRI, PET-MRI  An unsupervised enhanced medical   Brain
                                                                             image fusion network
            Lahoudand Süsstrunk [28]  2019  CNN   SPECT-MRI, CT-MRI, PET-MRI  A novel strategy based on deep   Brain
                                                                             feature maps extracted from a CNN
            Wang et al. [30]     2020  CNN        SPECT-MRI                  CNN and contrast pyramid  Brain
            Liu et al. [32]      2015  CSR        PET-MRI                    Zero-masking strategy for data   Brain
                                                                             fusion
            Xia et al. [50]      2020  CSR        SPECT-MRI, CT-MRI, PET-MRI  Parameter-adaptive pulse-coupled   Brain
                                                                             neural network (PAPCNN)
            Huang et al. [35]    2020  GAN        PET-MRI, SPECT- MRI, SPECT-CT  Multi-generator multi-discriminator   Brain
                                                                             conditional GAN (MGMDcGAN)
            Ma et al. [51]       2020  GAN        PET-MRI                    Dual-discriminator conditional   Brain
                                                                             GAN (DDcGAN)
            Kang et al.  [52]    2020  GAN        PET-MRI                    Tissue-aware conditional GAN   Brain
                                                                             (TA-cGAN)
            Suk et al. [40]      2014  RBM        PET-MRI                    MultiModal DBM            Brain
            Abbreviations: CNN: Convolutional neural network; CSR: Convolutional sparse representation; CT: Computed tomography; DBM: Deep Boltzmann
            machine; GAN: Generative adversarial networks; MRI: Magnetic resonance imaging; PET: Positron emission tomography; RBM: Restricted Boltzmann
            machines; SPECT: Single-photon emission computed tomography.
            to previous workflows, this method efficiently fuses   different resolutions, termed multi-generator multi-
            multimodal neuroimaging features in a single setting and   discriminator conditional GAN (MGMDcGAN). This
            has the potential to require less labeled data. A research   method enables the simultaneous preservation of functional
            study  proposed a CSR-based image fusion framework,   and structural information, including texture details and
                [31]
            where each source image is decomposed into a base layer   dense structure information, without introducing spectral
            and a detail layer, facilitating multi-focus image fusion and   distortion or information loss.
            multimodal image fusion.
                                                               3.4. RBM-based image fusion
            3.3. GAN-based image fusion                        RBM-based  image  fusion  technique  is an unsupervised
            GAN-based image fusion technology consists of two neural   learning-based method used for image fusion. It employs
            networks: A generator and a discriminator. The generator   two RBMs, with one RBM dedicated to extracting
            takes two input images and produces a fused image as   features from the first image and the other from the
            an output, while the discriminator’s role is to distinguish   second image. These two RBMs are interconnected to
            between the generated fused image and real images .   form a bidirectional image fusion model. The model
                                                        [33]
            Throughout the training process, the generator continually   leverages the energy function of RBMs to minimize the
            attempts to generate more realistic fused images to deceive   error between the fused image and the original images.
            the discriminator, while the discriminator attempts to   When presented with a new input image, the model
            identify the differences between real and generated images.   simultaneously extracts features from both RBMs and
            Through repeated iterations, the generator gradually   fuses them using the energy function [36,37] . The RBM
            learns how to generate more realistic fused images .   image fusion technique demonstrates good adaptability
                                                        [34]
            GAN-based image fusion technology excels ingenerating   and generalization ability, allowing it to effectively fuse
            more realistic fused images and demonstrates the ability   images of different types and resolutions. Moreover, it
            to adaptively fuse images of varying types and resolutions   preserves the image details, improves image quality, and
            while preserving the details and features of the original   reduces noise and artifacts [38,39] .
            images.                                              Suk  et al.  used deep Boltzmann machine (DBM),
                                                                          [40]
              Huang  et al. [35]  proposed a new deep learning-based   a deep network with a restricted Boltzmann machine
            fusion method for multi-modal medical images with   as a building block, to find a latent hierarchical feature



            Volume 1 Issue 2 (2023)                         6                       https://doi.org/10.36922/arnm.0870
   51   52   53   54   55   56   57   58   59   60   61