Page 55 - ARNM-1-2
P. 55

Advances in Radiotherapy
            & Nuclear Medicine                                                    Image fusion’s significance in medicine



            structure  in  the  cartoon  components  and  a  sparse-  fusion needs. For example, one can utilize classic CNN
            representation-based method for the texture components.   structures such as VGG and ResNet, or explore custom
            Their approach outperformed the state-of-art methods in   network structures and convolutional kernel sizes to achieve
                                                                                        [26]
            both visual and quantitative evaluations. Daneshvar and   more precise feature extraction . In addition, multi-scale
            Ghassemian  combined the IHS method with the retina-  feature fusion can be used to fuse feature maps of different
                      [22]
            inspired model (RIM) fusion technique, which enabled   scales, thereby enhancing the accuracy and robustness of
            the preservation of more spatial features and functional   the fusion image. Alternatively, pixel-level fusion methods
            information content, respectively. By integrating the   can be used to fuse information for each pixel to generate
            advantages of both IHS and RIM fusion methods, their   a more refined fusion image. Overall, CNN-based image
            algorithm successfully improved both functional and   fusion presents a highly promising image fusion method
            spatial information content.                       for achieving more accurate and efficient medical image
                                                               fusion. Continuous exploration of new network structures,
            3. Deep learning-based image fusion                hyperparameters,  and fusion methods  is essential  for
            methods                                            further improving its performance .
                                                                                          [27]
            The emergence of deep learning has further advanced the   Lahoud and Süsstrunk  proposed a real-time image
                                                                                    [28]
            development of image fusion technology. Deep learning-  fusion method that utilizes pre-trained neural networks
            based image fusion method utilizes deep neural network   to generate a single image containing features from multi-
            models  to  learn  image  features  and  weights,  achieving   modal sources. In contrast, Teng et al.  presented a fusion
                                                                                             [29]
            automatic  image fusion. First,  multiple  modal  medical   algorithm based on neuro-fuzzy logic and utilized a hybrid
            images  are  input into  the  deep  neural network, which   algorithm that combines the back propagation algorithm
            captures  key  information  by  extracting and  learning   with the least mean square (LMS) algorithm to train the
            features through multiple layers. Subsequently, the   parameters of the membership function. The fused images
            extracted features are fused using a fusion layer to generate   based on neuro-fuzzy logic not only reserve more texture
            the fused image. Finally, through training and optimizing   features but also enhance the information characteristics
            the network weights, the fused image achieves the highest   of the two original images. Wang  et al.  proposed an
                                                                                                [30]
            quality and information preservation.              algorithm that utilizes a trained Siamese CNN to fuse the
              Deep  learning  offers  new  feature  representation   pixel activity information of source images, enabling the
            methods,  addressing  the  limitations  of multiscale  and   generation of a weight map. In addition, they implemented
            spatial variability present in traditional methods. The   a contrast pyramid to decompose the source image.
            main deep learning-based fusion methods include    3.2. CSR-based image fusion
            convolutional neural networks (CNN), convolutional
            sparse  representation (CSR),  generative adversarial   The core idea of CSR-based image fusion technology is to
            networks (GAN), and deep restricted Boltzmann machines   leverage the sparsity of CNN to decompose image features
            (RBM) . In Section 3, we focus on the application of   into a set of sparse coefficients and then fuse them by
                 [23]
            deep learning-based image fusion methods in the field of   weighted summation.  First, CNN  is  used  to extract  the
            NMMI.  Table 2 provides an overview of the works that   features of two images to be fused, and subsequently, the
            report on the deep learning-based image fusion methods   extracted features are decomposed to obtain the sparse
            in NMMI.                                           coefficients for each feature. For each pixel, the sparse
                                                               coefficients of the two images are weighted and summed
            3.1. CNN-based image fusion                        based on their feature weights in the two images, resulting
            CNN-based image fusion is a common method for      in new sparse coefficients. Finally, these new sparse
            medical image fusion. CNNs learn features from raw data,   coefficients are restored into an image to obtain the fused
            and thus learn different types of image features, ultimately   image. CSR-based image fusion technology effectively
            achieving image fusion . In this approach, different types   retains the main features of both images during the fusion
                              [24]
            of medical images are first extracted for features through   process while demonstrating robust detail preservation
                                                                                    [31]
            CNNs, resulting in feature representations of different   ability and noise resistance .
            types of images. Then, these feature representations are   Liu  et  al.  designed a novel diagnostic framework
                                                                         [32]
            fused to generate the final fusion image . The flexibility   with deep learning architecture to aid the diagnosis of
                                            [25]
            of CNNs allows for the use of different network structures   Alzheimer’s disease (AD). This framework employs a zero-
            and hyperparameters during the feature extraction process,   masking strategy for data fusion, extracting complementary
            enabling adaptation to various types of medical images and   information from multiple data modalities. Compared


            Volume 1 Issue 2 (2023)                         5                       https://doi.org/10.36922/arnm.0870
   50   51   52   53   54   55   56   57   58   59   60