Page 102 - AN-4-4
P. 102

Advanced Neurology                                                Diffusion model for brain tumor classification



                                                               iterative refinement ensures training stability, avoids mode
                                                               collapse, and produces highly detailed synthetic images
                                                               that preserve tumor morphology and anatomical features.
                                                               Moreover, DDM does not require adversarial competition
                                                               between generator and discriminator networks, which is
                                                               the primary source of instability in GAN training. Instead,
                                                               it uses a maximum likelihood-based training objective,
                                                               resulting in a more predictable and stable convergence.
                                                               This stability ensures that the generated synthetic MRI
                                                               slices  maintain  consistent quality  across large batches,
                                                               enabling the downstream CDCNN classifier to learn from
                                                               a broader and more representative feature space.
                                                               4.1. Rationale for combining the DDM with a CDCNN
                                                               The CDCNN architecture is inherently powerful for medical
                                                               imaging tasks because it integrates deep convolutional
            Figure 8. Hexagonal radar chart of model performance illustrating the   feature extraction with conditional processing, allowing
            balanced performance of the conditional deep convolutional neural   it to focus on class-specific discriminative features. When
            network model across accuracy, sensitivity, specificity, F1 score, positive   trained on DDM-generated images alongside real MRI
            predictive value (PPV), and negative predictive value (NPV). Larger   data, the CDCNN benefits from:
            enclosed areas indicate superior performance.
                                                               (i)  Increased dataset size: DDM produced 10,000
                                                                  synthetic MRI images across four tumor classes,
            Table 4. Summary of different models’ results         effectively expanding the dataset and improving the

            Model      Accuracy   Sensitivity   Specificity   F1 Score   network’s generalization ability.
                         (%)      (%)       (%)      (%)       (ii)  Improved feature diversity: DDM-generated images
            CDCNN        96.2     95.8      96.7     96.0         introduce subtle variations in tumor shapes, textures,
            ResNet50     91.5     90.8      92.0     91.2         and intensities, enabling CDCNN to learn richer
                                                                  decision boundaries.
            VGG16        89.7     88.9      90.5     89.3      (iii) Balanced class representation: Synthetic augmentation
            VGG19        90.3     89.7      91.0     90.0         corrected class imbalances in the original dataset,
            InceptionV3  92.8     91.9      93.5     92.2         reducing bias and improving per-class recall and
            Abbreviations: CDCNN: Conditional deep convolutional neural   precision.
            network; VGG: Visual Geometry Group.
                                                                 This synergy between high-fidelity synthetic data and a
                                                               robust DL architecture resulted in balanced performance
            Onakpojeruo et al.,  where the same CDCNN model was   across all tumor classes, with particularly strong results
                           52
            used with Pix2Pix for synthetic data generation, achieving   for the challenging glioma class, which historically suffers
            an  accuracy  of  only  86%.  The  10%  increase  in  accuracy   from high misclassification rates.
            strongly highlights the effectiveness of DDM in generating
            high-quality, realistic, and diverse synthetic datasets that   The DDM-CDCNN framework demonstrated high
            significantly boost classification performance.    sensitivity and specificity across all four tumor categories.
                                                               For example, the no tumor class achieved a specificity of
              Traditional GAN-based methods, while powerful, are   97%, indicating the model’s ability to correctly identify
            known to suffer from issues, such as mode collapse, unstable   negative cases and avoid false positives. The balanced F1
            training dynamics, and sensitivity to hyperparameter   scores  further  underscore  the  robustness  of  the  model,
            tuning. These limitations often result in synthetic images   ensuring reliable classification without sacrificing either
            that either lack diversity or fail to capture fine structural   precision or recall. Importantly, the ablation study revealed
            details crucial for accurate medical image classification. In   that removing the DDM-generated synthetic data reduced
            contrast, DDM operates using a progressive noise addition   accuracy by 8.5%, confirming its critical role in enhancing
            and removal process. During the forward diffusion process,   classification performance. Statistical tests further
            Gaussian noise is added to the image over several timesteps,   validated these results: Paired t-tests and ANOVA showed
            gradually destroying image structure. The reverse diffusion   significant differences (p<0.05) between the proposed
            process  then  reconstructs  the  image  by  denoising  step-  method and baseline models, while confidence intervals
            by-step, conditioned on a learned distribution. This   demonstrated the stability of the performance metrics.


            Volume 4 Issue 4 (2025)                         96                           doi: 10.36922/AN025130025
   97   98   99   100   101   102   103   104   105   106   107