Page 100 - AN-4-4
P. 100

Advanced Neurology                                                Diffusion model for brain tumor classification



            an  ×64-based  system  equipped  with  an  11   Gen  Intel    generated synthetic data, learning curves were plotted for
                                                         
                                                th
            Core™ i7-11700KF CPU at 3.60 GHz. Experiments were   both accuracy and loss across 50 training epochs. These
            conducted  in  Python  3.10  using  PyTorch  2.0.1,  CUDA   curves offer critical insight into the convergence behavior
            11.7, and torchvision 0.15.2. All training procedures used a   and the stability of the model. As illustrated in Figure 5,
            fixed random seed of 42 for reproducibility. Finally, model   the training and validation accuracy curves show a steady
            training and evaluation were carried out on the synthetic   and consistent upward trajectory, ultimately converging at
            dataset generated using the Keras package and Python.   approximately 96% accuracy. The minimal gap between
            Here, the Jupyter Notebook environment was used as the   the training and validation curves throughout the training
            programming language.                              process suggests a low variance, indicating that the model
                                                               is learning meaningful patterns without overfitting to the
            3. Results                                         training data. Similarly, Figure 6 presents the training and
                                                               validation loss curves, both of which decrease smoothly over
            3.1. Model performance
                                                               epochs. The loss curves exhibit a converging trend, where
            The CDCNN model achieved an accuracy of 96.2%      the  training  loss  steadily  decreases  while  the  validation
            when trained with synthetic datasets generated using   loss remains closely aligned. This behavior further affirms
            DDM. Table 2 illustrates the performance metrics of the   the absence of overfitting and supports the claim that
            generated datasets. This marks a significant improvement
            over traditional GAN-based augmentation techniques,
            demonstrating the efficacy of the diffusion-based approach.

            3.2. Confusion matrix
            Figure 4 reports a row-normalized confusion matrix for
            the DDM-generated synthetic test set. Each cell shows the
            percentage of samples from a true class (row) predicted
            as a given class (column). The diagonal entries, therefore,
            represent per-class true-positive rates: Glioma 89%,
            meningioma 92%, pituitary tumor 95%, and no tumor
            97%. Given the 80/20 split of 2,500 synthetic images
            per class (i.e., 500 test images/class), these correspond
            to 445/500, 460/500, 475/500, and 485/500 correct
            classifications,  respectively.  Off-diagonal  percentages
            reflect the remaining misclassifications and are small in
            magnitude, indicating limited confusion across classes.
            The high diagonal values—particularly 97% for the no
            tumors class—confirm strong specificity; similarly, the   Figure  4.  Confusion matrix showcasing classification performance
                                                               across four tumor classes. Row-normalized confusion matrix (%) for the
            95% recall for the pituitary tumor class indicates robust   denoising diffusion model-generated synthetic test set. Each cell reports
            sensitivity. Overall, the confusion matrix corroborates the   the percentage of samples from the true class (row) predicted as the target
            model’s balanced performance and supports the reliability   class (column). The diagonal entries summarize per-class recall (TPR):
            of the CDCNN trained with DDM-augmented data.      Glioma 89% (≈445/500), Meningioma 92% (≈460/500), Pituitary tumor
                                                               95% (≈475/500), and No tumor 97% (≈485/500). Percentages may not
            3.3. Model convergence and learning curves         sum to exactly 100%/row due to rounding.
            To assess the training dynamics and generalization ability
            of the CDCNN model that was trained with DDM-


            Table 2. Performance metrics on synthetic datasets

            Class data  Precision   Recall   F1 score   Accuracy
                          (%)      (%)     (%)      (%)
            Glioma         89      91      90       96.2
            Meningioma     92      89      91
            Pituitary tumor  88    95      91
                                                               Figure 5. Training and validation accuracy across 50 epochs, showing
            No tumor       98      97      97
                                                               smooth convergence and consistent generalization

            Volume 4 Issue 4 (2025)                         94                           doi: 10.36922/AN025130025
   95   96   97   98   99   100   101   102   103   104   105