Page 122 - AN-4-2
P. 122

Advanced Neurology                                                          ML for EEG signal recognition



            metrics  underscore  LightGBM’s  suitability  for  accurate   and LGBM classifier models used for the classification of
            and reliable classification tasks. Because of its superior   epileptic and non-epileptic EEG signals, highlighting their
            performance, LightGMB was selected for further analysis   learning  behaviors  and  generalization  capabilities  across
            to be compared to the CNN and DNN.                 epochs and training instances.
            3.1. Learning curves                                 The performance of the three models – DNN, CNN,
                                                               and LGBM classifier – offers important insights into
            Learning curves are a fundamental tool used in machine   their applicability for distinguishing epileptic from non-
            learning (ML) to assess the performance of a model during   epileptic  EEG signals.  Both  the  DNN and  CNN  models
            training  and validation. They provide  insights into  how   exhibit  the expected trend where training accuracy
            well a model generalizes to unseen data and whether it is   steadily increases and plateaus around 98%. Similarly,
            overfitting or underfitting. A typical learning curve plots   their validation accuracies stabilize at approximately 82%
            training and validation accuracy (or loss) against epochs   and 80%, respectively, maintaining a reasonable gap from
            or the number of training instances. Ideally, the training   the training curves. This behavior indicates that both
            and validation accuracy should increase steadily and   models are learning meaningful patterns from the EEG
            plateau, with the validation accuracy consistently lower   data without severe overfitting, making them suitable for
            than the training accuracy, reflecting the model’s ability to   clinical applications where generalization to unseen data
            generalize while avoiding overfitting.
                                                               is critical. The slight fluctuation in validation accuracy,
              In the context of this study, learning curves are   particularly in the CNN model, may suggest sensitivity
            particularly relevant for evaluating the performance of   to hyperparameter selection or inherent variability in
            ML models in distinguishing epileptic from non-epileptic   the EEG dataset but remains within acceptable limits for
            EEG signals – a challenging clinical task requiring high   clinical decision-making.
            diagnostic accuracy and generalization. Monitoring these   The LGBM classifier, however, demonstrates a perfect
            curves ensures that the models are learning robust features
            from the EEG data rather than memorizing patterns   training score of 1.0 across all training instances, which
            specific to the training set, which is critical for reliable   is a potential sign of overfitting to the training data.
            deployment in clinical settings.  Figure  1 illustrates the   The cross-validation score stabilizes around 85%, with
            training and validation performance of the DNN, CNN,   minimal improvement as the number of training instances
                                                               increases. While the LGBM classifier achieves a slightly
                                                               higher validation score, its inability to show a steady rise
                                                               and plateau in both training and validation curves raises
                                                               concerns about its ability to generalize effectively to new
                                                               EEG data. This behavior may stem from insufficient
                                                               regularization or the model’s tendency to exploit spurious
                                                               correlations in the data.
                                                                 In the context of EEG-based classification for epilepsy
                                                               diagnosis, models such as DNN and CNN, which show a more
                                                               balanced trade-off between training and validation accuracy,
                                                               may offer greater clinical reliability. Their performance
                                                               aligns with the expectation that validation accuracy should
                                                               trail slightly behind training accuracy, reflecting a model’s
                                                               ability to generalize while avoiding overfitting. Conversely,
            Figure 1. The graphs compare the performance of three models: DNN,
            CNN, and LGBM classifier. Both the DNN and CNN models exhibit a   the LGBM classifier’s results highlight the importance of
            steady increase in training accuracy, eventually plateauing around 98%,   rigorous evaluation and regularization in high-stakes clinical
            while their validation accuracies stabilize at approximately 80% and 82%,   applications, as overly optimistic training performance does
            respectively. This is a typical and expected behavior, where the training   not  necessarily  translate to robust  diagnostic  capabilities.
            accuracy is slightly higher than the validation accuracy, reflecting proper   These findings underscore the need for careful model
            learning  without  severe  overfitting. In  contrast,  the  LGBM classifier
            shows a perfect training score of 1.0 across all training instances, which   selection and tuning in machine learning-based approaches
            is not ideal, as it suggests overfitting on the training data. The cross-  to EEG signal classification.
            validation score for the LGBM classifier stabilizes at around 85% but lacks
            the expected steady increase and plateau shape, which would indicate   3.2. Confusion matrices
            better generalization.
            Abbreviations: CNN: Convolutional neural network; DNN: Dense Neural   A confusion matrix is a widely used ML-based tool for
            Network; LightGBM: Light gradient boosting machine.  evaluating  the  performance  of classification models. It


            Volume 4 Issue 2 (2025)                        116                               doi: 10.36922/an.7941
   117   118   119   120   121   122   123   124   125   126   127