Page 50 - IJAMD-1-3
P. 50

International Journal of AI for
            Materials and Design
                                                                                  Metal AM porosity prediction using ML


            mechanism allows it to handle complex patterns in the   Repeated Stratified K‑Fold with the number of splits and
            data efficiently, while also offering adaptability through   repeats was employed. The system is only exposed to the
            hyperparameter tuning, making it a popular choice in ML   training  data  for hyperparameter  optimization.  Figure  9
            literature for large-scale classification tasks. Conversely,   illustrates the model accuracies across the training
            k-NN, though less efficient with large datasets due to   instances (i.e., iterations). Here, it can be observed that the
            the need to compute distances between all points, was   best parameters render a mean accuracy of the k-NN model
            used for its simplicity and interpretability. It provides a   around 69% on training data. The k‑NN model was trained
            straightforward approach to classification by relying on   on training data supplied with the best parameters obtained
            proximity in feature space. By employing both models,   in the previous step and evaluated it on the held-out test
            we ensure that XGBoost handles complex, non-linear   data. As a result, the model predicts the classes of the held-
            classifications efficiently, while k-NN provides a valuable   out samples with an accuracy of 65%. Undergoing a cross‑
            comparison through its simpler, distance-based approach.  validated hyper-parameter tuning process for XGBoost on
              For obtaining the best set of parameters of the k-NN   training data, the best set of parameters obtained are as
            and  XGBoost  model,  Grid  SearchCV  from  the  sklearn3   follows: learning rate = 0.01, estimators = 400, gamma = 5,
            library was employed. For the cross-validation mechanism,   max depth = 5, min child weight = 10, subsample = 1.0,
                                                               and colsample-bytree = 0.8. The XGBoost model attained
            Table 2. Summary of the advantages and disadvantages of   an accuracy, recall, precision, and F1 score of ~100% after
            models used for the predictive quality assessment of porosity   10-fold cross-validation as demonstrated by the confusion
            within metal additive manufacturing                matrix and classification report in Figure 11. The XGBoost
                                                               model proved to be a significantly better classifier of
            Model     Training   Real‑time         RMSE        porosity (high vs. low) compared to the k-NN model
                      speed     classification speed  (low porosity)
                                                               and is, therefore, the recommended choice for analyzing
            LR        Very Fast  Very Fast          0.05       pyrometer data.
            k-NN      Slow      Slow                0.06         In this study, pyrometer data was found to be valuable
            SVR       Medium    Medium              0.09       for predicting porosity percentage during AM, addressing
            DT        Fast      Very Fast           0.06       a major issue in printed parts. Porosity can lead to reduced
            RF        Slow      Fast                0.05       density and poor mechanical properties, but with the
            XT        Medium    Fast                0.04       assistance of ML models, porosity can be detected in
            GB        Slow      Medium              0.04       real-time, allowing for process parameters to be adjusted
            AdaBoost  Slow      Medium              0.05       accordingly.  The  two  in situ  pyrometers,  provided  by
            XGBoost   Fast      Fast                0.05       KLEIBER  Infrared  GmbH,  measured  reflected  light
                                                               from the melt-pool area, detecting heat emissions in the
            Abbreviations: AdaBoost: Adaptive Boosting; DT: Decision Tree;   1500 – 1700  nm range, which represent the melt pool
            GB: Gradient Boosting; k-NN: k-Nearest Neighbors; LR: Linear
            Regression; RF: Random Forest; SVR: Support Vector Regression;   temperature. The data captured was mapped in x and y
            XGBoost: eXtreme Gradient Boosting; XT: Extremely Randomized Trees.  coordinates for each layer, with a resolution calibrated

                         A                               B




















            Figure 11. Performance of eXtreme Gradient Boosting Classifier on validation data: (A) Confusion matrix and (B) classification report.


            Volume 1 Issue 3 (2024)                         44                             doi: 10.36922/ijamd.4812
   45   46   47   48   49   50   51   52   53   54   55