Page 44 - IJAMD-1-3
P. 44

International Journal of AI for
            Materials and Design
                                                                                  Metal AM porosity prediction using ML



                         A                                       B




















            Figure 3. Illustration of the data from a layer of an additive manufacturing-built NiTi block with 60-micron thickness. (A) Pyrometer data and (B) slice of
            µ-computed tomography scan representing the layer.

                         A                                               B




















                            Figure 4. Distribution of layer porosity through the dataset: (A) Histogram and (B) violin and boxplot.
              First, SMOTE randomly selects one sample (s ) from the   to Chawla et al.,  combining the random undersampling
                                                                            35
                                                  r
            minority class in dataset S (NiTi×3B in our case) and then   of  the  majority  class  and  SMOTE  (to  oversample  the
            finds the k nearest neighbors (NNs) of the selected sample s r   minority class) offers better performance than the simple
            in the feature space to create a set N. Next, it randomly selects   undersampling approach.
            a neighbor n (where n ∈ N) and connects s and n to form   SMOTE uses the NN strategy to synthesize new samples
                                              r
                      r
                              r
                                                   r
            a line segment (i.e. _ _ s rn r ) in the feature space. A new   and maintain closer proximity to the samples belonging to
            synthetic sample is created by selecting a random point (i.e.,   the same class in the feature space (i.e., samples sharing a class
            a vector in the feature space) along this line segment. The   are closer to one another than samples of a different class).
            process continues until the dataset is reasonably balanced   Like most NN algorithms, which are susceptible to noise,
            and the resultant modified dataset is D (Algorithm 1).  SMOTE also suffers from creating many noisy data points
              SMOTE’s  reasonable  strategy  for  synthetic  data   in the feature space. Regarding handling such a scenario,
                                                                   36
            generation, which directly involves minority samples, makes   ENN  is a popular algorithm that finds and removes
            the newly constructed data relatively close to the minority   ambiguous and noisy samples from the dataset.
            samples in the feature space, making the approach effective.  When used with SMOTE, ENN is an undersampling
                                                               method for extensive data cleaning over the oversampled
            2.3.1. SMOTE-edited NN (ENN)                       dataset to achieve much cleaner clustered samples. For each
            One drawback of SMOTE is that the algorithm overlooks   sample d in the SMOTE oversampled dataset D, ENN finds
                                                                      i
            the majority class during data augmentation. According   d’s NNs set N of size k. If the d’s class y and the majority
                                                                                        i
                                                                                               i
                                                                i
            Volume 1 Issue 3 (2024)                         38                             doi: 10.36922/ijamd.4812
   39   40   41   42   43   44   45   46   47   48   49