Page 67 - IJAMD-2-1
P. 67

International Journal of AI for
            Materials and Design
                                                                             Fatigue life prediction via contrastive learning


            types of data in the training and testing sets, simulating the   To investigate whether the model successfully learned
            uneven distribution of data in real-world conditions. To   the distribution of data features during training, a visual
            further mitigate the impact of this distribution imbalance,   analysis of the input and output features before and after
            additional sample data generated by a GAN model was   training was performed. Specifically, in the visualization
            included in the training set to augment the samples and   process, original data was used as the visualization samples,
            improve the model’s generalization ability. Each sample   excluding the augmented samples generated by GAN. The
            underwent  two  different  data  augmentation  methods.   t-distributed stochastic neighbor embedding (t-SNE)
                                                                                                            68
            The training loss of each contrastive learning model   method was employed for dimensionality reduction
            framework is shown in  Figure  7.  After 150 epochs, all   and visualization of the features, with the pre- and post-
            models converged well, validating the effectiveness of the   training visualization results shown in Figures 8 and 9. In
            proposed framework.                                these figures, data points with the same color represented
                                                               features of samples with the same amplitude but different
            Table 2. The detailed hyperparameters of contrastive   loading paths, while different colors corresponded to
            learning models with different network architectures  samples with different amplitudes. The visualization
                                                               results indicated that in the original data, samples from the
            Encoder  Projector       Hyperparameters           same loading path were clustered together in the feature
                             Configuration  Stride Padding  τ  space. However, for features generated by each encoder
            1D-CNN   ANN     Convolutional   1    1    0.3     from  the  initial  samples,  the  distribution of  features  for
                             Layers: 2,                        the same amplitude data was quite disordered, and no
                             Kernel Size: (3,3),               obvious clustering or separation of samples was observed
                             Filters: 64, 128
            2D-CNN           Convolutional   1    1            in the reduced-dimensional space. It was worth noting
                                                               that, although there was no clear separation of feature
                             Layers: 2, Kernel
                             Size: (3, 3),                     types in the visualization space, the learned features still
                             Filters: 64, 128                  showed significant application effects in downstream tasks.
            GRU              GRU Units: 128  \    \            This may have been because t-SNE’s low-dimensional
                             Number of layers:                 visualization lost important information from the high-
                             2                                 dimensional space, and the visualization results can
            ANN              Neurons: 256, 128  \  \           not accurately reflect the intrinsic structure of the high-
            Abbreviations: 1D-CNN: One-Dimensional Convolutional Neural   dimensional feature space.
            Network; 2D-CNN: Two-Dimensional Convolutional Neural Network;
            ANN: Artificial Neural Network; GRU: Gated Recurrent Unit.  4.2. The effectiveness of downstream fatigue life
                                                               prediction

                                                               To further find the optimal downstream regression model,
                                                               the linear regression model was considered, as it is simple,




















            Figure  7.  The training process of contrastive learning model between
            different architecture models
            Abbreviations: 1D-CNN:  One-dimensional  convolutional neural
            network; 2D-CNN: Two-dimensional convolutional neural network;
            ANN: Artificial neural network; GRU: Gated recurrent unit.  Figure 8. Visualization results of the original data



            Volume 2 Issue 1 (2025)                         61                        doi: 10.36922/IJAMD025040004
   62   63   64   65   66   67   68   69   70   71   72