Page 63 - IJAMD-2-1
P. 63

International Journal of AI for
            Materials and Design
                                                                             Fatigue life prediction via contrastive learning



                          A                   B               C                      D










                           Figure 2. The diagram of multiaxial loading paths: (A) uniaxial, (B) cross, (C) rhombus, and (D) circular

            Table 1. The detailed information of specimen for each   loading path, the strain amplitude of the augmented samples
            loading path type                                  was distributed between the maximum and minimum
                                                               values of the experimental data. Based on these augmented
            Loading path       εa         γa           Nf      samples, 100  samples were uniformly selected for each
            Uniaxial           0.2         -         161000    loading path. A  total of 400 augmented samples were
                               0.4         -          11865    selected for the training of the contrast learning model. Data
                               0.6         -          3339     augmentation effectively increases the training sample size
                               0.8         -          1719     and improves the performance of the contrastive learning
                               1.0         -          906      model. In addition,  only  stress-strain  hysteresis loops
            Cross              0.3        0.3         10023    are required, without the need for fatigue life data, which
                              0.35        0.35        5831     reduces the difficulty of using augmented samples. In the
                                                               subsequent downstream fatigue life prediction task, there
                               0.4        0.4         2038     are only the original 20 experimental datasets used for the
                               0.5        0.5         1378     training. In addition, in future work, constitutive material
                               0.6        0.6         1311     models or finite element simulations can also be employed
            Rhombus            0.3        0.3         7193     as alternative methods for data augmentation.
                              0.35        0.35        4041     2.3. Preprocessing
                               0.4        0.4         3944
                               0.5        0.5         1031     According to the previous introduction, the stress-strain
                               0.6        0.6         795      hysteresis loop will serve as input features for the training
                                                               of following contrastive learning and machine learning
            Circular           0.3        0.3         5045     model. The output feature of the downstream supervised
                              0.35        0.35        2963     model is the logarithmic fatigue life. It can avoid the effect
                               0.4        0.4         1802     of its large magnitude range. Furthermore, all input and
                               0.5        0.5         744      output features are normalized using the z-score method
                               0.6        0.6         396      to ensure that the model captures the relative relationships
                                                               between features rather than being influenced by their
                                                               absolute values. The z-score normalization is described as
            provides the amplitude and life information for each   Equation I:
            experimental specimen.
                                                                   x   x
            2.2. Data augmentation                             x                                           (I)
                                                                     x
            To ensure the effectiveness of contrastive learning, data
            augmentation was also applied to enhance the training   Where  x represents the normalized data, x represents
            samples. The generative adversarial network (GAN)-based   the raw data,  μ   is the mean of the raw data, and  σ
                                                                            x
                                                                                                             x
                                                               represents the standard deviation of the entire sample
            augmentation method used was derived from our previous   space.
            research.  By integrating CNN with the Fourier transform,
                   16
            the proposed model achieves synergistic augmentation   3. Algorithm
            of stress-strain hysteresis loops and fatigue life prediction.
            Based on the data augmentation capability of this model,   3.1. Contrastive learning architecture
            the 20 experimental samples were expanded to over   Contrastive learning is a self-supervised learning strategy
            1000  samples for each loading path. For each multiaxial   that enables the model to learn to distinguish between


            Volume 2 Issue 1 (2025)                         57                        doi: 10.36922/IJAMD025040004
   58   59   60   61   62   63   64   65   66   67   68