Page 66 - TD-4-1
P. 66

Tumor Discovery                                                  Drug repurposing for pancreatic cancer via AI




                          A                                    B











































            Figure 3. Flowchart of the systematic drug repurposing approach for treating PDAC. (A) Pre-training process of the deep neural network-based drug-
            target interaction model using drug-target interaction databases. (B) Prediction of candidate molecular drugs for PDAC biomarkers and selection of multi-
            molecular drug combination based on three drug design specifications
            Abbreviations: DNN: Deep neural network; DTI: Drug-target interaction; PDAC: Pancreatic ductal adenocarcinoma.

            75% for training and 25% for testing. We then utilized the   The output represents the probability of a drug-target
            DNN-DTI model to predict the probability of drug-target   interaction, ensuring the result lies within (0,1). The
            interactions. The architecture of the DNN model comprises   following binary cross-entropy was chosen as the loss
            an input layer with 996 neurons, followed by hidden layers   function to address the binary classification problem of
            with  512,  256,  128,  and 64  neurons,  respectively.  The   drug-target interactions:
            rectified linear unit (ReLU) activation function was applied   C  ) = , ˆ  −    log  + ˆ p  ˆ   p
                                                                                                   
                                                                   n
            in the hidden layers, and a dropout rate of 0.2 was set to   n ( pp n    n  n  ( − 1 p n ) ( −log 1 p n  )   (XXXIX)
            prevent overfitting. The output layer consists of a single   1  N
                                                                           C
                                                                                 ˆ ,
            neuron with a sigmoid activation function, as drug-target   L ( ,wb ) =  ∑ n ( p p n )        (XL)
                                                                               n
                                                                       N
            interaction prediction is a binary classification problem.   where  n represents the nth sample,  N is the total
                                                                          =1
                                                                         n
            The mathematical formula for each layer is as follows:  number of samples,  p n is the empirical probability of a
            O       n   b                    (XXXVIII)     positive interaction, 1−p n is the empirical probability of
                A wI
                                                                                  ˆ p
             n
                                                                                     ˆ p
              where  A is the activation function (either ReLU or   a negative interaction,   n  is the estimated probability of
                                                                                        is the estimated probability
                                                               a positive interaction, and
                                                                                      n
            Sigmoid); I n and O n represent the input and output of the   of a negative interaction. The loss function computed the
            n-th feature vector, respectively; w and b are the weight   average loss L(w, b) over all samples by averaging the loss
                                                                     ˆ ,
            vector and bias vector, respectively.              C n ( pp n ) .
                                                                   n
            Volume 4 Issue 1 (2025)                         58                                doi: 10.36922/td.4709
   61   62   63   64   65   66   67   68   69   70   71