Page 346 - IJB-9-4
P. 346

International Journal of Bioprinting                                      Bioprinting with machine learning



            attributes with enormous initial values from being larger   the opposite direction of the data propagation in the neural
            than those with smaller initial values.            network model. The essence of neural network training is
               The classification judgment rules in the algorithm   to adjust the parameters of each layer iteratively according
            are often majority voting . The most class value of the   to the error between the actual output and the expected
                                [35]
                                                                    [41]
            k nearest training pieces of the input data determines its   output . In the training stage, the gradient of the loss
            category. In the regression task, the average value of  k   function is calculated layer by layer through the back
            nearest neighbors can be used as the predictive value.  propagation algorithm, the output error of each layer in the
                                                               network is fed back to the upper layer through the iterative
            3.2. Artificial neural network                     method, and then the weight of the network is updated
            Compared with traditional machine learning algorithms   layer by layer, so that the theoretical value is closer to the
            such as the k-nearest neighbor, a neural network has better   sample value. After several iterations, the output error of
            generalization ability, and it can automatically extract the   the objective function is reduced to the minimum value,
            characteristics of sample data for classification, regression,   and the trained model is obtained .
                                                                                          [42]
            and nonlinear  operations.  The  simplest  artificial neural   Gradient  refers  to  the  expression  that  expresses  the
            network consists of three layers, an input layer, a hidden   partial derivative of a parameter in a function of multiple
            layer, and an output layer. The data comes in from the   variables in the form of a vector. In the geometric sense,
            input layer, and after being processed by the hidden layer,   it  represents  the  place  where  the  function  changes  and
            the output layer outputs the final result . The most critical   increases the most. The gradient descent algorithm is a
                                          [36]
            section is the hidden layer, where most of the data feature   search-based optimization  method,  which  is  a  common
            extraction and generalization work is done. An artificial   method to minimize the loss function. Its mathematical
            neural network containing multiple hidden layers is often   theory is the chain derivative rule of compound function.
            called a deep neural network, and the corresponding   The gradient descent algorithm mainly obtains the gradient
            learning process is called deep learning .         of the objective function for all variables in the process of
                                           [37]
               The data sequence from the input layer through the   training the neural network. In the operation of the neural
            hidden  layer and  finally output  in the  output layer  is   network gradient descent, the weight is updated through
            considered to be positive. Such a neural network model   the negative gradient direction, and the effective optimal
            is called a feedforward neural network. In the process   way of the objective function can be obtained .
                                                                                                   [43]
            of feedforward neural network model training, the     There are two commonly used gradient descent
            parameters of each neuron are constantly modified until   algorithms: batch gradient descent algorithm and random
            the final model can better fit the data in the training   gradient descent algorithm. The difference between them
            set. Feedforward neural network usually uses the back   mainly lies in the number of samples used to obtain gradient
            propagation algorithm based on gradient descent to update   and update parameters. The former uses all samples, while
            neuron parameters .                                the latter only selects one sample at random. Random
                           [38]
            3.2.1. Feedforward neural network                  gradient descent algorithm has advantages in training
            The feedforward neural network is  a classical neural   speed because only one sample is randomly selected to
            network model, which is also called multilayer perceptron.   update parameters, while the batch gradient descent
            Feedforward refers to the one-way propagation of   algorithm has more advantages in convergence speed and
                                                                                                   [44]
            parameters from the input end to the output end. The   can converge to the local optimal point faster .
            model itself and the output of the model do not form a   3.3. Convolutional neural network
            directed ring . There is no feedback connection, and the   The simple network structure of the artificial neural network
                      [39]
            information always flows to the output end. When there is   model is the reason for the loss of spatial information in
            a feedback connection in the network, the model does not   vector  space,  the difficulty  of multi-parameter  training,
            belong to the feedforward neural networks. In the forward   and the problems of network  overfitting.  However, a
            propagation stage, the input feature vector is processed,   convolutional neural network can better solve the defects
            and the output value of each node is calculated. If there is a   of artificial neural networks. Its main characteristics are
            deviation between the real value of the output layer and the   local connection and parameter sharing, and it is easier to
            expected value, the error is propagated back .     optimize the network by reducing the number of weights,
                                               [40]
            3.2.2. Back propagation algorithm                  thus reducing the risk of model overfitting. Convolutional
            A back propagation refers to the order in which neurons   neural networks have a significant improvement in large
            update parameters first from the output layer, then from   image processing performance compared with artificial
                                                                            [45]
            the hidden layer, and finally from the input layer, which is   neural networks .

            Volume 9 Issue 4 (2023)                        338                         https://doi.org/10.18063/ijb.739
   341   342   343   344   345   346   347   348   349   350   351