Page 92 - AIH-1-4
P. 92

Artificial Intelligence in Health                               Transformer-based radiology report summaries



            reports. Therefore, summarizing these  reports not  only   trained  BERT-based  models.  They  developed
            aids in the clinical workflow but also enhances patients’   ClinicalBioBERTSum, which incorporates domain-
            understanding of their medical conditions and improves   specific BERT models into the BERTSum architecture. This
            their engagement with health-care personnels.      model was applied to the MIMIC-CXR dataset, focusing
              In addition, although there is an evident need for   on predicting the “Impression” section of radiology reports
            effective summarization tools in radiology, advancements   based on the “Indication” and “Findings” sections. Their
            in  NLP  have  been  slow  to  penetrate  this  domain.  The   model achieved a recall-oriented understudy for gisting
            application of state-of-the-art NLP models to radiology   evaluation (ROUGE)-L F1 score of 57.37 and introduced
            report summarization remains relatively unexplored.   ClinicalBioBERTScore to better evaluate the semantic
            Addressing this gap presents a significant opportunity for   quality of the summaries. Their work emphasizes the
            innovation and improvement in medical communication   importance of domain-specific pre-training and fine-
            and patient care.                                  tuning in improving summarization performance for
                                                               clinical texts.  By leveraging ClinicalBioBERT, a model
              The present study addresses this gap by developing   fine-tuned on MIMIC-III clinical texts, they improved the
            and evaluating a novel summarization model tailored   representation of medical semantics in radiology reports.
            specifically for radiology reporting. By utilizing the MIMIC   The study also explored the use of custom tokenizers
            – chest X-ray (CXR) dataset,  we fine-tuned a bidirectional   and a two-stage fine-tuning process, which showed that
                                  3
            encoder  representation from transformers BERT-based   combining extractive and abstractive summarization
            model  to create Biomedical-BERT2BERT, achieving   objectives could enhance model performance.
                 2
            state-of-the-art performance in generating concise and          2
            accurate summaries of radiology findings. Our approach   Devlin  et al.  introduced BERT, a novel transformer-
            not only leverages advanced NLP techniques but also   based model pre-trained on large text corpora for various
            introduces a novel data augmentation strategy to enhance   NLP tasks. BERT’s bidirectional training allows it to capture
            model performance. This model takes multiple free-text   context from both left and right surroundings, significantly
            radiology report fields as input and uses a sequence-to-  enhancing its performance on tasks such as question
            sequence architecture to output abstract summaries. The   answering, language inference, and more. The model sets
            main contributions of our work are as follows:     new benchmarks on several NLP tasks, demonstrating the
            (1)  We developed Biomedical-BERT2BERT, an adjusted   effectiveness of pre-training on large datasets followed by
               BERT-based model with state-of-the-art performance   fine-tuning on specific tasks. The study’s impact on the field
               in radiology text summarization, assisting radiologists   is profound, as BERT has become a foundational model
               in generating concise impressions from reports  for many subsequent NLP advancements. Its architecture,
            (2)  We introduced a novel data augmentation strategy to   which utilizes multiple layers of transformers and a
               improve performance on related tasks with MIMIC-  masked language model objective, enables it to learn deep
               CXR reports                                     contextual representations. This work paved the way for
            (3)  We conducted an in-depth analysis of Biomedical-  numerous domain-specific adaptations, such as BioBERT
               BERT2BERT’s performance, knowledge gain, and    and ClinicalBERT, which tailor the model to specific fields
               limitations with respect  to disease  distribution and   by further pre-training on domain-specific texts.
               other architectures.                              Alsentzer et al.  explored the application of contextual
                                                                              6
              Our data processing pipeline and model training code   word embedding models, such as ELMo and BERT, to
            are also provided. 4                               clinical text, addressing the gap in publicly available pre-
                                                               trained models for clinical NLP tasks. They developed and
            2. Related works                                   released BERT models specifically trained on clinical text,
                                                               demonstrating significant improvements over traditional
            Initial work in this domain was conducted by Chen et al.,    BERT and BioBERT models on several clinical NLP tasks.
                                                          5
            who reported promising results in predicting radiologist   These tasks included named entity recognition (NER) and
            impressions from raw findings using fine-tuned BERT-  medical natural language inference, where their domain-
            based encoder–decoder models. We extended this work by   specific models outperformed general domain models. The
            experimenting with different architectures, understanding   study highlights the challenges of applying general pre-
            the limitations of applying language  models in this   trained models to domain-specific texts due to linguistic
            domain,  and investigating the  effectiveness  of modern   differences. By training BERT models on the MIMIC-
            linear attention mechanisms on MIMIC-CXR.          III dataset, which comprises approximately two million
              Chen  et al.  addressed  the challenge  of abstractive   clinical notes, they tailored the embeddings to better fit the
                        5
            summarization in radiology reporting using pre-    clinical context. This work is notable for providing publicly

            Volume 1 Issue 4 (2024)                         86                               doi: 10.36922/aih.3846
   87   88   89   90   91   92   93   94   95   96   97