Page 93 - AIH-1-4
P. 93
Artificial Intelligence in Health Transformer-based radiology report summaries
accessible resources that can be utilized by the wider our current work focuses on text-only data. Future work
research community to advance clinical NLP applications. could explore incorporating multimodal data to enhance
The models showed robust performance across various our model further.
tasks, although they noted limitations in de-identification Separately, a comprehensive review by Zhang et al.
12
tasks due to differences in data characteristics between recent advancements in NLP for medical text processing
training and task datasets. highlights the latest trends and future directions,
The T5 (Text-To-Text Transfer Transformer) model, contextualizing our work within the broader landscape
created by Raffel et al., frames all NLP tasks as text-to-text of NLP advancements in medical text processing. Our
8
problems. This approach allows for a unified framework contributions align with and extend these current
where both inputs and outputs are treated as text strings, trends, offering novel solutions for radiology report
simplifying the architecture and training process. T5 is pre- summarization.
trained on a large dataset (C4) and fine-tuned on various
downstream tasks, achieving state-of-the-art results across 2.1. Tokenizers
a wide range of benchmarks. The study highlights the BERT-based models have been trained on word-
versatility and efficiency of the text-to-text framework, splits tokenizers on several corpora, mainly wiki-data
demonstrating its applicability to tasks such as translation, and literature datasets in the process usually called
summarization, and question–answering. Using a tokenization. Tokenization is breaking the raw text into
consistent model structure for different tasks, T5 reduces small chunks. Tokenization breaks the raw text into
the complexity of developing task-specific models. The words and sentences called tokens. These tokens help in
success of T5 underscores the potential of transfer learning understanding the context or developing the model for the
and model unification in advancing the capabilities of NLP NLP. The tokenization helps in interpreting the meaning of
systems. the text by analyzing the sequence of the words.
Li et al. investigated the adaptation of long-sequence 2.2. Pre-trained language models (PLMs)
9
transformer models, such as Longformer and BigBird, to
clinical NLP tasks. These models address the limitations PLMs are large neural networks that are used in a wide
of traditional transformers such as BERT, which are variety of NLP tasks. They operate under a pre-train-
constrained by a maximum input sequence length of finetune paradigm: Models are first pre-trained over a large
512 tokens. By employing sparse attention mechanisms, text corpus and then fine-tuned on a downstream task using
Clinical-Longformer and Clinical-BigBird can handle additional datasets. Most common architectures, such as
8
6
sequences up to 4096 tokens, making them suitable for BERT and T5, have not been pre-trained on specialized
the lengthy documents common in clinical contexts. Their medical corpora. We have fine-tuned our model in the
study involved pre-training these models on large-scale MIMIC-CXR dataset, which is a large publicly available
clinical corpora and evaluating them on a variety of NLP dataset of chest radiographs, free-text radiology reports,
tasks, including NER, question answering, and document and structured labels.
classification. The results demonstrated that both 2.3. Evaluation metrics
Clinical-Longformer and Clinical-BigBird significantly
outperformed ClinicalBERT and other short-sequence We evaluated summarization generation performance
transformers across all tasks. This work underscores with a recall-oriented understudy for gisting evaluation, or
7
the potential of long-sequence models to improve the ROUGE on F1 metrics. Historically, ROUGE has shown
processing and analysis of extensive clinical texts, paving a good correlation with human-evaluated summaries and
the way for more effective NLP tools in health care. is a canonical metric for summarization evaluation. We
focused on a variant of ROUGE, called ROUGE-L which
The application of transformer-based models, such measures the longest common subsequence (LCS) overlap
as BERT, GPT-3, and T5, in medical text summarization between the predicted and reference summaries to evaluate
has been explored by Yalunin et al. They found that fine- the informativeness of the summary.
10
tuning these models on medical datasets significantly
improves their performance. Compared to their findings, 3. Approaches
our Biomedical-BERT2BERT model demonstrates
superior performance due to our novel data augmentation 3.1. Text summarization
techniques. Kraljevic et al. proposed a multimodal Our task, text summarization for biomedical documents,
11
approach combining text and image data for summarizing can be approached by either extractive or abstractive
medical documents. While their method shows promise, methods. Extractive summaries are snippets taken directly
Volume 1 Issue 4 (2024) 87 doi: 10.36922/aih.3846

