Git Product home page Git Product logo

biomedical-text-summarization-survey's Introduction

A Survey on Biomedical Text Summarisation with Pre-trained Language Model (PLMs)

Resource

This repository contains a list of papers, codes, and datasets in Biomedical Text Summarisation based on PLM. If you found any errors, please don't hesitate to open an issue or pull a request.

Contributor

Resource Contributed by Qianqian Xie, Zheheng Luo, Benyou Wang,Sophia Ananiadou.

Introduction

Biomedical text summarization has long been a fundamental task in biomedical natural language processing (BioNLP), aiming at generating concise summaries that distil key information from one or multiple biomedical documents. In recent years, pre-trained language models (PLMs) have been the de facto standard of various natural language processing tasks in the general domain. Most recently, PLMs have been further investigated in the biomedical domain and brought new insights into the biomedical text summarization task.

To help researchers quickly grasp the development in this task and inspire further research, we line up available datasets, recent approaches and evaluation methods in this project.

At present, the project has been completely open source, including:

  1. BioTS dataset table: we listed the datasets in the BioTS field, You can find the category, size, content, and access of them in the table.
  2. PLM Based BioTS Methods: we classified and arranged papers based on the type of output summary, numbers and type of input documents. the current mainstream frontiers. Each line of the table contains the category, the strategy of applying PLM, the backbone model, the training type, and used datasets.
  3. BioTS Evaluation: we listed metrics that cover three essential aspects in the evaluation of biomedical text summarization: 1) relevancy 2) fluency 3) factuality.

The organization and our survey and the detailed background of biomedical text summarization are illustrated in the pictures below.

survey-overview

BTSwPLMs-background

Quick path

Survey Paper

  1. Text summarization in the biomedical domain: A systematic review of recent research J. biomedical informatics 2014 [html]
  2. Summarization from medical documents: a survey Artif. intelligence medicine 2005 [html]
  3. Automated methods for the summarization of electronic health records J Am Med Inform Assoc. 2015 [html]
  4. A systematic review of automatic text summarization for biomedical literature and ehrs J Am Med Inform Assoc. 2021 [html]

BTSwPLMs-Dataset

Dataset

Name Category Size Content Multi/Single Sum(M/S) Access
PubMed Biomedical literature 133,215 Full contents of articles Single https://github.com/armancohan/long-summarization
RCT Biomedical literature 4,528 Titles and abstracts of articles Multiple https://github.com/bwallace/RCT-summarization-data
MSˆ2 Biomedical literature 470,402 Abstracts of articles Multiple https://github.com/allenai/ms2/
CDSR Biomedical literature 7,805 Abstracts of articles Single hhttps://github.com/qiuweipku/Plain_language_summarization
SumPubMed Biomedical literature 33,772 Full contents of articles Single https://github.com/vgupta123/sumpubmed
S2ORC Biomedical literature 63,709 Full contents of articles Single https://github.com/jbshp/GenCompareSum
CORD-19 Biomedical literature - (constantly increasing) Full contents of articles Single https://github.com/allenai/cord19
MIMIC-CXR EHR 227,835 Full contents of reports Single https://physionet.org/content/mimic-cxr/2.0.0/
OpenI EHR 3599 Full contents of reports Single https://openi.nlm.nih.gov/faq#collection
MeQSum meidical question summarization 1000 Full contents of question Single https://github.com/abachaa/MeQSum/
CHQ-Summ meidical question summarization 1507 Full contents of question Single https://github.com/shwetanlp/Yahoo-CHQ-Summ

Methods

Paper Category Strategy Model Training Dataset
ContinualBERT [code] extractive fine-tuning BERT supervised PubMed, CORD-19
BioBERTSum extractive fine-tuning BioBERT supervised PubMed
KeBioSum [code] extractive adaption+fine-tuning PubMedBERT supervised PubMed, CORD-19, S2ORC
N. Kanwal and G. Rizzo [code] extractive fine-tuning BERT unsupervised MIMIC-III
M. Moradi et.al [code] extractive feature-base BERT unsupervised PubMed
M. Moradi et.al [code] extractive feature-base BioBERT unsupervised PubMed
GenCompareSum [code] extractive feature-base T5 unsupervised PubMed, CORD-19, S2ORC
RadBERT extractive feature-base RadBERT unsupervised -
B Tan et.al [code] hybrid adaption+fine-tuning BERT,GPT-2 supervised CORD-19
S. S. Gharebagh et.al abstractive feature-base BERT supervised MIMIC-CXR
Y. Guo et.al [code] hybrid adaption+fine-tuning BERT, BART supervised CDSR
L. Xu et.al abstractive,question adaption+fine-tuning BART,PEGASUS supervised MIMIC-CXR,OpenI,MeQSum
W. Zhu et.al abstractive fine-tuning BART,T5,PEGASUS supervised MIMIC-CXR,OpenI
R. Kondadadi et.al abstractive fine-tuning BART,T5,PEGASUS supervised MIMIC-CXR,OpenI
S. Dai et.al abstractive adaption+fine-tuning PEGASUS supervised MIMIC-CXR,OpenI
D. Mahajan et.al abstractive adaption+fine-tuning BioRoBERTa supervised MIMIC-CXR,OpenI
H. Jingpeng et.al [code] abstractive fine-tuning BioBERT supervised MIMIC-CXR,OpenI
X. Cai et.al abstractive fine-tuning SciBERT supervised CORD-19
A. Yalunin et.al abstractive adaption+fine-tuning BERT,Longformer supervised -
B. C. Wallace et.al [code] abstractive,multi-doc adaption+fine-tuning BART supervised RCT
J. DeYoung et.al [code] abstractive,multi-doc fine-tuning BART,Longformer supervised MSˆ2
A. Esteva et.al abstractive,multi-doc fine-tuning BERT,GPT-2 supervised CORD-19
CAiRE-COVID [code] hybrid,multi-doc fine-tuning,feature-base ALBERT,BART un+supervised CORD-19
HET [code] extractive,dialogue fine-tuning BERT supervised HET-MC
CLUSTER2SENT [code] abstractive,dialogue fine-tuning BERT,T5 supervised -
L. Zhang et.al [code] abstractive,dialogue fine-tuning BART supervised -
B. Chintagunt et.al abstractive,dialogue fine-tuning GPT-3 supervised -
D. F. Navarro et.al abstractive,dialogue fine-tuning BART,T5, PEGASUS supervised -
BioBART [code] abstractive,dialogue fine-tuning BioBART supervised -
Y. He et.al abstractive,question fine-tuning BART,T5,PEGASUS supervised MeQSum,MIMIC-CXR,OpenI
S. Yadav et.al abstractive,question fine-tuning BERT,ProphetNet supervised MeQSum
S. Yadav et.al abstractive,question adaption+fine-tuning Minilm supervised MeQSum
K. Mrini et.al [code] abstractive,question adaption+fine-tuning BART,BioBERT supervised MeQSum
"-" in Dataset stands for "not accessible"

Evaluation

Common metrics

ROUGE:

  • ROUGE-N: N-gram overlap between generated summaries of summarizers and gold summaries(relevancy)
  • ROUGE-L: the longest common subsequences between generated summaries of summarizers and gold summaries(fluency)

BertScore

Factual Consistency

Automatic:

Human Involved

Leader Board

Pubmed

Model ROUGE-1 ROUGE-2 ROUGE-L Paper Code Like Source
Top Down Transformer(AdaPool) 51.05 23.26 46.47 Long Document Summarization with Top-down and Bottom-up Inference (https://arxiv.org/pdf/2203.07586v1.pdf) https://github.com/kangbrilliant/DCA-Net arxiv
LongT5 50.23 24.76 46.67 LongT5: Efficient Text-To-Text Transformer for Long Sequences (https://arxiv.org/pdf/2112.07916v2.pdf) https://github.com/google-research/longt5 NAACL
MemSum (extractive) 49.25 22.94 44.42 MemSum: Extractive Summarization of Long Documents Using Multi-Step Episodic Markov Decision Processes(https://arxiv.org/pdf/2107.08929v2.pdf) https://github.com/nianlonggu/memsum ACL
HAT-BART 48.25 21.35 36.69 Hierarchical Learning for Generation with Long Source Sequences(https://arxiv.org/pdf/2104.07545v2.pdf) arxiv
DeepPyramidion 47.81 21.14 46.47 Sparsifying Transformer Models with Trainable Representation Pooling (https://aclanthology.org/2022.acl-long.590.pdf) https://github.com/applicaai/pyramidions
HiStruct+ 46.59 20.39 42.11 HiStruct+: Improving Extractive Text Summarization with Hierarchical Structure Information(https://aclanthology.org/2022.findings-acl.102.pdf) acl
DANCER PEGASUS 46.34 19.97 42.42 A Divide-and-Conquer Approach to the Summarization of Long Documents[[pdf]](https://arxiv.org/pdf/2004.06190v3.pdf) https://github.com/AlexGidiotis/DANCER-summ IEEE/ACM Transactions on Audio Speech and Language Processing
BigBird-Pegasus 46.32 20.65 42.33 Big Bird: Transformers for Longer Sequences(https://arxiv.org/pdf/2007.14062v2.pdf) https://github.com/google-research/bigbird NeuralIPS
ExtSum-LG+MMR-Select+ 45.39 20.37 40.99 Systematically Exploring Redundancy Reduction in Summarizing Long Documents(https://arxiv.org/pdf/2012.00052v1.pdf) https://github.com/Wendy-Xiao/redundancy_reduction_longdoc AACL
ExtSum-LG+RdLoss 45.3 20.42 40.95 Systematically Exploring Redundancy Reduction in Summarizing Long Documents(https://arxiv.org/pdf/2012.00052v1.pdf) https://github.com/Wendy-Xiao/redundancy_reduction_longdoc AACL

MS^2

Model ROUGE-1 ROUGE-2 ROUGE-L Paper Code Like Source
DAMEN 28.95 9.72 21.83 MDiscriminative Marginalized Probabilistic Neural Method for Multi-Document Summarization of Medical Literature (https://aclanthology.org/2022.acl-long.15.pdf) https://disi-unibo-nlp.github.io/projects/damen/ ACL
BART Hierarchical 27.56 9.40 20.80 MSˆ2: A Dataset for Multi-Document Summarization of Medical Studies (https://aclanthology.org/2021.emnlp-main.594.pdf) https://github.com/allenai/ms2/ EMNLP
LED Flat 26.89 8.91 20.32 MSˆ2: A Dataset for Multi-Document Summarization of Medical Studies (https://aclanthology.org/2021.emnlp-main.594.pdf) https://github.com/allenai/ms2/ EMNLP

MIMIC-CXR

Model ROUGE-1 ROUGE-2 ROUGE-L Paper Code Like Source
ClinicalBioBERTSumAbs 58.97 47.06 57.37 Predicting Doctor’s Impression For Radiology Reports with Abstractive Text Summarization (https://web.stanford.edu/class/cs224n/reports/final_reports/report005.pdf) Stanford CS224N
Attend to Medical Ontologies 53.57 40.78 51.81 Attend to Medical Ontologies: Content Selection for Clinical Abstractive Summarization (https://aclanthology.org/2020.acl-main.172.pdf) ACL

MEQSum

Model ROUGE-1 ROUGE-2 ROUGE-L Paper Code Source
Gradually Soft MTL + Data Aug 54.5 37.9 50.2 A Gradually Soft Multi-Task and Data-Augmented Approach to Medical Question Understanding (https://aclanthology.org/2021.acl-long.119.pdf) https://github.com/KhalilMrini/Medical-Question-Understanding ACL
Explicit QTA Knowledge-Infusion 45.20 28.38 48.76 Question-aware Transformer Models for Consumer Health Question Summarization (https://arxiv.org/pdf/2106.00219.pdf) J. Biomed. Informatics
ProphetNet + QTR + QFR 45.52 27.54 48.19 Reinforcement Learning for Abstractive Question Summarization with Question-aware Semantic Rewards (https://aclanthology.org/2021.acl-short.33.pdf) https://github.com/shwetanlp/CHQ-Summ ACL
ProphetNet + QFR 45.36 27.33 47.96 Reinforcement Learning for Abstractive Question Summarization with Question-aware Semantic Rewards (https://aclanthology.org/2021.acl-short.33.pdf) https://github.com/shwetanlp/CHQ-Summ ACL
Multi-Cloze Fusion 44.58 27.02 47.81 Question-aware Transformer Models for Consumer Health Question Summarization (https://arxiv.org/pdf/2106.00219.pdf) J. Biomed. Informatics
ProphetNet + QTR 44.60 26.69 47.38 Reinforcement Learning for Abstractive Question Summarization with Question-aware Semantic Rewards (https://aclanthology.org/2021.acl-short.33.pdf) https://github.com/shwetanlp/CHQ-Summ ACL
Implicit QTA Knowledge-Infusion 44.44 26.98 47.66 Question-aware Transformer Models for Consumer Health Question Summarization (https://arxiv.org/pdf/2106.00219.pdf) J. Biomed. Informatics
Minilm 43.13 26.03 46.39 Minilm: Deep self-attention distillation for task-agnostic compression of pretrained transformers (https://arxiv.org/pdf/2106.00219.pdf) https://github.com/microsoft/unilm/tree/master/minilm NIPS

biomedical-text-summarization-survey's People

Contributors

xashely avatar zhehengluok avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.