Top Research Papers on Text Summarization
Dive into the top research papers on text summarization to stay informed about cutting-edge developments in the field. These papers cover a range of methodologies and applications, providing valuable insights for researchers, professionals, and enthusiasts. Enhance your knowledge and keep up with the latest trends and breakthroughs in text summarization.
Looking for research-backed answers?Try AI Search
Extractive Summarization as Text Matching
402 Citations 2020Ming Zhong, Pengfei Liu, Yiran Chen + 3 more
journal unavailable
This paper forms the extractive summarization task as a semantic text matching problem, in which a source document and candidate summaries will be matched in a semantic space to create a semantic matching framework.
Automatic text summarization: A comprehensive survey
754 Citations 2020Wafaa S. El-Kassas, Cherif Salama, Ahmed Rafea + 1 more
Expert Systems with Applications
This research provides a comprehensive survey for the researchers by presenting the different aspects of ATS: approaches, methods, building blocks, techniques, datasets, evaluation methods, and future research directions.
Abstractive Text Summarization Using GAN
735 Citations 2024Tanushree Bharti, Satyam Kumar Sinha, Harshit Singhal + 2 more
International Journal of Innovative Science and Research Technology (IJISRT)
A new approach to collecting abstract data using artificial neural networks (GANs), a class of deep learning models known for their ability to create patterns of real information, shows its promise in paving the way for advanced applications in fields.
Discourse-Aware Neural Extractive Text Summarization
263 Citations 2020Jiacheng Xu, Zhe Gan, Yu Cheng + 1 more
journal unavailable
DiscoBert extracts sub-sentential discourse units (instead of sentences) as candidates for extractive selection on a finer granularity and outperforms state-of-the-art methods by a significant margin on popular summarization benchmarks compared to other BERT-base models.
A Survey of Automatic Text Summarization: Progress, Process and Challenges
128 Citations 2021M. F. Mridha, Aklima Akter Lima, Kamruddin Nur + 3 more
IEEE Access
Every modern text summarization approach’s workflow and significance are reviewed with the limitations with potential recovery methods, including the feature extraction approaches, datasets, performance measurement techniques, and challenges of the ATS domain, etc.
Review of automatic text summarization techniques & methods
205 Citations 2020Adhika Pramita Widyassari, Supriadi Rustad, Guruh Fajar Shidik + 4 more
Journal of King Saud University - Computer and Information Sciences
This paper provides a broad and systematic review of research in the field of text summarization published from 2008 to 2019 and describes the techniques and methods that are often used by researchers as a comparison and means for developing methods.
Entity-level Factual Consistency of Abstractive Text Summarization
105 Citations 2021Nan Feng, Ramesh Nallapati, Zhiguo Wang + 5 more
journal unavailable
A set of new metrics are proposed to quantify the entity-level factual consistency of generated summaries and it is shown that the entity hallucination problem can be alleviated by simply filtering the training data.
Neural Abstractive Text Summarization with Sequence-to-Sequence Models
191 Citations 2021Tian Shi, Yaser Keneshloo, Naren Ramakrishnan + 1 more
ACM/IMS Transactions on Data Science
This article provides a comprehensive literature survey on different seq2seq models for abstractive text summarization from the viewpoint of network structures, training strategies, and summary generation algorithms.
T-BERTSum: Topic-Aware Text Summarization Based on BERT
115 Citations 2021Tinghuai Ma, Qian Pan, Huan Rong + 3 more
IEEE Transactions on Computational Social Systems
A topic-aware extractive and abstractive summarization model named T-BERTSum, based on Bidirectional Encoder Representations from Transformers (BERTs), which achieves new state-of-the-art results while generating consistent topics compared with the most advanced method.
Improving text summarization of online hotel reviews with review helpfulness and sentiment
150 Citations 2020Chih‐Fong Tsai, Kuanchin Chen, Ya‐Han Hu + 1 more
Tourism Management
A systematic approach is proposed that first constructs classifiers to identify helpful reviews and then classifies the sentences in the helpful reviews into six hotel features, and the sentiment polarities of sentences are analyzed to generate the review summaries.
Context-Aware Multi-View Summarization Network for Image-Text Matching
149 Citations 2020Leigang Qu, Meng Liu, Da Cao + 2 more
journal unavailable
A novel context-aware multi-view summarization network to summarize context-enhanced visual region information from multiple views and designs an adaptive gating self-attention module to extract representations of visual regions and words.
Adapted large language models can outperform medical experts in clinical text summarization
542 Citations 2024Dave Van Veen, Cara Van Uden, Louis Blankemeier + 16 more
Nature Medicine
Comparative performance assessment of large language models identified ChatGPT-4 as the best-adapted model across a diverse set of clinical text summarization tasks, and it outperformed 10 medical experts in a reader study.
DSNet: A Flexible Detect-to-Summarize Network for Video Summarization
186 Citations 2020Wencheng Zhu, Jiwen Lu, Jiahao Li + 1 more
IEEE Transactions on Image Processing
This paper proposes a Detect-to-Summarize network (DSNet) framework for supervised video summarization that contains anchor-based and anchor-free counterparts, and provides a dense sampling of temporal interest proposals with multi-scale intervals that accommodate interest variations in length.
MLSUM: The multilingual summarization corpus
116 Citations 2020Thomas Scialom
Institutional Research Information System (Università degli Studi di Trento)
We present MLSUM, the first large-scale MultiLingual SUMmarization dataset. Obtained from online newspapers, it contains 1.5M+ article/summary pairs in five different languages – namely, French, German, Spanish, Russian, Turkish. Together with English news articles from the popular CNN/Daily mail dataset, the collected data form a large scale multilingual dataset which can enable new research directions for the text summarization community. We report cross-lingual comparative analyses based on state-of-the-art systems. These highlight existing biases which motivate the use of a multi-lingual d...
Event Summarization Using Tweets
292 Citations 2021Deepayan Chakrabarti, Kunal Punera
Proceedings of the International AAAI Conference on Web and Social Media
It is argued that for some highly structured and recurring events, such as sports, it is better to use more sophisticated techniques to summarize the relevant tweets, and a solution based on learning the underlying hidden state representation of the event via Hidden Markov Models is given.
Document Summarization Based on Data Reconstruction
119 Citations 2021Zhanying He, Chun Chen, Jiajun Bu + 4 more
Proceedings of the AAAI Conference on Artificial Intelligence
This paper proposes a novel framework named Document Summarization based on Data Reconstruction (DSDR), which generates a summary which consist of those sentences that can best reconstruct the original document.
Enhancing Factual Consistency of Abstractive Summarization
122 Citations 2021Chenguang Zhu, William Hinthorn, Ruochen Xu + 4 more
journal unavailable
A fact-aware summarization model FASum is proposed to extract and integrate factual relations into the summary generation process via graph attention and a factual corrector model FC is designed to automatically correct factual errors from summaries generated by existing systems.
BRIO: Bringing Order to Abstractive Summarization
193 Citations 2022Yixin Liu, Pengfei Liu, Dragomir Radev + 1 more
Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
ive summarization models are commonly trained using maximum likelihood estimation, which assumes a deterministic (one-point) target distribution in which an ideal model will assign all the probability mass to the reference summary. This assumption may lead to performance degradation during inference, where the model needs to compare several system-generated (candidate) summaries that have deviated from the reference summary. To address this problem, we propose a novel training paradigm which assumes a non-deterministic distribution so that different candidate summaries are assigned probability...
Efficient Attentions for Long Document Summarization
125 Citations 2021Luyang Huang, Shuyang Cao, Nikolaus Nova Parulian + 2 more
journal unavailable
Hepos, a novel efficient encoder-decoder attention with head-wise positional strides to effectively pinpoint salient information from the source is proposed, able to process ten times more tokens than existing models that use full attentions.
A comprehensive survey of multi-view video summarization
144 Citations 2020Tanveer Hussain, Khan Muhammad, Weiping Ding + 3 more
Pattern Recognition
An overview of the existing strategies proposed for MVS is presented, including their advantages and drawbacks, and the genericsteps in MVS, such as the pre-processing of video data, feature extraction, and post-processing followed by summary generation are described.
Benchmarking Large Language Models for News Summarization
257 Citations 2024Tianyi Zhang, Faisal Ladhak, Esin Durmus + 3 more
Transactions of the Association for Computational Linguistics
It is found instruction tuning, not model size, is the key to the LLM’s zero-shot summarization capability and summaries are judged to be on par with human written summaries.
Retrieval-based neural source code summarization
235 Citations 2020Jian Zhang, Xu Wang, Hongyu Zhang + 2 more
journal unavailable
This paper proposes a retrieval-based neural source code summarization approach where the neural model is enhanced with the most similar code snippets retrieved from the training set, and the experimental results show that the proposed approach can improve the state-of-the-art methods.
Factual Error Correction for Abstractive Summarization Models
121 Citations 2020Meng Cao, Yue Dong, Jiapeng Wu + 1 more
journal unavailable
This work proposes a post-editing corrector module that is able to correct factual errors in summaries generated by other neural summarization models and outperforms previous models on factual consistency evaluation on the CNN/DailyMail dataset.
A Transformer-based Approach for Source Code Summarization
349 Citations 2020Wasi Uddin Ahmad, Saikat Chakraborty, Baishakhi Ray + 1 more
journal unavailable
This work explores the Transformer model that uses a self-attention mechanism and has shown to be effective in capturing long-range dependencies in source code summarization, and shows that despite the approach is simple, it outperforms the state-of-the-art techniques by a significant margin.
Reconstructive Sequence-Graph Network for Video Summarization
121 Citations 2021Bin Zhao, Haopeng Li, Xiaoqiang Lu + 1 more
IEEE Transactions on Pattern Analysis and Machine Intelligence
A reconstructive sequence-graph network (RSGN) is proposed to encode the frames and shots as sequence and graph hierarchically, where the frame-level dependencies are encoded by long short-term memory (LSTM), and the shot- level dependencies are captured by the graph convolutional network (GCN).
News Summarization and Evaluation in the Era of GPT-3
180 Citations 2022Tanya Goyal, Junyi Jessy Li, Greg Durrett
arXiv (Cornell University)
It is shown that not only do humans overwhelmingly prefer GPT-3 summaries, prompted using only a task description, but these also do not suffer from common dataset-specific issues such as poor factuality.
Text-to-Text Pre-Training for Data-to-Text Tasks
138 Citations 2020Mihir Kale, Abhinav Rastogi
journal unavailable
It is indicated that text-to-text pre-training in the form of T5 enables simple, end- to-end transformer based models to outperform pipelined neural architectures tailored for data-to/text generation, as well as alternatives such as BERT and GPT-2.
Heterogeneous Graph Neural Networks for Extractive Document Summarization
273 Citations 2020Danqing Wang, Pengfei Liu, Yining Zheng + 2 more
journal unavailable
This paper presents a heterogeneous graph-based neural network for extractive summarization (HETERSUMGRAPH), which contains semantic nodes of different granularity levels apart from sentences that act as the intermediary between sentences and enrich the cross-sentence relations.
DialogSum: A Real-Life Scenario Dialogue Summarization Dataset
130 Citations 2021Yulong Chen, Yang Liu, Chen Liang + 1 more
journal unavailable
Experimental results show unique challenges in dialogue summarization, such as spoken terms, special discourse structures, coreferences and ellipsis, pragmatics and social common sense, which require specific representation learning technologies to better deal with.
Evaluating large language models on medical evidence summarization
313 Citations 2023Liyan Tang, Zhaoyi Sun, Betina Idnay + 9 more
npj Digital Medicine
Abstract Recent advances in large language models (LLMs) have demonstrated remarkable successes in zero- and few-shot performance on various downstream tasks, paving the way for applications in high-stakes domains. In this study, we systematically examine the capabilities and limitations of LLMs, specifically GPT-3.5 and ChatGPT, in performing zero-shot medical evidence summarization across six clinical domains. We conduct both automatic and human evaluations, covering several dimensions of summary quality. Our study demonstrates that automatic metrics often do not strongly correlate with the ...
GSum: A General Framework for Guided Neural Abstractive Summarization
199 Citations 2021Zi-Yi Dou, Pengfei Liu, Hiroaki Hayashi + 2 more
journal unavailable
A general and extensible guided summarization framework that can effectively take different kinds of external guidance as input is proposed and demonstrated, and how different types of guidance generate qualitatively different summaries is demonstrated, lending a degree of controllability to the learned models.
Leveraging Graph to Improve Abstractive Multi-Document Summarization
117 Citations 2020Wei Li, Xinyan Xiao, Jiachen Liu + 3 more
journal unavailable
A neural abstractive multi-document summarization (MDS) model which can leverage well-known graph representations of documents, to more effectively process multiple input documents and produce abstractive summaries is developed.
SimCLS: A Simple Framework for Contrastive Learning of Abstractive Summarization
209 Citations 2021Yixin Liu, Pengfei Liu
journal unavailable
SimCLS can bridge the gap between the learning objective and evaluation metrics resulting from the currently dominated sequence-to-sequence learning framework by formulating text generation as a reference-free evaluation problem assisted by contrastive learning.
Mining Feature-Opinion in Online Customer Reviews for Opinion Summarization
121 Citations 2020Gamgarn Somprasertsri, Pattarachai Lalitrojwong
TUGraz OPEN Library (Graz University of Technology)
This paper proposed an approach for mining product feature and opinion based on the consideration of syntactic information and semantic information and shows that the approach is more flexible and effective.
BARTScore: Evaluating Generated Text as Text Generation
318 Citations 2021Weizhe Yuan, Graham Neubig, Pengfei Liu
arXiv (Cornell University)
A wide variety of NLP applications, such as machine translation, summarization, and dialog, involve text generation. One major challenge for these applications is how to evaluate whether such generated texts are actually fluent, accurate, or effective. In this work, we conceptualize the evaluation of generated text as a text generation problem, modeled using pre-trained sequence-to-sequence models. The general idea is that models trained to convert the generated text to/from a reference output or the source text will achieve higher scores when the generated text is better. We operationalize th...
Few-shot training LLMs for project-specific code-summarization
185 Citations 2022Toufique Ahmed, Prémkumar Dévanbu
journal unavailable
This paper investigates the use few-shot training with the very large GPT (Generative Pre-trained Transformer) Codex model, and finds evidence suggesting that one can significantly surpass state-of-the-art models for code-summarization, leveraging project-specific training.
PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization
370 Citations 2020Jingqing Zhang, Yao Zhao, Mohammad Saleh + 1 more
International Conference on Machine Learning
This work proposes pre-training large Transformer-based encoder-decoder models on massive text corpora with a new self-supervised objective, PEGASUS, and demonstrates it achieves state-of-the-art performance on all 12 downstream datasets measured by ROUGE scores.
Multi-document Summarization via Deep Learning Techniques: A Survey
114 Citations 2022Congbo Ma, Wei Emma Zhang, Mingyu Guo + 2 more
ACM Computing Surveys
This survey, the first of its kind, systematically overviews the recent deep-learning-based MDS models and proposes a novel taxonomy to summarize the design strategies of neural networks and conduct a comprehensive summary of the state of the art.
On Extractive and Abstractive Neural Document Summarization with Transformer Language Models
184 Citations 2020Jonathan Pilault, Raymond Li, Sandeep Subramanian + 1 more
journal unavailable
A method to produce abstractive summaries of long documents that exceed several thousand words via neural abstractive summarization via transformer language model is presented, finding that transformers are ranked highly for coherence and fluency, but purely extractive methods score higher for informativeness and relevance.
CLIFF: Contrastive Learning for Improving Faithfulness and Factuality in Abstractive Summarization
129 Citations 2021Shuyang Cao, Lu Wang
Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing
It is found that the contrastive learning framework consistently produces more factual summaries than strong comparisons with post error correction, entailment-based reranking, and unlikelihood training, according to QA-based factuality evaluation.