Top Research Papers on Text Summarization
Dive into the top research papers on text summarization to stay informed about cutting-edge developments in the field. These papers cover a range of methodologies and applications, providing valuable insights for researchers, professionals, and enthusiasts. Enhance your knowledge and keep up with the latest trends and breakthroughs in text summarization.
Looking for research-backed answers?Try AI Search
Extractive Summarization as Text Matching
402 Citations 2020Ming Zhong, Pengfei Liu, Yiran Chen + 3 more
journal unavailable
This paper forms the extractive summarization task as a semantic text matching problem, in which a source document and candidate summaries will be matched in a semantic space to create a semantic matching framework.
Automatic text summarization: A comprehensive survey
754 Citations 2020Wafaa S. El-Kassas, Cherif Salama, Ahmed Rafea + 1 more
Expert Systems with Applications
This research provides a comprehensive survey for the researchers by presenting the different aspects of ATS: approaches, methods, building blocks, techniques, datasets, evaluation methods, and future research directions.
Abstractive Text Summarization Using GAN
735 Citations 2024Tanushree Bharti, Satyam Kumar Sinha, Harshit Singhal + 2 more
International Journal of Innovative Science and Research Technology (IJISRT)
A new approach to collecting abstract data using artificial neural networks (GANs), a class of deep learning models known for their ability to create patterns of real information, shows its promise in paving the way for advanced applications in fields.
Discourse-Aware Neural Extractive Text Summarization
263 Citations 2020Jiacheng Xu, Zhe Gan, Yu Cheng + 1 more
journal unavailable
DiscoBert extracts sub-sentential discourse units (instead of sentences) as candidates for extractive selection on a finer granularity and outperforms state-of-the-art methods by a significant margin on popular summarization benchmarks compared to other BERT-base models.
A Survey of Automatic Text Summarization: Progress, Process and Challenges
128 Citations 2021M. F. Mridha, Aklima Akter Lima, Kamruddin Nur + 3 more
IEEE Access
Every modern text summarization approach’s workflow and significance are reviewed with the limitations with potential recovery methods, including the feature extraction approaches, datasets, performance measurement techniques, and challenges of the ATS domain, etc.
Review of automatic text summarization techniques & methods
205 Citations 2020Adhika Pramita Widyassari, Supriadi Rustad, Guruh Fajar Shidik + 4 more
Journal of King Saud University - Computer and Information Sciences
This paper provides a broad and systematic review of research in the field of text summarization published from 2008 to 2019 and describes the techniques and methods that are often used by researchers as a comparison and means for developing methods.
Entity-level Factual Consistency of Abstractive Text Summarization
105 Citations 2021Nan Feng, Ramesh Nallapati, Zhiguo Wang + 5 more
journal unavailable
A set of new metrics are proposed to quantify the entity-level factual consistency of generated summaries and it is shown that the entity hallucination problem can be alleviated by simply filtering the training data.
Neural Abstractive Text Summarization with Sequence-to-Sequence Models
191 Citations 2021Tian Shi, Yaser Keneshloo, Naren Ramakrishnan + 1 more
ACM/IMS Transactions on Data Science
This article provides a comprehensive literature survey on different seq2seq models for abstractive text summarization from the viewpoint of network structures, training strategies, and summary generation algorithms.
T-BERTSum: Topic-Aware Text Summarization Based on BERT
115 Citations 2021Tinghuai Ma, Qian Pan, Huan Rong + 3 more
IEEE Transactions on Computational Social Systems
A topic-aware extractive and abstractive summarization model named T-BERTSum, based on Bidirectional Encoder Representations from Transformers (BERTs), which achieves new state-of-the-art results while generating consistent topics compared with the most advanced method.
Improving text summarization of online hotel reviews with review helpfulness and sentiment
150 Citations 2020Chih‐Fong Tsai, Kuanchin Chen, Ya‐Han Hu + 1 more
Tourism Management
A systematic approach is proposed that first constructs classifiers to identify helpful reviews and then classifies the sentences in the helpful reviews into six hotel features, and the sentiment polarities of sentences are analyzed to generate the review summaries.
Context-Aware Multi-View Summarization Network for Image-Text Matching
149 Citations 2020Leigang Qu, Meng Liu, Da Cao + 2 more
journal unavailable
A novel context-aware multi-view summarization network to summarize context-enhanced visual region information from multiple views and designs an adaptive gating self-attention module to extract representations of visual regions and words.
Adapted large language models can outperform medical experts in clinical text summarization
542 Citations 2024Dave Van Veen, Cara Van Uden, Louis Blankemeier + 16 more
Nature Medicine
Comparative performance assessment of large language models identified ChatGPT-4 as the best-adapted model across a diverse set of clinical text summarization tasks, and it outperformed 10 medical experts in a reader study.
DSNet: A Flexible Detect-to-Summarize Network for Video Summarization
186 Citations 2020Wencheng Zhu, Jiwen Lu, Jiahao Li + 1 more
IEEE Transactions on Image Processing
This paper proposes a Detect-to-Summarize network (DSNet) framework for supervised video summarization that contains anchor-based and anchor-free counterparts, and provides a dense sampling of temporal interest proposals with multi-scale intervals that accommodate interest variations in length.
MLSUM: The multilingual summarization corpus
116 Citations 2020Thomas Scialom
Institutional Research Information System (Università degli Studi di Trento)
We present MLSUM, the first large-scale MultiLingual SUMmarization dataset. Obtained from online newspapers, it contains 1.5M+ article/summary pairs in five different languages – namely, French, German, Spanish, Russian, Turkish. Together with English news articles from the popular CNN/Daily mail dataset, the collected data form a large scale multilingual dataset which can enable new research directions for the text summarization community. We report cross-lingual comparative analyses based on state-of-the-art systems. These highlight existing biases which motivate the use of a multi-lingual d...
Event Summarization Using Tweets
292 Citations 2021Deepayan Chakrabarti, Kunal Punera
Proceedings of the International AAAI Conference on Web and Social Media
It is argued that for some highly structured and recurring events, such as sports, it is better to use more sophisticated techniques to summarize the relevant tweets, and a solution based on learning the underlying hidden state representation of the event via Hidden Markov Models is given.
Document Summarization Based on Data Reconstruction
119 Citations 2021Zhanying He, Chun Chen, Jiajun Bu + 4 more
Proceedings of the AAAI Conference on Artificial Intelligence
This paper proposes a novel framework named Document Summarization based on Data Reconstruction (DSDR), which generates a summary which consist of those sentences that can best reconstruct the original document.
Enhancing Factual Consistency of Abstractive Summarization
122 Citations 2021Chenguang Zhu, William Hinthorn, Ruochen Xu + 4 more
journal unavailable
A fact-aware summarization model FASum is proposed to extract and integrate factual relations into the summary generation process via graph attention and a factual corrector model FC is designed to automatically correct factual errors from summaries generated by existing systems.
BRIO: Bringing Order to Abstractive Summarization
193 Citations 2022Yixin Liu, Pengfei Liu, Dragomir Radev + 1 more
Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
ive summarization models are commonly trained using maximum likelihood estimation, which assumes a deterministic (one-point) target distribution in which an ideal model will assign all the probability mass to the reference summary. This assumption may lead to performance degradation during inference, where the model needs to compare several system-generated (candidate) summaries that have deviated from the reference summary. To address this problem, we propose a novel training paradigm which assumes a non-deterministic distribution so that different candidate summaries are assigned probability...
Efficient Attentions for Long Document Summarization
125 Citations 2021Luyang Huang, Shuyang Cao, Nikolaus Nova Parulian + 2 more
journal unavailable
Hepos, a novel efficient encoder-decoder attention with head-wise positional strides to effectively pinpoint salient information from the source is proposed, able to process ten times more tokens than existing models that use full attentions.
Retrieval-based neural source code summarization
235 Citations 2020Jian Zhang, Xu Wang, Hongyu Zhang + 2 more
journal unavailable
This paper proposes a retrieval-based neural source code summarization approach where the neural model is enhanced with the most similar code snippets retrieved from the training set, and the experimental results show that the proposed approach can improve the state-of-the-art methods.