4.6 Article

Somun: entity-centric summarization incorporating pre-trained language models

Journal

NEURAL COMPUTING & APPLICATIONS
Volume 33, Issue 10, Pages 5301-5311

Publisher

SPRINGER LONDON LTD
DOI: 10.1007/s00521-020-05319-2

Keywords

Automatic text summarization; Language models; Harmonic centrality

Ask authors/readers for more resources

This study introduces an entity-centric summarization method that extracts named entities using a dependency parser and generates a small graph, ranking entities using the harmonic centrality algorithm. Experimental results show that it outperforms state-of-the-art unsupervised learning baselines by more than 10% for ROUGE-1 and more than 50% for ROUGE-2 scores, achieving comparable results to recent end-to-end models.
Text summarization resolves the issue of capturing essential information from a large volume of text data. Existing methods either depend on the end-to-end models or hand-crafted preprocessing steps. In this study, we propose an entity-centric summarization method which extracts named entities and produces a small graph with a dependency parser. To extract entities, we employ well-known pre-trained language models. After generating the graph, we perform the summarization by ranking entities using the harmonic centrality algorithm. Experiments illustrate that we outperform the state-of-the-art unsupervised learning baselines by improving the performance more than 10% for ROUGE-1 and more than 50% for ROUGE-2 scores. Moreover, we achieve comparable results to recent end-to-end models.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available