Abstract
In the era of Big Data, we are faced with an inevitable and challenging problem of “overload information ”. To alleviate this problem, it is important to use effective automatic text summarization techniques to obtain the key information quickly and efficiently from the huge amount of text. In this paper, we propose a hybrid method of extractive text summarization based on deep learning and graph ranking algorithms (ETSDG). In this method, a pre‑trained deep learning model is designed to yield useful sentence embeddings. Given the association between sentences in raw documents, a traditional LexRank algorithm with fine‑tuning is adopted fin ETSDG. In order to improve the performance of the extractive text summarization method, we further integrate the traditional LexRank algorithm with deep learning. Testing results on the data set DUC2004 show that ETSDG has better performance in ROUGE metrics compared with certain benchmark methods.
With the rapid development of Internet in the 21st century and the exponential growth of text data, it has become an urgent problem for users to quickly and effectively distill the required useful information from the massive informatio
There are two main forms of text summarization generation, namely extractive and abstractive. Abstractive summarizations require complex language processing and generate sentences with poor readability. In contrast, extractive summarizations generate sentences with better readability and now more maturely developed. Hence, in this paper, we focus on extractive summarizations. Extractive summarizations, as the name implies, consist of abstracting key text units, including but not limit to words, phrases and sentences, etc., from original texts. Extractive text summarizations usually retain the salient information of the source text, have correct syntax, and are suitable for long text summarizations.
Most existing extractive summarization methods use sentences as the basic text unit for extraction, since they are the smallest grammatical unit that can be expressed. However, this kind of methods usually faces two challenges:(1) How to extract and represent sentences correctly?(2) How to select and sort sentences that reflect the meaning of the source text?
To address these two problems, we propose a new extractive text summarization method: A hybrid method of extractive text summarization combining deep learning and graph ranking algorithms (ETSDG). ETSDG is a combination of a deep learning pre-trained model and the traditional algorithm LexRank. The main contributions of this work are as follows:
(1) To better express sentences, we propose a new sentence representation method using deep learning technologies. A pooling operation is added to yield interpretable sentence embeddings.
(2) To select the sentences that are more relevant to the whole text, we obtain a threshold of 0.3 for setting the best result while fine-tuning in ETSDG.
(3) We propose a hybrid method which integrates the traditional graph ranking algorithm with deep learning technologies.
Current mainstream extractive text summarization techniques can be classified into unsupervised learning methods (feature scoring, graph ranking, etc.) and supervised learning methods (classification algorithms, deep learning, etc.).
Unsupervised learning methods aim to generate summarizations by ranking sentences. Most of early unsupervised methods extracted summarizations by analyzing features of original texts, which include word frequency, similarity of the first sentence to the title, and elements such as sentence length and sentence centrality. The work of Luh
With the continuous development of extractive summarization research, a variety of unsupervised extractive summarization methods have emerged. Typical unsupervised learning methods are graph ranking methods. Graph ranking-based methods, such as TextRan
Supervised learning methods treat text summarization as a classification problem. Traditional machine learning algorithms, which belong to supervised methods, are applied to extractive text summarization. Extractive summarizations can be converted into a traditional natural language processing problem of text classification. Classification-based methods use classification models such as support vector machine (SVM
In recent years supervised extractive summarization methods are more based on neural networks, where neural network models can learn vector representations of words and sentences to form abstract features, and then use these features for ranking. Yin et al
ETSDG is inspired from the work of Yin et al
The framework of the proposed ETSDG is shown in

Fig.1 Framework of the proposed ETSDG
(1) Pre-process raw documents to get sentences;
(2) Calculate sentence embeddings using deep learning;
(3) Calculate the pair-wise cosine similarities;
(4) Calculate the centrality for each sentence;
(5) Sort sentences so that the first selected sentence will be the one with the highest score.
In the first step, we pre-process raw documents. ETSDG uses sentences as the basic text unit for extraction and splits the raw document into sentences with the separator “.”, “?”and“!”, etc. However, sentence splitting is not a simple problem. The punctuation marks “?”and “!” can split sentences without having ambiguous meanings. But “.” has different uses and is not necessarily the end of a sentence, so we need to exclude some situations. For example, “Mr. and Mrs. Smith” went to the ball. In this sentence, “Mr. And Mrs. Smith” is a person phrase, where the “· ” is part of the person phrase, then it is not the end of the sentence. To solve this punctuation ambiguity, a simple rule of text break is used in ETSDG, namely “.” is not preceded by an abbreviation and is not followed by a number, then it indicates that the end of the sentence is detected.
There are dozens of proposed neural network-based methods to calculate sentence embeddings, such as Skip-Though
In order to reduce calculation time, memory usage and better express sentences, we propose a new model based on a pre-trained DistilRoBERTa network. A pooling operation is added in ETSDG to yield interpretable sentence embeddings.
ETSDG first uses a DistilRoBERTa network, which is a pre-trained Transformer network, and then fine-tune it to get sentence embeddings. The basic model Transformer replaces recurrent layers with self-attention in an encoder-endecoder framework and has achieved state-of-the-art results in language modeling. Based on Transformer, ETSDG aims to eliminate recurrence to allow more direct interaction among words in a sentence.
Moreover, a pooling operation is added to the output of DistilRoBERTa to derive sentence embeddings, which helps ETSDG reduce calculation time and memory usage.
Based on this pre-trained deep learning model, ETSDG can map sentences or paragraphs to a 768 dimensional dense vector space such that similarity sentences are close.

Fig.2 Architecture of calculating sentence embedding
After transforming sentences into sentence embeddings, we aim to consider similarities between sentences more effectively and exclude the influence of noisy sentences that have little to do with the center of the source text. Thus, we calculate similarities between pairwise sentences based on LexRan
As shown in

Fig.3 Schematic diagram of LexRank algorith
ETSDG uses cosine similarity to calculate the similarity of pair-wise sentences. The score of the pair-wise cosine similarity ranges from 0 to 1. According to scores of the pair-wise cosine similarities, a connection matrix between sentences is formed to iteratively calculate the amount of information contained in sentences.
The main goal of this calculation module is to find key sentences of documents by calculating the centrality for each sentence. When calculating the sentence centrality, we fully consider the number of connecting edges of nodes corresponding to each sentence and weights of the connecting edges, i.e., the size of the centrality and relevance of the sentence. Finally, according to a certain threshold, sentences with a higher score are selected as the key sentences of the document.
In order to better set the threshold in ETSDG, we conduct an experiment on a benchmark dataset “DUC2004 Task2”. In the process of fine-tuning, a pilot study is experimented to set the threshold ranging from 0 to 1.0. We set the final threshold according to the scores of commonly used text summarization automatic evaluation indicators ROUGE-1, ROUGE-2, and ROUGE-L. The experiment results are shown in

Fig.4 Testing results of ETSDG with different thresholds
As shown in
In the last calculation module, ETSDG sorts sentences considering the centrality of each sentence and edge weights in the text graph structure to generate summarizations. More representative sentences are positioned higher in the final generated summarization. The Equation we used in sentence sorting is described as follows
(1) |
where weight(u) represents the centrality of sentence u, and d the “damping factor”, which is usually set to 0.85. N represents the total number of sentences, adj[u] the set of sentences that are adjacent to u, and deg(v) the degree of the sentence v.
Based on
In this section, we introduce eight benchmark methods, the data set “DUC”, the evaluation method “ROUGE”, experiment results and analyses of the proposed ETSDG.
In order to evaluate the performance of ETSDG, we select eight commonly used benchmark methods (i.e., Lead Baseline, Random, GraphRank, FreqSum, TsSu
Since some of these eight methods use different metrics in their evaluation, in order to be able to reasonably compare with them, this paper adopts the results on the DUC2004 dataset (from Ying et al
Document understanding conference (DUC) is a commonly used dataset for measuring text summarization methods and an important evaluation criterion applicable to measuring extractive text summarization tasks. Therefore, in this paper, the DUC dataset is chosen as the main corpus for experiments.
Considering that DUC-2004 is the most commonly used comparison dataset in the DUC corpus, we conducted experiments with DUC-2004 to evaluate the performance of ETSDG. It contains 500 documents, each with four different manually generated reference summarizations.
Current evaluation methods are divided into automatic evaluation and manual evaluation methods according to whether there are manual participations involved. A widely metric used in automatic evaluation methods is ROUGE. Hence, in this paper, we choose ROUGE to evaluate the performance of ETSDG.
ROUGE is an automatic digest evaluation method proposed by Li
(2) |
where n-gram denotes the n‑word, {Ref} the reference summarization, and the number of n-gram appearing in both the system summarization and the reference summarization, and the number of n-gram appearing in the reference summarization.
In the evaluation phase, we used the toolkit pyrouge to calculate the ROUGE scores of ETSDG, using commonly used evaluation metrics ROUGE-1, ROUGE-2, ROUGE-L for comparison with other benchmark methods.
We compare ETSDG on DUC2004 using ROUGE evaluation method with eight benchmark methods.
Data from the second to the fourth column are extracted from Ref.[
As shown in
Lead Baseline is simple and outperforms Random in all of ROUGE measures, however, it only selects the leading sentences of a document to generate a summarization while ignoring connections between sentences. GraphRank takes connections between sentences into consideration and performances a little better than Lead Baseline in terms of ROUGE-1 metric. However, since GraphRank fails to better express sentences, it gets poor performance in terms of ROUGE-2 and ROUGE-L metrics. The above three traditional methods, Lead Baseline, Random and GraphRank, perform much worse than our proposed ETSDG due to their single feature considered.
Although the data of Centroid, FreqSum and TsSum methods are significantly better than the three traditional methods mentioned above for ROUGE-1 and ROUGE-2, they still inferior to ETSDG. The reason is that these three approaches fail to output sentences embeddings well and thus the overall performance is poor, while ETSDG does a better job in representing sentence embeddings.
BasicSum and TSSM slightly outperform ETSDG on ROUGE-2, but still underperform ETSDG on ROUGE-1. The reason for the gap in ETSDG on ROUGE-2 may lie in the fact that it only considers correlations between sentences, without further considering common words frequency features and named entities in sentences. These are also some of the directions we will consider to further optimize ETSDG in the future.
Compared with eight benchmark methods above, the overall performance of our proposed ETSDG is still good. It not only better expresses sentences using a pretrained model, but also considers connections between sentences using LexRank. Moreover, the threshold of “0.3” we set in ETSDG helps us both filter out some weakly similar sentences and avoid losing too many sentences related to the source text. These reasons explain why ETSDG can obtain better performances.
Experiment results show our method which is a fusion of deep learning and traditional LexRank algorithms can be used for text summarizations effectively.
The social value of automatic text summarization, which belongs to the domain of text generation in natural language processing, demonstrates the importance of automatic text summarization in the natural language domain. In this paper, a hybrid approach combining deep learning and graph ranking algorithms is introduced from the perspective of extractive methods. ETSDG firstly uses a pre-trained language model to represent sentence embeddings, secondly fully considers connections between sentences in documents based on the LexRank algorithm with a threshold, and finally finds more representative sentences to express the source text. According to the general text summarization evaluation method ROUGE, the results on “DUC2004” data set show that ETSDG has a better performance than certain benchmark methods.In the future, we hope to generate better summarizations using richer semantic information and leveraging external knowledge.
Contributions Statement
Ms. SHI Hui designed the study, complied the models and wrote the manuscript. Dr. WANG Tiexin reviewed the draft. All authors commented on the manuscript draft and approved the submission.
Conflict of Interest
The authors declare no competing interests.
References
LI Jinpeng, ZHANG Chuang,CHEN Xiaojun, et al. Survey on automatic text summarization[J]. Journal of Computer Research and Development,2021,58(1):1-21. [Baidu Scholar]
LUHN H P. The automatic creation of literature abstracts[J]. IBM Journal of Research and Development, 1958, 2(2): 159-165. [Baidu Scholar]
MIHALCEA R, TARAU P. Textrank: Bringing order into text[C]//Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing. Stroudsburg: ACL, 2004: 404-411. [Baidu Scholar]
PAGE L, BRIN S, MOTWANI R, et al. The PageRank citation ranking: Bringing order to the web:SIDL‑WP‑1999‑01‑20 [R].[S.l.]: Stanford InfoLab, 1999. [Baidu Scholar]
CORTES C, VAPNIK V. Support vector machine[J]. Machine learning, 1995, 20(3): 273-297. [Baidu Scholar]
ITTI L, BALDI P F. Bayesian surprise attracts human attention[J]. Advances in Neural Information Processing Systems, 2005,18: 547-554. [Baidu Scholar]
LOUIS A P. A bayesian method to incorporate background knowledge during automatic text summarization[C] //Proceedings of the 52nd Annual Meeting of the ACL. Stroudsburg: ACL, 2014: 333-338. [Baidu Scholar]
YIN W, PEI Y. Optimizing sentence modeling and selection for document summarization[C]// Proceedings of Int Joint Conference on Artificial Intelligence. Menlo Park: AAAI, 2015: 1383-1389. [Baidu Scholar]
CAO Z, LI W, LI S, et al. Improving multi-document summarization via text classification[C]//Proceedings of the AAAI Conference on Artificial Intelligence. Menlo Park: AAAI, 2017: 3053-3059. [Baidu Scholar]
ABDEL‑SALAM S, RAFEA A. Performance study on extractive text summarization using bert models[J]. Information, 2022,13(2): 67. [Baidu Scholar]
WANG Ershen, SONG Yuanshang, XU Song, et al. ADS-B anomaly data detection model based on deep learning and difference of Gaussian approach[J]. Transactions of Nanjing University of Aeronautics & Astronautics, 2020,37(4): 550-561. [Baidu Scholar]
YADAV A K, SINGH A, Dhiman M, et al Extractive text summarization using deep learning approach[J]. International Journal of Information Technology, 2022,14(5): 1‑9. [Baidu Scholar]
KIROS R, ZHU Y, SALAKHUTDINOV R R, et al. Skip-thought vectors[J]. Advances in Neural Information Processing Systems, 2015,28: 3294-3302. [Baidu Scholar]
CONNEAU A,KIELA D,SCHWENK H,et al.Supervised learning of universal sentence representations from natural language inference data[C]//Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing. Copenhagen, Denmark: Association for Computational Linguistics, 2017:670-680. [Baidu Scholar]
CER D, YANG Y, KONG S, et al. Universal sentence encoder for English[C]//Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing: System Demonstrations. Brussels, Belgium: Association for Computational Linguistics, 2018: 169-174. [Baidu Scholar]
REIMERS N, GUREVYCH I. Sentence-BERT: Sentence embeddings using siamese BERT-Networks[J]. arxiv Preprint arxiv: 1908.10084, 2019. [Baidu Scholar]
ERKAN G, RADEV D R. Lexrank: Graph-based lexical centrality as salience in text summarization[J]. Journal of Artificial Intelligence Research, 2004, 22: 457-479. [Baidu Scholar]
DING Xiaofei. Research on the generation of review summary based on text summarization technology [D]. Changsha: Hunan University, 2019. [Baidu Scholar]
HAN X, LV T, HU Z, et al. Text summarization using FrameNet-based semantic graph model[J]. Scientific Programming, 2016. DOI: 10.1155/2016/5130603. [Baidu Scholar]
HAN X, LV T, JIANG Q, et al. Text summarization using sentence-level semantic graph model[C]// Proceedings of 2016 4th International Conference on Cloud Computing and Intelligence Systems (CCIS). [S.l.]: IEEE, 2016: 171-176. [Baidu Scholar]
NENKOVA A,VANDERWENDE L, MCKEOWN K. A compositional context sensitive multi-document summarizer: Exploring the factors that influence summarization [C]// Proceedings of the 29th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval. [S.l.]: ACM, 2006: 573-580. [Baidu Scholar]
LIN C T, HOVY E. The automated acquisition of topic signatures for text summarization[J]. Proceedings of the 18th Conference on Computational Linguistics. Saarbrücken, Germany: Coling,2000: 495-501. [Baidu Scholar]
YIN Wenhao, LI Sujian, SUI Zhifang. A topic-sensitive extractive method for multi-document summarization [J]. Journal of Chinese Information Processing, 2017,31(6):155-161. [Baidu Scholar]
LIN C Y. Rouge: A package for automatic evaluation of summarizations[C]// Proceedings of the Workshop of ACL 2004. Stroudsburg: ACL, 2004: 74-81. [Baidu Scholar]