TOMDS (Topic-Oriented Multi-Document Summarization): Enabling Personalized Customization of Multi-Document Summaries
<p>The overall framework of TOMDS.</p> "> Figure 2
<p>TOMDS generates partial structure.</p> "> Figure 3
<p>Example of discourse segmentation and RST tree conversion. The original paragraph is segmented into 9 EDUs in box (<b>a</b>) and then parsed into an RST discourse tree in box (<b>b</b>). Elab, Interp, Conc, Seq, and Cond are relation labels (Elab = elaboration, Interp = interpretation, Conc = concession, Seq = sequence, and Cond = condition). The converted primary–secondary relationship-based RST discourse tree is shown in box (<b>c</b>). Nucleus nodes including ①, ③, ⑤, ⑦, and ⑨, and satellite nodes including ②, ④, ⑥, and ⑧, are denoted by solid lines and dashed lines.</p> "> Figure 4
<p>Discourse-aware attention (DaAtt) mechanism.</p> "> Figure 5
<p>Discourse graph attention (DGAtt) mechanism.</p> "> Figure 6
<p>Topic attention mechanism.</p> ">
Abstract
:1. Introduction
- (1)
- We incorporated and emphasized topic information in the extraction and generation phases of the model to further enhance the focus of multi-document summaries on specific topics.
- (2)
- We designed a discourse parser based on the rhetorical structure theory and analyzed it on the micro- and macro-levels to obtain the primary and secondary relationships between elementary discourse units (EDUs) within and between paragraphs.
- (3)
- We extended the transformer model and added a discourse-aware attention mechanism in the encoder part to combine the intra-paragraph and inter-paragraph relationships with the implicit relationships of the source documents to generate richer semantic features in a complementary manner.
- (4)
- The experimental results of the WikiSum dataset demonstrate that our model achieved advanced performance in recall-oriented understudy for gisting evaluation (ROUGE) scores and human evaluations. While improving the subject focus, it also received high ratings in terms of grammatical accuracy.
2. Literature Review
2.1. Multi-Document Summarization
2.2. Discourse Structure
3. Topic-Oriented Multi-Document Summarization (TOMDS) Model
3.1. Overview
3.2. Extractive Stage
3.3. Abstractive Stage
3.3.1. Embedding
3.3.2. Primary–Secondary Relationship Matrix
3.3.3. Local Transformer Layer
3.3.4. Global Transformer Layer
- Multi-head Pooling
- Discourse Graph Attention
3.3.5. Topic Attention Mechanism
4. Experiments
4.1. Experimental Setup
4.1.1. Dataset
4.1.2. Evaluation Method
4.1.3. Training Configure
4.1.4. Model Comparison
4.2. Experimental Results
4.3. Human Evaluation
4.4. Grammar Check
5. Conclusions and Future Work
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- IDC: Expect 175 Zettabytes of Data Worldwide by 2025. Networkworld. 2018. Available online: https://www.networkworld.com/article/3325397/idc-expect-175-zettabytes-of-data-worldwide-by-2025.html (accessed on 20 October 2023).
- Aggarwal, C.C. Neural Networks and Deep Learning; Springer: Cham, Switzerland, 2018; p. 3. [Google Scholar]
- Zopf, M. Auto-hmds: Automatic construction of a large heterogeneous multilingual multi-document summarization corpus. In Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018), Miyazaki, Japan, 1 May 2018; Available online: https://aclanthology.org/L18-1510 (accessed on 20 October 2023).
- Liu, P.J.; Saleh, M.; Pot, E.; Goodrich, B.; Sepassi, R.; Kaiser, L.; Shazeer, N. Generating wikipedia by summarizing long sequences. In Proceedings of the 6th International Conference on Learning Representations (ICLR 2018), Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
- Fabbri, A.R.; Li, I.; She, T.; Li, S.; Radev, D.R. Multi-news: A large-scale multi-document summarization dataset and abstractive hierarchical model. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 4 June 2019. [Google Scholar] [CrossRef]
- Radev, D.R. A common theory of information fusion from multiple text sources step one: Cross-document structure. In Proceedings of the 1st SIGdial Workshop on Discourse and Dialogue, Hong Kong, China, 7–8 October 2000; pp. 74–83. [Google Scholar]
- Antognini, D.; Faltings, B. Learning to create sentence semantic relation graphs for multi-document summarization. In Proceedings of the 2nd Workshop on New Frontiers in Summarization, arXiv: Computation and Language, Association for Computational Linguistics, Hong Kong, China, 20 September 2019. [Google Scholar] [CrossRef]
- Yin, Y.; Song, L.; Su, J.; Zeng, J.; Zhou, C.; Luo, J. Graph-based neural sentence ordering. In Proceedings of the 28th International Joint Conference on Artificial Intelligence, Macao, China, 16 December 2019; pp. 5387–5393. [Google Scholar] [CrossRef]
- Li, W.; Xiao, X.; Liu, J.; Wu, H.; Wang, H.; Du, J. Leveraging Graph to Improve Abstractive Multi-Document Summarization. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 20 May 2020. [Google Scholar] [CrossRef]
- Russell, S.J.; Norvig, P. Artificial Intelligence a Modern Approach; Prentice Hall: London, UK, 2010. [Google Scholar]
- Erkan, G.; Radev, D.R. Lexrank: Graph-based lexical centrality as salience in text summarization. J. Artif. Intell. Res. 2004, 22, 457–479. [Google Scholar] [CrossRef]
- Mihalcea, R.; Tarau, P. TextRank: Bringing order into text. In Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing, Barcelona, Spain, 25–26 July 2004; ACL: Stroudsburg, PA, USA, 2004; pp. 404–411. [Google Scholar]
- Radev, D.R.; Jing, H.; Styś, M.; Tam, D. Centroid-based summarization of multiple documents. Inf. Process. Manag. 2004, 40, 919–938. [Google Scholar] [CrossRef]
- Liu, N.; Peng, X.; Lu, Y.; Tang, X.J.; Wang, H.W. A Topic Approach to Sentence Ordering for Multi-document Summarization. In Proceedings of the 2016 IEEE Trustcom/BigDataSE/ISPA, Tianjin, China, 23–26 August 2016; pp. 1390–1395. [Google Scholar] [CrossRef]
- Baralis, E.; Cagliero, L.; Jabeen, S.; Fiori, A. Multi-document summarization exploiting frequent itemsets. In Proceedings of the 27th Annual ACM Symposium on Applied Computing, Trento, Italy, 26–30 March 2012; pp. 782–786. [Google Scholar]
- Goldstein, J.; Mittal, V.O.; Carbonell, J.G.; Kantrowitz, M. Multi-document summarization by sentence extraction. In Proceedings of the NAACL-ANLP 2000 Workshop: Automatic Summarization, Seattle, WA, USA, 30 April 2000. [Google Scholar]
- Wan, X.; Yang, J. Multi-document summarization using cluster-based link analysis. In Proceedings of the 31st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Singapore, 20–24 July 2008; pp. 299–306. [Google Scholar]
- Haghigh, A.; Vanderwende, L. Exploring content models for multi-document summarization. In Proceedings of the Human Language Technologies: The 2009 Annual Conference of the North American Chapter of the Association for Computational Linguistics, New York, NY, USA, 4–9 June 2009; pp. 362–370. [Google Scholar]
- Arora, R.; Ravindran, B. Latent dirichlet allocation and singular value decomposition based multi-document summarization. In Proceedings of the 2008 Eighth IEEE International Conference on Data Mining, Pisa, Italy, 15–19 December 2008; pp. 713–718. [Google Scholar] [CrossRef]
- Jin, H.; Wang, T.; Wan, X. Multi-granularity interaction network for extractive and abstractive multi-document summarization. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 6244–6254. [Google Scholar]
- Barzilay, R.; McKeown, K.R. Sentence fusion for multidocument news summarization. Comput. Linguist. 2005, 31, 297–327. [Google Scholar] [CrossRef]
- Tan, J.; Wan, X.; Xiao, J. Abstractive Document Summarization with a Graph-Based Attentional Neural Model. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, Vancouver, BC, Canada, 30 July–4 August 2017; pp. 1171–1181. [Google Scholar] [CrossRef]
- Liu, Y.; Lapata, M. Hierarchical Transformers for Multi-Document Summarization. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 30 May 2019. [Google Scholar] [CrossRef]
- Ma, C.B.; Zhang, W.E.; Wang, H.; Gupta, S.; Guo, M. Incorporating linguistic knowledge for abstractive multi-document summarization. In Proceedings of the 36th Pacific Asia Conference on Language, Information and Computation, Manila, Philippines, 20–22 October 2022; pp. 147–156. [Google Scholar]
- Wan, X. Subtopic-based multi-modality ranking for topic-focused multi-document summarization. Comput. Intell. 2013, 29, 627–648. [Google Scholar] [CrossRef]
- Li, J.; Li, S. A novel feature-based Bayesian model for query focused multi-document summarization. Trans. Assoc. Comput. Linguist. 2013, 1, 89–98. [Google Scholar] [CrossRef]
- Lee, S.; Belkasim, S.; Zhang, Y. Multi-document text summarization using topic model and fuzzy logic. In Proceedings of the 9th International Conference on Machine Learning and Data Mining in Pattern Recognition (MLDM 13), New York, NY, USA, 23–25 July 2013; pp. 159–168. [Google Scholar]
- Liu, N.; Tang, X.J.; Lu, Y.; Li, M.X.; Wang, H.W.; Xiao, P. Topic-sensitive multi-document summarization algorithm. In Proceedings of the 2014 Sixth International Symposium on Parallel Architectures, Algorithms and Programming, Beijing, China, 13–15 July 2014; pp. 69–74. [Google Scholar]
- Cui, P.; Hu, L. Topic-Guided Abstractive Multi-Document Summarization. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Punta Cana, Dominican Republic, 7–11 November 2021. [Google Scholar]
- Chomsky, N. New Horizons in the Study of Language and Mind; Cambridge University Press: Cambridge, UK, 2000. [Google Scholar]
- Chomsky, N. Language and Mind; Cambridge University Press: Cambridge, UK, 2006. [Google Scholar]
- Xu, J. Text Linguistics in Contemporary Chinese; The Commercial Press: Beijing, China, 2010; pp. 2–4. [Google Scholar]
- Mann, W.C.; Thompson, S.A. Relational propositions in discourse. Discourse Process. 1986, 9, 57–90. [Google Scholar] [CrossRef]
- Van Dijk, T.A. Macrostructures: An Interdisciplinary Study of Global Structures in Discourse, Interaction, and Cognition; Lawrence Erlbaum Associates, Inc.: Hillsdale, NJ, USA, 2019; pp. 67–93. [Google Scholar]
- Zhang, Y.; Wang, Z.; Wang, H. Single Document Extractive Summarization with Satellite and Nuclear Relations. J. Chin. Inf. Process. 2019, 33, 67–76. [Google Scholar]
- Liu, K.; Wang, H. Research Automatics Summarization Coherence Based on Discourse Rhetoric Structure. J. Chin. Inf. Process. 2019, 33, 77–84. [Google Scholar]
- Xu, J.C.; Gan, Z.; Cheng, Y.; Liu, J. Discourse-Aware Neural Extractive Text Summarization. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 5021–5031. [Google Scholar] [CrossRef]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Aidan, N.G.; Kaiser, L.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
- Zhang, X.; Wei, Q.; Song, Q.; Zhang, P. An Extractive Text Summarization Model Based on Rhetorical Structure Theory. In Proceedings of the 26th ACIS International Winter Conference on Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing (SNPD-Winter), Taiyuan, China, 6–8 July 2023. [Google Scholar]
- Kobayashi, N.; Hirao, T.; Kamigaito, H.; Okumura, M.; Nagata, M. Top-down RST parsing utilizing granularity levels in documents. AAAI Conf. Artif. Intell. 2020, 34, 8099–8106. [Google Scholar] [CrossRef]
- Lin, C.Y. ROUGE: A Package for Automatic Evaluation of Summaries. In Proceedings of the Workshop on Text Summarization Branches Out, Barcelona, Spain, 25–26 July 2004; pp. 74–81. Available online: http://anthology.aclweb.org/W/W04/W04-1013.pdf (accessed on 25 November 2023).
- Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the 3rd International Conference for Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar] [CrossRef]
- Baralis, E.; Cagliero, L.; Mahoto, N.; Fiori, A. GraphSum: Discovering correlations among multiple terms for graph-based summarization. Inf. Sci. 2013, 249, 96–109. [Google Scholar] [CrossRef]
- Clarke, J.; Lapata, M. Discourse constraints for document compression. Comput. Linguist. 2010, 36, 411–441. [Google Scholar] [CrossRef]
- Narayan, S.; Shay, B.C.; Lapata, M. Ranking sentences for extractive summarization with reinforcement learning. In Proceedings of the NAACL 2018, New Orleans, USA, 1–6 June 2018. [Google Scholar] [CrossRef]
- Xu, J.; Durrett, G. Neural Extractive Text Summarization with Syntactic Compression. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019. [Google Scholar] [CrossRef]
Dataset | Size |
---|---|
Training set | 1,490,138 |
Validation set | 82,785 |
Test set | 82,520 |
Model | ROUGE-1 | ROUGE-2 | ROUGE-L |
---|---|---|---|
Extractive methods | |||
Lead | 38.22 | 16.85 | 26.89 |
LexRank | 36.12 | 11.67 | 22.52 |
GraphSum | 42.63 | 27.70 | 36.97 |
Abstractive models | |||
T-DMCA | 40.77 | 25.60 | 34.90 |
FT | 39.55 | 24.63 | 33.99 |
HT | 40.82 | 25.99 | 35.08 |
Topic-based models | |||
TG-MultiSum | 42.34 | 26.91 | 35.85 |
S-sLDA | 37.28 | 15.97 | 26.31 |
HierSum | 36.27 | 12.84 | 25.79 |
Ours | |||
Our Model | 38.57 | 18.93 | 28.64 |
OurModel + DisParser | 41.28 | 25.61 | 33.79 |
OurModel + TA | 39.35 | 20.98 | 30.12 |
OurModel + DisParser + TA | 43.23 | 28.27 | 37.86 |
Model | QA | Rating |
---|---|---|
Lead | 32.50 | −0.323 |
HT | 55.05 | 0.243 |
TG-MultiSum | 56.36 | 0.254 |
OurModel + TA + Datt | 57.21 | 0.251 |
Model | CR | PV | PT | O |
---|---|---|---|---|
T-DMCA | 17.6 | 2.8 | 2.3 | 3.0 |
FT | 18.0 | 2.8 | 2.4 | 3.0 |
HT | 18.1 | 2.9 | 2.4 | 2.8 |
OurModel + TA + Datt | 18.3 | 3.0 | 2.4 | 3.1 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhang, X.; Wei, Q.; Song, Q.; Zhang, P. TOMDS (Topic-Oriented Multi-Document Summarization): Enabling Personalized Customization of Multi-Document Summaries. Appl. Sci. 2024, 14, 1880. https://doi.org/10.3390/app14051880
Zhang X, Wei Q, Song Q, Zhang P. TOMDS (Topic-Oriented Multi-Document Summarization): Enabling Personalized Customization of Multi-Document Summaries. Applied Sciences. 2024; 14(5):1880. https://doi.org/10.3390/app14051880
Chicago/Turabian StyleZhang, Xin, Qiyi Wei, Qing Song, and Pengzhou Zhang. 2024. "TOMDS (Topic-Oriented Multi-Document Summarization): Enabling Personalized Customization of Multi-Document Summaries" Applied Sciences 14, no. 5: 1880. https://doi.org/10.3390/app14051880
APA StyleZhang, X., Wei, Q., Song, Q., & Zhang, P. (2024). TOMDS (Topic-Oriented Multi-Document Summarization): Enabling Personalized Customization of Multi-Document Summaries. Applied Sciences, 14(5), 1880. https://doi.org/10.3390/app14051880