Nothing Special   »   [go: up one dir, main page]

TANet: Thread-Aware Pretraining for Abstractive Conversational Summarization

Ze Yang, Christian Wang, Zhoujin Tian, Wei Wu, Zhoujun Li


Abstract
Although pre-trained language models (PLMs) have achieved great success and become a milestone in NLP, abstractive conversational summarization remains a challenging but less studied task. The difficulty lies in two aspects. One is the lack of large-scale conversational summary data. Another is that applying the existing pre-trained models to this task is tricky because of the structural dependence within the conversation and its informal expression, etc. In this work, we first build a large-scale (11M) pretraining dataset called RCSum, based on the multi-person discussions in the Reddit community. We then present TANet, a thread-aware Transformer-based network. Unlike the existing pre-trained models that treat a conversation as a sequence of sentences, we argue that the inherent contextual dependency among the utterances plays an essential role in understanding the entire conversation and thus propose two new techniques to incorporate the structural information into our model. The first is thread-aware attention which is computed by taking into account the contextual dependency within utterances. Second, we apply thread prediction loss to predict the relations between utterances. We evaluate our model on four datasets of real conversations, covering types of meeting transcripts, customer-service records, and forum threads. Experimental results demonstrate that TANet achieves a new state-of-the-art in terms of both automatic evaluation and human judgment.
Anthology ID:
2022.findings-naacl.198
Volume:
Findings of the Association for Computational Linguistics: NAACL 2022
Month:
July
Year:
2022
Address:
Seattle, United States
Editors:
Marine Carpuat, Marie-Catherine de Marneffe, Ivan Vladimir Meza Ruiz
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
2594–2607
Language:
URL:
https://aclanthology.org/2022.findings-naacl.198
DOI:
10.18653/v1/2022.findings-naacl.198
Bibkey:
Cite (ACL):
Ze Yang, Christian Wang, Zhoujin Tian, Wei Wu, and Zhoujun Li. 2022. TANet: Thread-Aware Pretraining for Abstractive Conversational Summarization. In Findings of the Association for Computational Linguistics: NAACL 2022, pages 2594–2607, Seattle, United States. Association for Computational Linguistics.
Cite (Informal):
TANet: Thread-Aware Pretraining for Abstractive Conversational Summarization (Yang et al., Findings 2022)
Copy Citation:
PDF:
https://aclanthology.org/2022.findings-naacl.198.pdf
Video:
 https://aclanthology.org/2022.findings-naacl.198.mp4