Abstract
Pre-trained conversation models (PCMs) have achieved promising progress in recent years. However, existing PCMs for Task-oriented dialog (TOD) are insufficient for capturing the sequential nature of the TOD-related tasks, as well as for learning dialog policy information. To alleviate these problems, this paper proposes a task-progressive PCM with two policy-aware pre-training tasks. The model is pre-trained through three stages where TOD-related tasks are progressively employed according to the task logic of the TOD system. A global policy consistency task is designed to capture the multi-turn dialog policy sequential relation, and an act-based contrastive learning task is designed to capture similarities among samples with the same dialog policy. Our model achieves better results on both MultiWOZ and In-Car end-to-end dialog modeling benchmarks with only 18% parameters and 25% pre-training data compared to the previous state-of-the-art PCM, GALAXY. We make our code and data publicly available (https://github.com/lucenzhong/TPLD).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Tian, X., Huang, L., Lin, Y., et al.: Amendable generation for dialogue state tracking. In: Proceedings of the 3rd Workshop on Natural Language Processing for Conversational AI, pp. 80–92 (2021)
Takanobu, R., Liang, R., Huang, M.: Multi-agent task-oriented dialog policy learning with role-aware reward decomposition. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 625–638 (2020)
Peng, B., Zhu, C., Li, C., et al.: Few-shot natural language generation for task-oriented dialog. In: Findings of the Association for Computational Linguistics: EMNLP 2020, pp. 172–182 (2020)
Madotto, A., Wu, C.S., Fung, P.: Mem2Seq: effectively incorporating knowledge bases into end-to-end task-oriented dialog systems. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 1468–1478 (2018)
Lei, W., Jin, X., Kan, M.Y., et al.: Sequicity: simplifying task-oriented dialogue systems with single sequence-to-sequence architectures. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 1437–1447 (2018)
Yang, Y., Li, Y., Quan, X.: UBAR: towards fully end-to-end task-oriented dialog system with GPT-2. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, no. 16, pp. 14230–14238 (2021)
Sun, H., Bao, J., Wu, Y., et al.: Mars: semantic-aware contrastive learning for end-to-end task-oriented dialog. arXiv preprint arXiv:2210.08917 (2022)
Wu, C.S., Hoi, S.C., Socher, R., et al.: TOD-BERT: pre-trained natural language understanding for task-oriented dialogue. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 917–929 (2020)
Peng, B., Li, C., Li, J., et al.: Soloist: building task bots at scale with transfer learning and machine teaching. Trans. Assoc. Comput. Linguist. 9, 807–824 (2021)
Su, Y., Shu, L., Mansimov, E., et al.: Multi-task pre-training for plug-and-play task-oriented dialogue system. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 4661–4676 (2022)
He, W., Dai, Y., Zheng, Y., et al.: Galaxy: a generative pre-trained model for task-oriented dialog with semi-supervised learning and explicit policy injection. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 36, no. 10, pp. 10749–10757 (2022)
He, W., Dai, Y., Hui, B., et al.: SPACE-2: tree-structured semi-supervised contrastive pre-training for task-oriented dialog understanding. In: Proceedings of the 29th International Conference on Computational Linguistics, pp. 553–569 (2022)
He, W., Dai, Y., Yang, M., et al.: Unified dialog model pre-training for task-oriented dialog understanding and generation. In: Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 187–200 (2022)
Raffel, C., Shazeer, N., Roberts, A., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21(1), 5485–5551 (2020)
Budzianowski, P., Wen, T.H., Tseng, B.H., et al.: MultiWOZ-a large-scale multi-domain Wizard-of-Oz dataset for task-oriented dialogue modelling. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 5016–5026 (2018)
Eric, M., Krishnan, L., Charette, F., et al.: Key-value retrieval networks for task-oriented dialogue. In: Proceedings of the 18th Annual SIGdial Meeting on Discourse and Dialogue, pp. 37–49 (2017)
Li, X., Wang, Y., Sun, S., et al.: Microsoft dialogue challenge: building end-to-end task-completion dialogue systems. arXiv preprint arXiv:1807.11125 (2018)
El Asri, L., Schulz, H., Sarma, S.K., et al.: Frames: a corpus for adding memory to goal-oriented dialogue systems. In: Proceedings of the 18th Annual SIGdial Meeting on Discourse and Dialogue, pp. 207–219 (2017)
Wen, T.H., Vandyke, D., Mrkšić, N., et al.: A network-based end-to-end trainable task-oriented dialogue system. In: Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 1, Long Papers, pp. 438–449 (2017)
Eric, M., Goel, R., Paul, S., Sethi, et al.: Multiwoz 2.1: multi-domain dialogue state corrections and state tracking baselines. arXiv preprint arXiv:1907.01669 (2019)
Radford, A., Wu, J., Child, R., et al.: Language models are unsupervised multitask learners. OpenAI Blog 1(8), 9 (2019)
Henderson, M., Casanueva, I., Mrkšić, N., et al.: ConveRT: efficient and accurate conversational representations from transformers. In: Findings of the Association for Computational Linguistics: EMNLP 2020, pp. 2161–2174 (2020)
Adiwardana, D., Luong, M.T., So, D.R., et al.: Towards a human-like open-domain chatbot. arXiv preprint arXiv:2001.09977 (2020)
Chen, Z., Liu, Y., Chen, L., et al.: OPAL: ontology-aware pretrained language model for end-to-end task-oriented dialogue. arXiv preprint arXiv:2209.04595 (2022)
Papineni, K., Roukos, S., Ward, T., et al.: Bleu: a method for automatic evaluation of machine translation. In: Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, pp. 311–318 (2002)
Mehri, S., Srinivasan, T., Eskenazi, M.: Structured fusion networks for dialog. In: Proceedings of the 20th Annual SIGdial Meeting on Discourse and Dialogue, pp. 165–177 (2019)
Nekvinda, T., Dušek, O.: Shades of BLEU, flavours of success: the case of MultiWOZ. In: Proceedings of the 1st Workshop on Natural Language Generation, Evaluation, and Metrics (GEM 2021), pp. 34–46 (2021)
Wolf, T., Debut, L., Sanh, V., et al.: Huggingface’s transformers: state-of-the-art natural language processing. arXiv preprint arXiv:1910.03771 (2019)
Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)
Acknowledgements
We are grateful to the anonymous reviewers for their insightful comments and suggestions.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Zhong, L. et al. (2023). A Task-Oriented Dialog Model with Task-Progressive and Policy-Aware Pre-training. In: Liu, F., Duan, N., Xu, Q., Hong, Y. (eds) Natural Language Processing and Chinese Computing. NLPCC 2023. Lecture Notes in Computer Science(), vol 14302. Springer, Cham. https://doi.org/10.1007/978-3-031-44693-1_1
Download citation
DOI: https://doi.org/10.1007/978-3-031-44693-1_1
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-44692-4
Online ISBN: 978-3-031-44693-1
eBook Packages: Computer ScienceComputer Science (R0)