Nothing Special   »   [go: up one dir, main page]

Skip to main content

A Task-Oriented Dialog Model with Task-Progressive and Policy-Aware Pre-training

  • Conference paper
  • First Online:
Natural Language Processing and Chinese Computing (NLPCC 2023)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 14302))

  • 2209 Accesses

Abstract

Pre-trained conversation models (PCMs) have achieved promising progress in recent years. However, existing PCMs for Task-oriented dialog (TOD) are insufficient for capturing the sequential nature of the TOD-related tasks, as well as for learning dialog policy information. To alleviate these problems, this paper proposes a task-progressive PCM with two policy-aware pre-training tasks. The model is pre-trained through three stages where TOD-related tasks are progressively employed according to the task logic of the TOD system. A global policy consistency task is designed to capture the multi-turn dialog policy sequential relation, and an act-based contrastive learning task is designed to capture similarities among samples with the same dialog policy. Our model achieves better results on both MultiWOZ and In-Car end-to-end dialog modeling benchmarks with only 18% parameters and 25% pre-training data compared to the previous state-of-the-art PCM, GALAXY. We make our code and data publicly available (https://github.com/lucenzhong/TPLD).

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 99.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 129.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Tian, X., Huang, L., Lin, Y., et al.: Amendable generation for dialogue state tracking. In: Proceedings of the 3rd Workshop on Natural Language Processing for Conversational AI, pp. 80–92 (2021)

    Google Scholar 

  2. Takanobu, R., Liang, R., Huang, M.: Multi-agent task-oriented dialog policy learning with role-aware reward decomposition. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 625–638 (2020)

    Google Scholar 

  3. Peng, B., Zhu, C., Li, C., et al.: Few-shot natural language generation for task-oriented dialog. In: Findings of the Association for Computational Linguistics: EMNLP 2020, pp. 172–182 (2020)

    Google Scholar 

  4. Madotto, A., Wu, C.S., Fung, P.: Mem2Seq: effectively incorporating knowledge bases into end-to-end task-oriented dialog systems. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 1468–1478 (2018)

    Google Scholar 

  5. Lei, W., Jin, X., Kan, M.Y., et al.: Sequicity: simplifying task-oriented dialogue systems with single sequence-to-sequence architectures. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 1437–1447 (2018)

    Google Scholar 

  6. Yang, Y., Li, Y., Quan, X.: UBAR: towards fully end-to-end task-oriented dialog system with GPT-2. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, no. 16, pp. 14230–14238 (2021)

    Google Scholar 

  7. Sun, H., Bao, J., Wu, Y., et al.: Mars: semantic-aware contrastive learning for end-to-end task-oriented dialog. arXiv preprint arXiv:2210.08917 (2022)

  8. Wu, C.S., Hoi, S.C., Socher, R., et al.: TOD-BERT: pre-trained natural language understanding for task-oriented dialogue. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 917–929 (2020)

    Google Scholar 

  9. Peng, B., Li, C., Li, J., et al.: Soloist: building task bots at scale with transfer learning and machine teaching. Trans. Assoc. Comput. Linguist. 9, 807–824 (2021)

    Article  Google Scholar 

  10. Su, Y., Shu, L., Mansimov, E., et al.: Multi-task pre-training for plug-and-play task-oriented dialogue system. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 4661–4676 (2022)

    Google Scholar 

  11. He, W., Dai, Y., Zheng, Y., et al.: Galaxy: a generative pre-trained model for task-oriented dialog with semi-supervised learning and explicit policy injection. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 36, no. 10, pp. 10749–10757 (2022)

    Google Scholar 

  12. He, W., Dai, Y., Hui, B., et al.: SPACE-2: tree-structured semi-supervised contrastive pre-training for task-oriented dialog understanding. In: Proceedings of the 29th International Conference on Computational Linguistics, pp. 553–569 (2022)

    Google Scholar 

  13. He, W., Dai, Y., Yang, M., et al.: Unified dialog model pre-training for task-oriented dialog understanding and generation. In: Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 187–200 (2022)

    Google Scholar 

  14. Raffel, C., Shazeer, N., Roberts, A., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21(1), 5485–5551 (2020)

    MathSciNet  Google Scholar 

  15. Budzianowski, P., Wen, T.H., Tseng, B.H., et al.: MultiWOZ-a large-scale multi-domain Wizard-of-Oz dataset for task-oriented dialogue modelling. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 5016–5026 (2018)

    Google Scholar 

  16. Eric, M., Krishnan, L., Charette, F., et al.: Key-value retrieval networks for task-oriented dialogue. In: Proceedings of the 18th Annual SIGdial Meeting on Discourse and Dialogue, pp. 37–49 (2017)

    Google Scholar 

  17. Li, X., Wang, Y., Sun, S., et al.: Microsoft dialogue challenge: building end-to-end task-completion dialogue systems. arXiv preprint arXiv:1807.11125 (2018)

  18. El Asri, L., Schulz, H., Sarma, S.K., et al.: Frames: a corpus for adding memory to goal-oriented dialogue systems. In: Proceedings of the 18th Annual SIGdial Meeting on Discourse and Dialogue, pp. 207–219 (2017)

    Google Scholar 

  19. Wen, T.H., Vandyke, D., Mrkšić, N., et al.: A network-based end-to-end trainable task-oriented dialogue system. In: Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 1, Long Papers, pp. 438–449 (2017)

    Google Scholar 

  20. Eric, M., Goel, R., Paul, S., Sethi, et al.: Multiwoz 2.1: multi-domain dialogue state corrections and state tracking baselines. arXiv preprint arXiv:1907.01669 (2019)

  21. Radford, A., Wu, J., Child, R., et al.: Language models are unsupervised multitask learners. OpenAI Blog 1(8), 9 (2019)

    Google Scholar 

  22. Henderson, M., Casanueva, I., Mrkšić, N., et al.: ConveRT: efficient and accurate conversational representations from transformers. In: Findings of the Association for Computational Linguistics: EMNLP 2020, pp. 2161–2174 (2020)

    Google Scholar 

  23. Adiwardana, D., Luong, M.T., So, D.R., et al.: Towards a human-like open-domain chatbot. arXiv preprint arXiv:2001.09977 (2020)

  24. Chen, Z., Liu, Y., Chen, L., et al.: OPAL: ontology-aware pretrained language model for end-to-end task-oriented dialogue. arXiv preprint arXiv:2209.04595 (2022)

  25. Papineni, K., Roukos, S., Ward, T., et al.: Bleu: a method for automatic evaluation of machine translation. In: Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, pp. 311–318 (2002)

    Google Scholar 

  26. Mehri, S., Srinivasan, T., Eskenazi, M.: Structured fusion networks for dialog. In: Proceedings of the 20th Annual SIGdial Meeting on Discourse and Dialogue, pp. 165–177 (2019)

    Google Scholar 

  27. Nekvinda, T., Dušek, O.: Shades of BLEU, flavours of success: the case of MultiWOZ. In: Proceedings of the 1st Workshop on Natural Language Generation, Evaluation, and Metrics (GEM 2021), pp. 34–46 (2021)

    Google Scholar 

  28. Wolf, T., Debut, L., Sanh, V., et al.: Huggingface’s transformers: state-of-the-art natural language processing. arXiv preprint arXiv:1910.03771 (2019)

  29. Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)

Download references

Acknowledgements

We are grateful to the anonymous reviewers for their insightful comments and suggestions.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Xiaojie Wang .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Zhong, L. et al. (2023). A Task-Oriented Dialog Model with Task-Progressive and Policy-Aware Pre-training. In: Liu, F., Duan, N., Xu, Q., Hong, Y. (eds) Natural Language Processing and Chinese Computing. NLPCC 2023. Lecture Notes in Computer Science(), vol 14302. Springer, Cham. https://doi.org/10.1007/978-3-031-44693-1_1

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-44693-1_1

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-44692-4

  • Online ISBN: 978-3-031-44693-1

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics