skip to main content
10.1145/3483207.3483231acmotherconferencesArticle/Chapter ViewAbstractPublication PagesspmlConference Proceedingsconference-collections
research-article

Learning Dense Entity-Aware Dialogue Intentions with Rewritten Utterance for External Knowledge Documents Retrieval

Published:28 October 2021Publication History

ABSTRACT

External knowledge-enhanced task-oriented dialogue systems aim to cover user requests beyond pre-defined DBs/APIs. Recently, existing dialogue systems have focused more on retrieving external knowledge sources relevant to dialogue contexts, achieving competitive results. However, due to the lack of modeling entity-aware dialogue intention, such dialogue systems are hard to accurately and efficiently link the out-of-API functions in real-world scenarios. To tackle this problem, this paper investigates learning dense entity-aware dialogue intentions for external knowledge documents retrieval in task-oriented dialogues. To this end, we propose an intention-guided two-stage training approach that includes intention-guided training and knowledge transfer stages. This approach, which leverages rewritten utterances that explicitly convey entity-aware user intentions, can improve the performance of existing Bi-Encoder retrievers such as DPR (Deep Passage Retriever). In intention-guided training stage, a posterior history encoder is initialized and guided by inputting rewritten utterances for learning discriminative dense representations. In knowledge transfer stage, these representations are transferred to a newly initialized prior encoder for inference via an extra intent consistency loss. In addition, negative sampling in test knowledge documents is used to learn more discriminative dense representations of the unseen domain. The advantages of our approach are no need for response annotations and extra response generator, additionally, it provides great scalability. The experimental results on augmented MultiWOZ 2.1 dataset show that our approach outperforms baseline models except for relevance classifiers in retrieval accuracy and has reasonably high efficiency.

References

  1. Seokhwan Kim, Mihail Eric, Karthik Gopalakrishnan, Behnam Hedayatnia, Yang Liu, and Dilek Hakkani-Tür. 2020. Beyond Domain APIs: Task-oriented conversational modeling with unstructured knowledge access. In SIGdial. 278–289.Google ScholarGoogle Scholar
  2. Mihail Eric, Rahul Goel, Shachi Paul, Abhishek Sethi, Sanchit Agarwal, Shuyang Gao, Adarsh Kumar, Anuj Goyal, Peter Ku, and Dilek Hakkani-Tur. 2020. MultiWOZ 2.1: A Consolidated Multi-Domain Dialogue Dataset with State Corrections and State Tracking Baselines. In Proceedings of the 12th Language Resources and Evaluation Conference. European Language Resources Association, Marseille, France, 422–428. https://www.aclweb.org/anthology/2020.lrec-1.53Google ScholarGoogle Scholar
  3. H. He, Hua Lu, Siqi Bao, Fan Wang, Hua Wu, Zheng-Yu Niu, and H. Wang. 2021. Learning to Select External Knowledge with Multi-Scale Negative Sampling. ArXiv abs/2102.02096(2021).Google ScholarGoogle Scholar
  4. Chaohong Tan, Xiaoyu Yang, Zi’ou Zheng, Tianda Li, Yufei Feng, Jia-Chen Gu, QUAN LIU, Dan Liu, Zhenhua Ling, and Xiao-Dan Zhu. 2020. Learning to Retrieve Entity-Aware Knowledge and Generate Responses with Copy Mechanism for Task-Oriented Dialogue Systems. ArXiv abs/2012.11937(2020).Google ScholarGoogle Scholar
  5. David Thulke, Nico Daheim, Christian Dugast, and H. Ney. 2021. Efficient Retrieval Augmented Generation from Unstructured Knowledge for Task-Oriented Dialog. ArXiv abs/2102.04643(2021).Google ScholarGoogle Scholar
  6. Patrick S. H. Lewis, Ethan Perez, Aleksandra Piktus, Fabio Petroni, Vladimir Karpukhin, Naman Goyal, Heinrich Küttler, Mike Lewis, Wen tau Yih, Tim Rocktäschel, Sebastian Riedel, and Douwe Kiela. 2020. Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks. In Advances in Neural Information Processing Systems, Vol. 33. 9459–9474.Google ScholarGoogle Scholar
  7. Daniel Gillick, Sayali Kulkarni, Larry Lansing, Alessandro Presta, Jason Baldridge, Eugene Ie, and Diego Garcia-Olano. 2019. Learning Dense Representations for Entity Retrieval. In Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL). 528–537.Google ScholarGoogle ScholarCross RefCross Ref
  8. Tiancheng Zhao and Maxine Eskenazi. 2018. Zero-Shot Dialog Generation with Cross-Domain Latent Actions. In Proceedings of the 19th Annual SIGdial Meeting on Discourse and Dialogue. 1–10.Google ScholarGoogle ScholarCross RefCross Ref
  9. Hang Liu, Meng Chen, Youzheng Wu, X. He, and B. Zhou. 2021. Conversational Query Rewriting with Self-supervised Learning. ArXiv abs/2102.04708(2021).Google ScholarGoogle Scholar
  10. Vladimir Karpukhin, Barlas Oguz, Sewon Min, Patrick S. H. Lewis, Ledell Wu, Sergey Edunov, Danqi Chen, and Wen tau Yih. 2020. Dense Passage Retrieval for Open-Domain Question Answering. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP). 6769–6781.Google ScholarGoogle ScholarCross RefCross Ref
  11. Tsung-Hsien Wen, David Vandyke, Nikola Mrksic, Milica Gasic, Lina Maria Rojas-Barahona, Pei-Hao Su, Stefan Ultes, and Steve J. Young. 2017. A Network-based End-to-End Trainable Task-oriented Dialogue System. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 1, Long Papers, Vol. 1. 438–449.Google ScholarGoogle ScholarCross RefCross Ref
  12. Chien-Sheng Wu, Andrea Madotto, Ehsan Hosseini-Asl, Caiming Xiong, Richard Socher, and Pascale Fung. 2019. Transferable Multi-Domain State Generator for Task-Oriented Dialogue Systems. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 808–819.Google ScholarGoogle ScholarCross RefCross Ref
  13. Hung Le, Steven C.H. Hoi, and Richard Socher. 2020. Non-Autoregressive Dialog State Tracking. In ICLR 2020 : Eighth International Conference on Learning Representations.Google ScholarGoogle Scholar
  14. Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is All You Need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Vol. 30. 5998–6008.Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Haoyu Song, Weinan Zhang, Yiming Cui, Dong Wang, and Ting Liu. 2019. Exploiting Persona Information for Diverse Generation of Conversational Responses.. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence. 5190–5196.Google ScholarGoogle ScholarCross RefCross Ref
  16. Rongzhong Lian, Min Xie, Fan Wang, Jinhua Peng, and Hua Wu. 2019. Learning to Select Knowledge for Response Generation in Dialog Systems. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence. 5081–5087.Google ScholarGoogle ScholarCross RefCross Ref
  17. Siqi Bao, Huang He, Fan Wang, Rongzhong Lian, and Hua Wu. 2019. Know More about Each Other: Evolving Dialogue Strategy via Compound Assessment.. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 5382–5391.Google ScholarGoogle ScholarCross RefCross Ref
  18. Siqi Bao, Huang He, Fan Wang, Hua Wu, and Haifeng Wang. 2020. PLATO: Pre-trained Dialogue Generation Model with Discrete Latent Variable. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 85–96.Google ScholarGoogle ScholarCross RefCross Ref
  19. Byeongchang Kim, Jaewoo Ahn, and Gunhee Kim. 2020. Sequential Latent Knowledge Selection for Knowledge-Grounded Dialogue. In ICLR 2020 : Eighth International Conference on Learning Representations.Google ScholarGoogle Scholar
  20. Xiaoxue Zang, Abhinav Rastogi, Srinivas Sunkara, Raghav Gupta, Jianguo Zhang, and Jindong Chen. 2020. MultiWOZ 2.2 : A Dialogue Dataset with Additional Annotation Corrections and State Tracking Baselines. In Proceedings of the 2nd Workshop on Natural Language Processing for Conversational AI. 109–117.Google ScholarGoogle ScholarCross RefCross Ref
  21. Tsung-Hsien Wen, Yishu Miao, Phil Blunsom, and Steve Young. 2017. Latent Intention Dialogue Models. In ICML’17 Proceedings of the 34th International Conference on Machine Learning - Volume 70. 3732–3741.Google ScholarGoogle Scholar
  22. Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv preprint arXiv:1907.11692(2019).Google ScholarGoogle Scholar
  23. Alexandr Andoni and Piotr Indyk. 2008. Near-optimal hashing algorithms for approximate nearest neighbor in high dimensions. Communications of The ACM 51, 1 (2008), 117–122.Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. Kelvin Guu, Kenton Lee, Zora Tung, Panupong Pasupat, and Mingwei Chang. 2020. Retrieval Augmented Language Model Pre-Training. In ICML 2020: 37th International Conference on Machine Learning, Vol. 1. 3929–3938.Google ScholarGoogle Scholar
  25. Daya Guo, Duyu Tang, Nan Duan, Jian Yin, Daxin Jiang, and Ming Zhou. 2020. Evidence-Aware Inferential Text Generation with Vector Quantised Variational AutoEncoder. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 6118–6129.Google ScholarGoogle ScholarCross RefCross Ref
  26. Davis Liang, Peng Xu, Siamak Shakeri, C. D. Santos, Ramesh Nallapati, Zhiheng Huang, and Bing Xiang. 2020. Embedding-based Zero-shot Retrieval through Query Generation. ArXiv abs/2009.10270(2020).Google ScholarGoogle Scholar
  27. Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Veselin Stoyanov, and Luke Zettlemoyer. 2020. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 7871–7880.Google ScholarGoogle ScholarCross RefCross Ref
  28. Hui Su, Xiaoyu Shen, Rongzhi Zhang, Fei Sun, Pengwei Hu, Cheng Niu, and Jie Zhou. 2019. Improving Multi-turn Dialogue Modelling with Utterance ReWriter. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Florence, Italy, 22–31. https://doi.org/10.18653/v1/P19-1003Google ScholarGoogle ScholarCross RefCross Ref
  29. Stephen Robertson and Hugo Zaragoza. 2009. The Probabilistic Relevance Framework.Google ScholarGoogle Scholar
  30. Chulaka Gunasekara, Seokhwan Kim, Luis Fernando D’Haro, Abhinav Rastogi, Y. Chen, M. Eric, Behnam Hedayatnia, Karthik Gopalakrishnan, Y. Liu, Chao-Wei Huang, D. Hakkani-Tur, Jinchao Li, Qi Zhu, Lingxiao Luo, L. Liden, Kaili Huang, Shahin Shayandeh, Runze Liang, Baolin Peng, Zheng Zhang, Swadheen Shukla, Minlie Huang, Jianfeng Gao, Shikib Mehri, Y. Feng, Carla Gordon, S. Alavi, David Traum, M. Eskénazi, A. Beirami, Eunjoon Cho, Paul A. Crook, Ankita De, A. Geramifard, S. Kottur, Seungwhan Moon, S. Poddar, and Rajen Subba. 2020. Overview of the Ninth Dialog System Technology Challenge: DSTC9. ArXiv abs/2011.06486(2020).Google ScholarGoogle Scholar
  31. Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, Rémi Louf, Morgan Funtowicz, and Jamie Brew. 2019. HuggingFace’s Transformers: State-of-the-art Natural Language Processing.arXiv preprint arXiv:1910.03771(2019).Google ScholarGoogle Scholar
  32. Diederik P. Kingma and Jimmy Lei Ba. 2015. Adam: A Method for Stochastic Optimization. In 2015 International Conference on Learning Representations (ICLR).Google ScholarGoogle Scholar

Index Terms

  1. Learning Dense Entity-Aware Dialogue Intentions with Rewritten Utterance for External Knowledge Documents Retrieval
          Index terms have been assigned to the content through auto-classification.

          Recommendations

          Comments

          Login options

          Check if you have access through your login credentials or your institution to get full access on this article.

          Sign in
          • Published in

            cover image ACM Other conferences
            SPML '21: Proceedings of the 2021 4th International Conference on Signal Processing and Machine Learning
            August 2021
            183 pages
            ISBN:9781450390170
            DOI:10.1145/3483207

            Copyright © 2021 ACM

            Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

            Publisher

            Association for Computing Machinery

            New York, NY, United States

            Publication History

            • Published: 28 October 2021

            Permissions

            Request permissions about this article.

            Request Permissions

            Check for updates

            Qualifiers

            • research-article
            • Research
            • Refereed limited

          PDF Format

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader

          HTML Format

          View this article in HTML Format .

          View HTML Format