Skip to main content

Choose Your Words Wisely: Leveraging Embedded Dialog Trajectories to Enhance Performance in Open-Domain Conversations

  • Conference paper
  • First Online:
Book cover Agents and Artificial Intelligence (ICAART 2020)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 12613))

Included in the following conference series:

  • 707 Accesses

Abstract

Human conversations are notoriously nondeterministic, and identical conversation histories can nevertheless accept dozens, if not hundreds, of distinct valid responses. In this paper, we present and expand upon Conversational Scaffolding, a response scoring method that capitalizes on this fundamental linguistic property. We envision a conversation as a set of trajectories through embedding space. Our method leverages the analogical structure encoded within language model representations to prioritize possible conversational responses with respect to these trajectories. Specifically, we locate candidate responses based on their linear offsets relative to the scaffold sentence pair with the greatest cosine similarity to the current conversation history. In an open-domain dialog setting, we are able to show that our method outperforms both an Approximate Nearest-Neighbor approach and a naive nearest neighbor baseline. We demonstrate our method’s performance on a retrieval-based dialog task using a retrieval dataset containing 19,665 randomly-selected sentences. We further introduce a comparative analysis of algorithm performance as a function of contextual alignment strategy, with accompanying discussion.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    https://github.com/BYU-PCCL/chitchat-dataset.

  2. 2.

    https://aclanthology.coli.uni-saarland.de/papers/I17-1099/i17-1099.

  3. 3.

    http://files.pushshift.io/reddit/.

  4. 4.

    https://www.kaggle.com/rtatman/ubuntu-dialogue-corpus.

  5. 5.

    Due to the massive size of Reddit, we only used a subset of the comments and posts from June 2014 to November 2014.

  6. 6.

    Similarity was defined as Euclidean distance < \(\tau \), where \(\tau \) is a hand-selected threshold value.

References

  1. Abadi, M., et al.: TensorFlow: large-scale machine learning on heterogeneous systems, software available from tensorflow.org (2015). https://www.tensorflow.org/

  2. Bak, J., Oh, A.: Variational hierarchical user-based conversation model. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 1941–1950. Association for Computational Linguistics, Hong Kong, China, November 2019. https://doi.org/10.18653/v1/D19-1202, https://www.aclweb.org/anthology/D19-1202

  3. Banerjee, S., Lavie, A.: Meteor: an automatic metric for MT evaluation with improved correlation with human judgments. In: Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization, pp. 65–72 (2005)

    Google Scholar 

  4. Bojanowski, P., Grave, E., Joulin, A., Mikolov, T.: Enriching word vectors with subword information. arXiv preprint arXiv:1607.04606 (2016)

  5. Bowman, S.R., Vilnis, L., Vinyals, O., Dai, A.M., Jozefowicz, R., Bengio, S.: Generating sentences from a continuous space. CoRR abs/1511.06349 (2015). http://arxiv.org/abs/1511.06349

  6. Buckels, E.E., Trapnell, P.D., Paulhus, D.L.: Trolls just want to have fun. Pers. Individ. Differ. 67, 97–102 (2014)

    Article  Google Scholar 

  7. Cer, D., Diab, M., Agirre, E., Lopez-Gazpio, I., Specia, L.: Semeval-2017 task 1: semantic textual similarity - multilingual and crosslingual focused evaluation. In: Proceedings of SemEval, vol. 2017 (2017)

    Google Scholar 

  8. Cer, D., et al.: Universal sentence encoder. CoRR abs/1803.11175 (2018). http://arxiv.org/abs/1803.11175

  9. Cho, D., Acquisti, A.: The more social cues, the less trolling? an empirical study of online commenting behavior (2013)

    Google Scholar 

  10. Conneau, A., Kiela, D., Schwenk, H., Barrault, L., Bordes, A.: Supervised learning of universal sentence representations from natural language inference data. arXiv preprint arXiv:1705.02364 (2017)

  11. Conneau, A., Kruszewski, G., Lample, G., Barrault, L., Baroni, M.: What you can cram into a single vector: Probing sentence embeddings for linguistic properties. arXiv preprint arXiv:1805.01070 (2018)

  12. Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)

  13. Fulda, N., et al.: Byu-eve: mixed initiative dialog via structured knowledge graph traversal and conversational scaffolding. In: Proceedings of the 2018 Amazon Alexa Prize, November 2018

    Google Scholar 

  14. Gladkova, A., Drozd, A., Matsuoka, S.: Analogy-based detection of morphological and semantic relations with word embeddings: what works and what doesn’t. In: Proceedings of the NAACL Student Research Workshop, pp. 8–15 (2016)

    Google Scholar 

  15. Kiros, R., et al.: Skip-thought vectors. CoRR abs/1506.06726 (2015)

    Google Scholar 

  16. Li, Y., Su, H., Shen, X., Li, W., Cao, Z., Niu, S.: DailyDialog: A Manually Labelled Multi-turn Dialogue Dataset. arXiv e-prints arXiv:1710.03957 (2017)

  17. Logeswaran, L., Lee, H.: An efficient framework for learning sentence representations. In: International Conference on Learning Representations (2018). https://openreview.net/forum?id=rJvJXZb0W

  18. Lowe, R., Pow, N., Serban, I., Pineau, J.: The Ubuntu Dialogue Corpus: A Large Dataset for Research in Unstructured Multi-Turn Dialogue Systems. arXiv e-prints arXiv:1506.08909 (2015)

  19. Mikolov, T., Chen, K., Corrado, G., Dean, J.: Efficient estimation of word representations in vector space. CoRR abs/1301.3781 (2013)

    Google Scholar 

  20. Mikolov, T., tau Yih, W., Zweig, G.: Linguistic regularities in continuous space word representations. Association for Computational Linguistics, May 2013

    Google Scholar 

  21. Park, Y., Cho, J., Kim, G.: A hierarchical latent structure for variational conversation modeling. In: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Long Papers), vol. 1, pp. 1792–1801. Association for Computational Linguistics, New Orleans, Louisiana, June 2018. https://doi.org/10.18653/v1/N18-1162, https://www.aclweb.org/anthology/N18-1162

  22. Pennington, J., Socher, R., Manning, C.D.: Glove: global vectors for word representation. In: Empirical Methods in Natural Language Processing (EMNLP), pp. 1532–1543 (2014). http://www.aclweb.org/anthology/D14-1162

  23. Peters, M.E., et al.: Deep contextualized word representations. In: Proceedings of NAACL (2018)

    Google Scholar 

  24. Rainie, H., Anderson, J.Q.: The future of free speech, trolls, anonymity and fake news online (2017)

    Google Scholar 

  25. Reddit: Reddit datasets. https://www.reddit.com/r/datasets/

  26. Shen, X., Su, H., Niu, S., Demberg, V.: Improving variational encoder-decoders in dialogue generation. In: Thirty-Second AAAI Conference on Artificial Intelligence (2018)

    Google Scholar 

  27. Will, M., Tyler, E., Nancy, F.: Conversational scaffolding: an analogy-based approach to response prioritization in open-domain dialogs. In: Proceedings of the 12th International Conference on Agents and Artificial Intelligence (ICAART) (2020)

    Google Scholar 

  28. Xupeng Tong, Y.L., Yen, C.M.: Variational neural conversational model. In: ICML (2014). https://www.cs.cmu.edu/epxing/Class/10708-17/project-reports/project12.pdf

  29. Zhang, T., Kishore, V., Wu, F., Weinberger, K.Q., Artzi, Y.: Bertscore: evaluating text generation with bert. arXiv preprint arXiv:1904.09675 (2019)

  30. Zhao, T., Zhao, R., Eskenazi, M.: Learning discourse-level diversity for neural dialog models using conditional variational autoencoders. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Long Papers), vol. 1, pp. 654–664. Association for Computational Linguistics, Vancouver, Canada, July 2017. https://doi.org/10.18653/v1/P17-1061, https://www.aclweb.org/anthology/P17-1061

  31. Zhu, X., Li, T., De Melo, G.: Exploring semantic properties of sentence embeddings. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Short Papers), vol. 2, pp. 632–637 (2018)

    Google Scholar 

  32. Zhu, Y., et al.: Aligning books and movies: towards story-like visual explanations by watching movies and reading books. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 19–27 (2015)

    Google Scholar 

Download references

Acknowledgements

We wish to thank David Wingate and his students in the BYU Perception, Control and Cognition laboratory for their role in creating and hosting the Chit-Chat dataset, and Daniel Ricks for his contributions to Fig. 3.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Nancy Fulda .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2021 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Fulda, N., Etchart, T., Myers, W. (2021). Choose Your Words Wisely: Leveraging Embedded Dialog Trajectories to Enhance Performance in Open-Domain Conversations. In: Rocha, A.P., Steels, L., van den Herik, J. (eds) Agents and Artificial Intelligence. ICAART 2020. Lecture Notes in Computer Science(), vol 12613. Springer, Cham. https://doi.org/10.1007/978-3-030-71158-0_11

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-71158-0_11

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-71157-3

  • Online ISBN: 978-3-030-71158-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics