Skip to main content

Fine-Tuning BERT for Generative Dialogue Domain Adaptation

  • Conference paper
  • First Online:
Text, Speech, and Dialogue (TSD 2022)

Abstract

Current data-driven Dialogue State Tracking (DST) models exhibit a poor capacity to adapt themselves to domain changes, resulting in a significant degradation in performance. We propose a methodology, called Generative Dialogue Domain Adaptation, which significantly simplifies the creation of training data when a number of changes (e.g., new slot-values or new instances) occur in a domain Knowledge Base. We start from dialogues for a source domain and apply generative methods based on language models such as BERT, fine-tuned on task-related data and generate slot-values substitutions for a target domain. We have experimented dialogue domain adaptation in a few-shot setting showing promising results, although the task is still very challenging. We provide a deep analysis of the quality of the generated data and of the features that affect this task, and we emphasise that DST models are very sensitive to the distribution of slot-values in the corpus.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

Notes

  1. 1.

    http://www.datiopen.it/.

References

  1. Andrews, G.E.: The geometric series in calculus. Am. Math. Mon. 105(1), 36–40 (1998). https://doi.org/10.1080/00029890.1998.12004846

  2. Balaraman, V., Magnini, B.: Domain-aware dialogue state tracker for multi-domain dialogue systems. IEEE/ACM Trans. Audio Speech Lang. Process. 29, 866–873 (2021). https://doi.org/10.1109/TASLP.2021.3054309

    Article  Google Scholar 

  3. Balaraman, V., Sheikhalishahi, S., Magnini, B.: Recent neural methods on dialogue state tracking for task-oriented dialogue systems: a survey. In: Li, H., et al. (eds.) Proceedings of the 22nd Annual Meeting of the Special Interest Group on Discourse and Dialogue, SIGdial 2021, Singapore and Online, 29–31 July 2021, pp. 239–251. Association for Computational Linguistics (2021). https://aclanthology.org/2021.sigdial-1.25

  4. Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp. 4171–4186. Association for Computational Linguistics, Minneapolis, June 2019. https://doi.org/10.18653/v1/N19-1423, https://www.aclweb.org/anthology/N19-1423

  5. Feng, Y., Wang, Y., Li, H.: A sequence-to-sequence approach to dialogue state tracking. arXiv preprint arXiv:2011.09553 (2020)

  6. Han, T., et al.: Multiwoz 2.3: a multi-domain task-oriented dataset enhanced with annotation corrections and co-reference annotation. arXiv preprint arXiv:2010.05594 (2020)

  7. Henderson, M., Thomson, B., Williams, J.D.: The second dialog state tracking challenge. In: Proceedings of the 15th Annual Meeting of the Special Interest Group on Discourse and Dialogue (SIGDIAL), pp. 263–272. Association for Computational Linguistics, Philadelphia, June 2014. https://doi.org/10.3115/v1/W14-4337, https://www.aclweb.org/anthology/W14-4337

  8. Labruna, T., Magnini, B.: Addressing slot-value changes in task-oriented dialogue systems through dialogue domain adaptation. In: Proceedings of RANLP 2021 (2021)

    Google Scholar 

  9. Labruna, T., Magnini, B.: From Cambridge to Pisa: a journey into cross-lingual dialogue domain adaptation for conversational agents (2021)

    Google Scholar 

  10. Madotto, A., Liu, Z., Lin, Z., Fung, P.: Language models as few-shot learner for task-oriented dialogue systems. arXiv preprint arXiv:2008.06239 (2020)

  11. Magnini, B., Louvan, S.: Understanding Dialogue for Human Communication, pp. 1–43. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-44982-7_20-1

  12. McTear, M.: Conversational AI: Dialogue Systems, Conversational Agents, and Chatbots. Morgan and Claypool Publishers, San Rafael (2020). https://doi.org/10.2200/S01060ED1V01Y202010HLT048

  13. Rastogi, A., Zang, X., Sunkara, S., Gupta, R., Khaitan, P.: Towards scalable multi-domain conversational agents: the schema-guided dialogue dataset. In: The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, 7–12 February 2020, pp. 8689–8696. AAAI Press (2020). https://aaai.org/ojs/index.php/AAAI/article/view/6394

  14. Wu, C.S., Madotto, A., Hosseini-Asl, E., Xiong, C., Socher, R., Fung, P.: Transferable multi-domain state generator for task-oriented dialogue systems. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 808–819. Association for Computational Linguistics, Florence, Italy, July 2019. https://doi.org/10.18653/v1/P19-1078, https://www.aclweb.org/anthology/P19-1078

  15. Young, S., et al.: The hidden information state model: a practical framework for POMDP-based spoken dialogue management. Comput. Speech Lang. 24(2), 150–174 (2010)

    Article  Google Scholar 

  16. Zhao, J., Mahdieh, M., Zhang, Y., Cao, Y., Wu, Y.: Effective sequence-to-sequence dialogue state tracking. arXiv preprint arXiv:2108.13990 (2021)

Download references

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to Tiziano Labruna or Bernardo Magnini .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Labruna, T., Magnini, B. (2022). Fine-Tuning BERT for Generative Dialogue Domain Adaptation. In: Sojka, P., Horák, A., Kopeček, I., Pala, K. (eds) Text, Speech, and Dialogue. TSD 2022. Lecture Notes in Computer Science(), vol 13502. Springer, Cham. https://doi.org/10.1007/978-3-031-16270-1_42

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-16270-1_42

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-16269-5

  • Online ISBN: 978-3-031-16270-1

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics