Skip to main content

ASIM: Explicit Slot-Intent Mapping with Attention for Joint Multi-intent Detection and Slot Filling

  • Conference paper
  • First Online:
International Conference on Neural Computing for Advanced Applications (NCAA 2023)

Abstract

The accurate analysis of a user’s natural language statement, including their potential intentions and corresponding slot tags, is crucial for cognitive intelligence services. In real-world applications, a user’s statement often contains multiple intentions, and most existing approaches either mainly focus on the single-intent research problems or utilizes an overall encoder directly to capture the relationship between intents and slot tags, which ignore the explicit slot-intent mapping relation. In this paper, we propose a novel Attention-based Slot-Intent Mapping Method (ASIM) for joint multi-intent detection and slot filling task. The ASIM model not only models the correlation among sequence tags while considering the mutual influence between two tasks but also maps specific intents to each semantic slot. The ASIM model can balance multi-intent knowledge to guide slot filling and further increase the interaction between the two tasks. Experimental results on the MixATIS dataset demonstrate that our ASIM model achieves substantial improvement and state-of-the-art performance.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Liu, B., Lane, I.: Attention-Based Recurrent Neural Network Models for Joint Intent Detection and Slot Filling (2016)

    Google Scholar 

  2. Gangadharaiah, R., Narayanaswamy, B.: Joint multiple intent detection and slot labeling for goal-oriented dialog. In: Proceedings of the 2019 Conference of the North (2019)

    Google Scholar 

  3. Goo, C.-W., et al.: Slot-gated modeling for joint slot filling and intent prediction. In: Proceedings of NAACL (2018)

    Google Scholar 

  4. Li, C., Li, L., Qi, J.: A selfattentive model with gate mechanism for spoken language understanding. In: Proceedings of EMNLP (2018)

    Google Scholar 

  5. Wang, Y., Shen, Y., Jin, H.: A bi-model based RNN semantic frame parsing model for intent detection and slot filling. In: Proceedings of NAACL (2018)

    Google Scholar 

  6. Haihong, E., Niu, P., Chen, Z., Song, M.: A novel bi-directional interrelated model for joint intent detection and slot filling. In: Proceedings of ACL (2019)

    Google Scholar 

  7. Qin, L., Xu, X., Che, W., Liu, T.: AGIF: an adaptive graph-interactive framework for joint multiple intent detection and slot filling. In: EMNLP Findings (2020)

    Google Scholar 

  8. Mccallum, A., Nigam, K.: A comparison of event models for Naive Bayes text classification. In: AAAI-98 Workshop on Learning for Text Categorization, pp. 41–48 (1998)

    Google Scholar 

  9. Haffner, P., Tur, G., Wright, J.H.: Optimizing SVMs for complex call classification. In: IEEE International Conference on Acoustics, pp. 632–635 (2003)

    Google Scholar 

  10. Genkin, A., Lewis, D.D., Madigan, D.: Large-scale Bayesian logistic regression for text categorization. Technometrics 49(3), 291–304 (2007)

    Article  MathSciNet  Google Scholar 

  11. Dowding, J., et al.: Gemini: a natural language system for spoken-language understanding (1994)

    Google Scholar 

  12. Pengju, Y.: Research on Natural Language Understanding in Conversational Systems. Tsinghua University, Beijing (2002)

    Google Scholar 

  13. Xu, P., Sarikaya, R.: Convolutional neural network based triangular CRF for joint intent detection and slot filling. In: 2013 IEEE Workshop on Automatic Speech Recognition and Understanding, pp. 78–83 (2013)

    Google Scholar 

  14. Ravuri, S., Stolcke, A.: Recurrent neural network and LSTM models for lexical utterance classification. In: Interspeech (2015)

    Google Scholar 

  15. Wang, Q.: Biological named entity recognition combining dictionary and machine learning. Dalian University of Technology (2009)

    Google Scholar 

  16. Collins, M., Singer, Y.: Unsupervised models for named entity classification. In: Joint SIGDAT Conference on Empirical Methods in Natural Language Processing and Very Large Corpora, pp. 100–110 (1999)

    Google Scholar 

  17. Cucerzan, S., Yarowsky, D.: Language independent named entity recognition combining morphological and contextual evidence. In: Joint SIGDAT Conference on Empirical Methods in Natural Language Processing and Very Large Corpora, pp. 90–99 (1999)

    Google Scholar 

  18. Mikheev, A., Moens, M., Grover, C.: Named entity recognition without gazetteers. In: Proceedings of the Ninth Conference on European Chapter of the Association for Computational Linguistics. Association for Computational Linguistics, pp. 1–8 (1999)

    Google Scholar 

  19. Wei, L., Mccallum, A.: Rapid development of Hindi named entity recognition using conditional random fields and feature induction. ACM Trans. Asian Lang. Inf. Process. 2(3), 290–294 (2003)

    Article  Google Scholar 

  20. Bikel, D.M., Miller, S., Schwartz, R., et al.: Nymble: a high-performance learning name-finder. Anlp, pp. 194–201 (1998)

    Google Scholar 

  21. Bikel, D.M., Schwartz, R., Weischedel, R.M.: An algorithm that learns what’s in a name. Mach. Learn. 34(1–3), 211–231 (1999)

    Article  MATH  Google Scholar 

  22. Borthwick, A., Grishman, R.: A maximum entropy approach to named entity recognition. Graduate School of Arts and Science. New York University (1999)

    Google Scholar 

  23. Yao, K., Zweig, G., Hwang, M.-Y., Shi, Y., Yu, D.: Recurrent neural networks for language understanding. In: Interspeech (2013)

    Google Scholar 

  24. Mesnil, G., He, X., Deng, L., Bengio, Y.: Investigation of recurrentneural-network architectures and learning methods for spoken language understanding. In: Interspeech (2013)

    Google Scholar 

  25. Yao, K., Peng, B., Zhang, Y., Yu, D., Zweig, G., Shi, Y.: Spoken language understanding using long short-term memory neural networks. In: SLT (2014)

    Google Scholar 

  26. Qin, L., Che, W., Li, Y., et al.: A stack-propagation framework with token-level intent detection for spoken language understanding (2019)

    Google Scholar 

  27. Guo, D., Tur, G., Yih, W.-T., Zweig, G.: Joint semantic utterance classification and slot filling with recursive neural networks. In: 2014 IEEE Spoken Language Technology Workshop (SLT), pp. 554–559 (2014)

    Google Scholar 

  28. Zhang, X., Wang, H.: A joint model of intent determination and slot filling for spoken language understanding. In: IJCAI, vol. 16, pp. 2993–2999 (2016)

    Google Scholar 

  29. Liu, B., Lane, I.: Joint online spoken language understanding and language modeling with recurrent neural networks. arXiv preprint arXiv:1609.01462 (2016)

  30. Qin, L., Ni, M., Zhang, Y., Che, W.: Cosda-ml: multi-lingual code-switching data augmentation for zero-shot cross-lingual NLP. arXiv preprint arXiv:2006.06402 (2020)

  31. Devlin, J., Chang, M.-W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: NAACL-HLT (1) (2019)

    Google Scholar 

  32. Sun, Y., et al.: Ernie 2.0: a continual pre-training framework for language understanding. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, no. 05, pp. 8968–8975 (2020)

    Google Scholar 

  33. Yang, Z., Dai, Z., Yang, Y., Carbonell, J., Salakhutdinov, R.R., Le, Q.V.: Xlnet: generalized autoregressive pretraining for language understanding. In: Advances in Neural Information Processing Systems, vol. 32 (2019)

    Google Scholar 

  34. Chen, Q., Zhuo, Z., Wang, W.: BERT for joint intent classification and slot filling. arXiv preprint arXiv:1902.10909 (2019)

  35. Zhang, Z., Zhang, Z., Chen, H., Zhang, Z.: A joint learning framework with BERT for spoken language understanding. IEEE Access 7, 168:849–168:858 (2019)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Mingzhi Wang .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Chen, J. et al. (2023). ASIM: Explicit Slot-Intent Mapping with Attention for Joint Multi-intent Detection and Slot Filling. In: Zhang, H., et al. International Conference on Neural Computing for Advanced Applications. NCAA 2023. Communications in Computer and Information Science, vol 1870. Springer, Singapore. https://doi.org/10.1007/978-981-99-5847-4_16

Download citation

  • DOI: https://doi.org/10.1007/978-981-99-5847-4_16

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-99-5846-7

  • Online ISBN: 978-981-99-5847-4

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics