Skip to main content

Binary Label-Aware Transfer Learning for Cross-Domain Slot Filling

  • Conference paper
  • First Online:
  • 1599 Accesses

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 13109))

Abstract

Slot filling plays an important role in spoken language understanding. Slot prediction need to use a lot of labeled data in a specific field for training, but in the real situation, there is often a lack of training data in a specific field, which is the biggest problem in cross-domain slot filling. In the previous works on cross-domain slot filling, many methods train their model through the sufficient source domain data, so that the model could predict the slot type in the unknown domain. However, previous approaches do not make good use of the small amount of labeled target domain data. In this paper, we proposed a cross-domain slot filling model with label-aware transfer learning. First, we classify words into three categories based on their BIO labels, calculate the MMD (maximum mean discrepancy) by computing hidden representations between two domains with the same ground truth label, which can participate in the loss function calculation, so that the model can better capture the overall characteristics of the target domain. Experimental results show that our proposed models significantly outperform other methods on average F1-score.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. Liu, B., Lane, I.: Recurrent neural network structured output prediction for spoken language understanding. In: NIPS Workshop on Machine Learning for Spoken Language Understanding and Interactions (2015)

    Google Scholar 

  2. Bapna, A., Tür, G., Hakkani-Tür, D., Heck, L.: Towards zero-shot frame semantic parsing for domain scaling. In: Interspeech 2017, August 2017

    Google Scholar 

  3. Shah, D., Gupta, R, Fayazi, A., Hakkani-Tür, D.: Robust zero-shot cross-domain slot filling with example values. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 5484–5490 (2019)

    Google Scholar 

  4. Lee, S., Jha, R.: Zero-shot adaptive transfer for conversational language understanding. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, pp. 6642–6649 (2019)

    Google Scholar 

  5. Liu, Z., Winata, G., Fung, P.: Zero-resource cross-domain named entity recognition. In: ACL, pp. 19–25 (2020)

    Google Scholar 

  6. Wang, Z., Qu, Y., Chen, L.: Label-aware double transfer learning for cross-specialty medi-cal named entity recognition. In: NAACL, pp. 1–15 (2018)

    Google Scholar 

  7. Gretton, A., Borgwardt, K.M., Rasch, M.J., Schölkopf, B., Smola, A.: A kernel two-sample test. J. Mach. Learn. Res. 13, 723–773 (2012)

    MathSciNet  MATH  Google Scholar 

  8. Elman, J.L.: Distributed representations, simple recurrent networks, and grammatical structure. Mach. Learn. 7, 195–225 (1991)

    Google Scholar 

  9. Hochreiter, S., Schmidhuber, J.: Long short- term memory. Neural Comput. 9(8), 1735–1780 (1997)

    Article  Google Scholar 

  10. Liu, B., Lane, I.: Attention-based recurrent neural network models for joint intent detection and slot filling. In: Interspeech, pp. 685–689 (2016)

    Google Scholar 

  11. Yao, K.: Applications of reproducing kernel Hilbert spaces-bandlimited signal models. Inf. Control 11(4), 429–444 (1967)

    Article  Google Scholar 

  12. Raymond, C., Riccardi, G.: Generative and discriminative algorithms for spoken language understanding. In: Interspeech, pp. 1605–1608 (2007)

    Google Scholar 

  13. Goldberg, Y., Levy, O.: Word2vec explained: deriving Mikolov et al’.s negative sampling word-embedding method. arXiv preprint arXiv:1402.3722 (2014)

  14. Coucke, A., et al.: Snips voice platform: an embedded spoken language understanding system for private-by-design voice interfaces. arXiv preprint arXiv:1805.10190 (2018)

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Shenggen Ju .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2021 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Liu, G., Ju, S., Chen, Y. (2021). Binary Label-Aware Transfer Learning for Cross-Domain Slot Filling. In: Mantoro, T., Lee, M., Ayu, M.A., Wong, K.W., Hidayanto, A.N. (eds) Neural Information Processing. ICONIP 2021. Lecture Notes in Computer Science(), vol 13109. Springer, Cham. https://doi.org/10.1007/978-3-030-92270-2_31

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-92270-2_31

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-92269-6

  • Online ISBN: 978-3-030-92270-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics