Abstract
Slot filling plays an important role in spoken language understanding. Slot prediction need to use a lot of labeled data in a specific field for training, but in the real situation, there is often a lack of training data in a specific field, which is the biggest problem in cross-domain slot filling. In the previous works on cross-domain slot filling, many methods train their model through the sufficient source domain data, so that the model could predict the slot type in the unknown domain. However, previous approaches do not make good use of the small amount of labeled target domain data. In this paper, we proposed a cross-domain slot filling model with label-aware transfer learning. First, we classify words into three categories based on their BIO labels, calculate the MMD (maximum mean discrepancy) by computing hidden representations between two domains with the same ground truth label, which can participate in the loss function calculation, so that the model can better capture the overall characteristics of the target domain. Experimental results show that our proposed models significantly outperform other methods on average F1-score.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Liu, B., Lane, I.: Recurrent neural network structured output prediction for spoken language understanding. In: NIPS Workshop on Machine Learning for Spoken Language Understanding and Interactions (2015)
Bapna, A., Tür, G., Hakkani-Tür, D., Heck, L.: Towards zero-shot frame semantic parsing for domain scaling. In: Interspeech 2017, August 2017
Shah, D., Gupta, R, Fayazi, A., Hakkani-Tür, D.: Robust zero-shot cross-domain slot filling with example values. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 5484–5490 (2019)
Lee, S., Jha, R.: Zero-shot adaptive transfer for conversational language understanding. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, pp. 6642–6649 (2019)
Liu, Z., Winata, G., Fung, P.: Zero-resource cross-domain named entity recognition. In: ACL, pp. 19–25 (2020)
Wang, Z., Qu, Y., Chen, L.: Label-aware double transfer learning for cross-specialty medi-cal named entity recognition. In: NAACL, pp. 1–15 (2018)
Gretton, A., Borgwardt, K.M., Rasch, M.J., Schölkopf, B., Smola, A.: A kernel two-sample test. J. Mach. Learn. Res. 13, 723–773 (2012)
Elman, J.L.: Distributed representations, simple recurrent networks, and grammatical structure. Mach. Learn. 7, 195–225 (1991)
Hochreiter, S., Schmidhuber, J.: Long short- term memory. Neural Comput. 9(8), 1735–1780 (1997)
Liu, B., Lane, I.: Attention-based recurrent neural network models for joint intent detection and slot filling. In: Interspeech, pp. 685–689 (2016)
Yao, K.: Applications of reproducing kernel Hilbert spaces-bandlimited signal models. Inf. Control 11(4), 429–444 (1967)
Raymond, C., Riccardi, G.: Generative and discriminative algorithms for spoken language understanding. In: Interspeech, pp. 1605–1608 (2007)
Goldberg, Y., Levy, O.: Word2vec explained: deriving Mikolov et al’.s negative sampling word-embedding method. arXiv preprint arXiv:1402.3722 (2014)
Coucke, A., et al.: Snips voice platform: an embedded spoken language understanding system for private-by-design voice interfaces. arXiv preprint arXiv:1805.10190 (2018)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Liu, G., Ju, S., Chen, Y. (2021). Binary Label-Aware Transfer Learning for Cross-Domain Slot Filling. In: Mantoro, T., Lee, M., Ayu, M.A., Wong, K.W., Hidayanto, A.N. (eds) Neural Information Processing. ICONIP 2021. Lecture Notes in Computer Science(), vol 13109. Springer, Cham. https://doi.org/10.1007/978-3-030-92270-2_31
Download citation
DOI: https://doi.org/10.1007/978-3-030-92270-2_31
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-92269-6
Online ISBN: 978-3-030-92270-2
eBook Packages: Computer ScienceComputer Science (R0)