Abstract
There is heavy dependence on the large amount of annotated data in most existing methods of relation classification, which is a serious problem. Besides, we cannot learn by leveraging past learned knowledge in most situation, which means it can only train from scratch to learn new tasks. Motivated from humans’ ability of learning effectively from few samples and learning quickly by utilizing learned knowledge, we use both meta network based on co-reference resolution and prototypical network based on co-reference resolution to resolve the problem of few-shot relation classification for crossing-sentence task. Both of the two network aim to learn a transferrable deep distance metric to recognize new relation categories given very few labelled samples. Instead of single sentence, paragraphs containing multi-sentence is a major concern in the experiment. The results demonstrate that our approach performs well and achieves high precision.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Mintz, M., Bills, S., Snow, R., Jurafsky, D.: Distant supervision for relation extraction without labeled data. In: Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP, vol. 2, pp. 1003–1011. Association for Computational Linguistics, Suntec, Singapore (2009)
Han, X., et al.: FewRel: a large-scale supervised few-shot relation classification dataset with state-of-the-art evaluation. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 4803–4809. Association for Computational Linguistics, Brussels, Belgium (2018)
Gao, T., Han, X., Liu, Z., Sun, M.: Hybrid attention-based prototypical networks for noisy few-shot relation classification. In: The Thirty-Third AAAI Conference on Artificial Intelligence. Association for the Advancement of Artificial Intelligence, Hawaii, USA (2019)
Koch, G., Zemel, R., Salakhutdinov, R.: Siamese neural networks for one-shot image recognition. In: Proceedings of the 32nd International Conference on Machine Learning, Lille, France, vol. 37. JMLR, July 2015
Clark, K., Manning, C.D.: Improving coreference resolution by learning entity-level distributed representations. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, pp. 643–653, Berlin, Germany. Association for Computational Linguistics (2016)
Vinyals, O., Blundell, C., Lillicrap, T., Wierstra, D.: Matching networks for one shot learning. In: Neural Information Processing Systems, pp. 3630–3638. Curran Associates Inc., Barcelona, Spain (2016)
Snell, J., Swersky, K., Zemel, R.: Prototypical networks for few-shot learning. In: Neural Information Processing Systems, pp. 4077–4087. Curran Associates Inc., Long Beach, USA (2017)
Sung, F., Yang, Y., Zhang, L., Xiang, T., Torr, P.H., Hospedales, T.M.: Learning to compare: relation network for few-shot learning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, UT, USA, pp. 1199–1208. IEEE (2018)
Ravi, S., Larochelle, H.: Optimization as a model for few-shot learning. In: International Conference on Learning Representations. Toulon, France (2017)
Luo, H., Glass, J.: Learning word representations with cross-sentence dependency for end-to-end co-reference resolution. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, pp. 4829–4833. Association for Computational Linguistics (2018)
Cheng, J., Dong, L., Lapata, M.: Long short-term memory-networks for machine reading. In: Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pp. 551–561. Association for Computational Linguistics, Austin, Texas (2016)
Lee, K., He, L., Lewis, M., Zettlemoyer, L.: End-to-end neural coreference resolution. In: Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pp. 188–197. Association for Computational Linguistics, Copenhagen, Denmark (2017)
Durrett, G., Klein, D.: Easy victories and uphill battles in coreference resolution. In: Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, Seattle, Washington, USA, pp. 1971–1982. Association for Computational Linguistics (2013)
Peters, M.E., et al.: Deep contextualized word representations. In: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, New Orleans, Louisiana, vol. 1, pp. 2227–2237. Association for Computational Linguistics (2018)
Wiseman, S., Rush, A.M., Shieber, S.M.: Learning global features for coreference resolution. In: Proceedings of NAACL-HLT 2016, pp. 994–1004, San Diego, California. Association for Computational Linguistics (2016)
Bengio, Y.: Deep learning of representations for unsupervised and transfer learning. In: Proceedings of ICML Workshop on Unsupervised and Transfer Learning, Washington, USA, pp. 17–36. JMLR.org (2011)
Feng, J., Huang, M., Zhao, L., Yang, Y., Zhu, X.: Reinforcement learning for relation classification from noisy data. In: Thirty-Second AAAI Conference on Artificial Intelligence. Association for the Advancement of Artificial Intelligence, New Orleans, Louisiana, USA, April 2018
Finn, C., Abbeel, P., Levine, S.: Model-agnostic meta-learning for fast adaptation of deep networks. In: Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, vol. 70, pp. 1126–1135. JMLR.org, August 2017
Ji, G., Liu, K., He, S., Zhao, J.: Distant supervision for relation extraction with sentence-level attention and entity descriptions. In: AAAI 2017 Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, California, USA, pp. 3060–3066. Association for the Advancement of Artificial Intelligence, February 2017
Luo, B., et al.: Learning with noise: Enhance distantly supervised relation extraction with dynamic transition matrix. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, Vancouver, Canada, pp. 430–439. Association for Computational Linguistics (2017)
Andrychowicz, M., et al.: Learning to learn by gradient descent by gradient descent. In: Neural Information Processing Systems, Barcelona, Spain, pp. 3981–3989. Curran Associates Inc. (2016)
Husken, M., Goerick, C.: Fast learning for problem classes using knowledge based network initialization. In: Proceedings of the IEEE-INNS-ENNS International Joint Conference on Neural Networks, IJCNN 2000. Neural Computing: New Challenges and Perspectives for the New Millennium, Como, Italy, vol. 6, pp. 619–624. IEEE (2000)
Munkhdalai, T., Yu, H.: Meta networks. In: Proceedings of the 34th International Conference on Machine Learning, vol. 70, pp. 2554–2563, Sydney, Australia. JMLR.org, August 2017
Acknowledgments
This research work is supported by National Natural Science Foundation of China (No. 61402220, No. 61502221), the Philosophy and Social Science Foundation of Hunan Province (No. 16YBA323), Scientific Research Fund of Hunan Provincial Education Department for excellent talents (No.18B279).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Wen, W., Liu, Y., Ouyang, C. (2020). Few-Shot Learning for Crossing-Sentence Relation Classification. In: Zhai, G., Zhou, J., Yang, H., An, P., Yang, X. (eds) Digital TV and Wireless Multimedia Communication. IFTC 2019. Communications in Computer and Information Science, vol 1181. Springer, Singapore. https://doi.org/10.1007/978-981-15-3341-9_13
Download citation
DOI: https://doi.org/10.1007/978-981-15-3341-9_13
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-15-3340-2
Online ISBN: 978-981-15-3341-9
eBook Packages: Computer ScienceComputer Science (R0)