Skip to main content

Extracting Structural Knowledge for Professional Text Inference

  • Conference paper
  • First Online:
Computer Supported Cooperative Work and Social Computing (ChineseCSCW 2023)

Part of the book series: Communications in Computer and Information Science ((CCIS,volume 2013))

  • 148 Accesses

Abstract

Grading subjective questions of specialty text is a kind of text inference task. Since there are many specialty terms and concepts, it is difficult to judge the knowledge contained in a text as the usual way on inferring a general text. In this paper, we propose a specialty text inference model by extracting the structural knowledge from text. We first propose a knowledge graph construction method for the extraction of knowledge from specialty texts. By combining the constructed knowledge features with the text semantic features, we design the specialty text inference model. Finally, we use real datasets from a national professional exam to validate the soundness of the knowledge graph construction method and the performance of the inference model. The experiments under different training set sizes and network structures are also conducted to detailly analyze the design of our method. The experimental results show the effectiveness and practicality of our approach.

This work was supported by the National Nature Science Foundation of China, NSFC (62376138) and the Innovative Development Joint Fund Key Projects of Shandong NSF (ZR2022LZH007).

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 79.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 99.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Di, S., Shen, Y., Chen, L.: Relation extraction via domain-aware transfer learning. In: Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pp. 1348–1357 (2019)

    Google Scholar 

  2. Ding, X., Lybarger, K., Tauscher, J., Cohen, T.: Improving classification of infrequent cognitive distortions: domain-specific model vs. data augmentation. In: Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: Student Research Workshop, pp. 68–75 (2022)

    Google Scholar 

  3. Duan, B., Wang, S., Liu, X., Xu, Y.: Cluster-aware pseudo-labeling for supervised open relation extraction. In: Proceedings of the 29th International Conference on Computational Linguistics, pp. 1834–1841 (2022)

    Google Scholar 

  4. Gu, J., Sun, F., Qian, L., Zhou, G.: Chemical-induced disease relation extraction via attention-based distant supervision. BMC Bioinform. 20, 1–14 (2019)

    Article  Google Scholar 

  5. Hu, X., Zhang, C., Xu, Y., Wen, L., Yu, P.S.: Selfore: self-supervised relational feature learning for open relation extraction. arXiv preprint arXiv:2004.02438 (2020)

  6. Jia, C., Liang, X., Zhang, Y.: Cross-domain NER using cross-domain language modeling. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 2464–2474 (2019)

    Google Scholar 

  7. Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: Bert: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of NAACL-HLT, vol. 1, p. 2 (2019)

    Google Scholar 

  8. Lee, J., et al.: Biobert: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics 36(4), 1234–1240 (2020)

    Article  MathSciNet  Google Scholar 

  9. Li, D., Liu, T., Pan, W., Liu, X., Sun, Y., Yuan, F.: Grading Chinese answers on specialty subjective questions. In: Sun, Y., Lu, T., Yu, Z., Fan, H., Gao, L. (eds.) ChineseCSCW 2019. CCIS, vol. 1042, pp. 670–682. Springer, Singapore (2019). https://doi.org/10.1007/978-981-15-1377-0_52

    Chapter  Google Scholar 

  10. Li, Z., Tomar, Y., Passonneau, R.J.: A semantic feature-wise transformation relation network for automatic short answer grading. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 6030–6040 (2021)

    Google Scholar 

  11. Liu, D., et al.: Tell me how to ask again: question data augmentation with controllable rewriting in continuous space. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 5798–5810 (2020)

    Google Scholar 

  12. Liu, Y., et al.: Roberta: a robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 (2019)

  13. Lun, J., Zhu, J., Tang, Y., Yang, M.: Multiple data augmentation strategies for improving performance on automatic short answer scoring. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, pp. 13389–13396 (2020)

    Google Scholar 

  14. Sennrich, R., Haddow, B., Birch, A.: Improving neural machine translation models with monolingual data. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 86–96 (2016)

    Google Scholar 

  15. Sung, C., Dhamecha, T., Saha, S., Ma, T., Reddy, V., Arora, R.: Pre-training bert on domain resources for short answer grading. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 6071–6075 (2019)

    Google Scholar 

  16. Tang, D., Qin, B., Liu, T.: Document modeling with gated recurrent neural network for sentiment classification. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pp. 1422–1432 (2015)

    Google Scholar 

  17. Tran, T.T., Le, P., Ananiadou, S.: Revisiting unsupervised relation extraction. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 7498–7505 (2020)

    Google Scholar 

  18. Wei, J., Zou, K.: EDA: easy data augmentation techniques for boosting performance on text classification tasks. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 6382–6388 (2019)

    Google Scholar 

  19. Wu, J., Liu, T., Sun, Y., Gong, B.: A light transfer model for Chinese named entity recognition for specialty domain. In: Sun, Y., Liu, D., Liao, H., Fan, H., Gao, L. (eds.) ChineseCSCW 2020. CCIS, vol. 1330, pp. 530–541. Springer, Singapore (2021). https://doi.org/10.1007/978-981-16-2540-4_38

    Chapter  Google Scholar 

  20. Zhang, H., Cisse, M., Dauphin, Y.N., Lopez-Paz, D.: mixup: beyond empirical risk minimization. arXiv preprint arXiv:1710.09412 (2017)

  21. Zhao, J., Gui, T., Zhang, Q., Zhou, Y.: A relation-oriented clustering method for open relation extraction. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 9707–9718 (2021)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yuqing Sun .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Xia, T., Wang, J., Liu, T., Jiang, H., Sun, Y. (2024). Extracting Structural Knowledge for Professional Text Inference. In: Sun, Y., Lu, T., Wang, T., Fan, H., Liu, D., Du, B. (eds) Computer Supported Cooperative Work and Social Computing. ChineseCSCW 2023. Communications in Computer and Information Science, vol 2013. Springer, Singapore. https://doi.org/10.1007/978-981-99-9640-7_25

Download citation

  • DOI: https://doi.org/10.1007/978-981-99-9640-7_25

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-99-9639-1

  • Online ISBN: 978-981-99-9640-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics