skip to main content
research-article

Meta-Information Fusion of Hierarchical Semantics Dependency and Graph Structure for Structured Text Classification

Published:20 February 2023Publication History
Skip Abstract Section

Abstract

Structured text with plentiful hierarchical structure information is an important part in real-world complex texts. Structured text classification is attracting more attention in natural language processing due to the increasing complexity of application scenarios. Most existing methods treat structured text from a local hierarchy perspective, focusing on the semantics dependency and the graph structure of the structured text independently. However, structured text has global hierarchical structures with sophisticated dependency when compared to unstructured text. According to the variety of structured texts, it is not appropriate to use the existing methods directly. The function of distinction information within semantics dependency and graph structure for structured text, referred to as meta-information, should be stated more precisely. In this article, we propose HGMETA, a novel meta-information embedding frame network for structured text classification, to obtain the fusion embedding of hierarchical semantics dependency and graph structure in a structured text, and to distill the meta-information from fusion characteristics. To integrate the global hierarchical features with fused structured text information, we design a hierarchical LDA module and a structured text embedding module. Specially, we employ a multi-hop message passing mechanism to explicitly incorporate complex dependency into a meta-graph. The meta-information is constructed from meta-graph via neighborhood-based propagation to distill redundant information. Furthermore, using an attention-based network, we investigate the complementarity of semantics dependency and graph structure based on global hierarchical characteristics and meta-information. Finally, the fusion embedding and the meta-information can be straightforwardly incorporated for structured text classification. Experiments conducted on three real-world datasets show the effectiveness of meta-information and demonstrate the superiority of our method.

REFERENCES

  1. [1] Ion Androutsopoulos, John Koutsias, Konstantinos Chandrinos, Georgios Paliouras, and Constantine D. Spyropoulos. 2000. An evaluation of Naive Bayesian anti-spam filtering. CoRR cs.CL/0006013 (2000). https://arxiv.org/abs/cs/0006013.Google ScholarGoogle Scholar
  2. [2] Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. 2015. Neural Machine Translation by Jointly Learning to Align and Translate. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Yoshua Bengio and Yann LeCun (Eds.). http://arxiv.org/abs/1409.0473.Google ScholarGoogle Scholar
  3. [3] Peter W. Battaglia, Jessica B. Hamrick, Victor Bapst, Alvaro Sanchez-Gonzalez, Vinícius Flores Zambaldi, Mateusz Malinowski, Andrea Tacchetti, David Raposo, Adam Santoro, Ryan Faulkner, Çaglar Gülçehre, H. Francis Song, Andrew J. Ballard, Justin Gilmer, George E. Dahl, Ashish Vaswani, Kelsey R. Allen, Charles Nash, Victoria Langston, Chris Dyer, Nicolas Heess, Daan Wierstra, Pushmeet Kohli, Matthew M. Botvinick, Oriol Vinyals, Yujia Li, and Razvan Pascanu. 2018. Relational inductive biases, deep learning, and graph networks. CoRR abs/1806.01261 (2018). arXiv:1806.01261 http://arxiv.org/abs/1806.01261.Google ScholarGoogle Scholar
  4. [4] Behera Ranjan Kumar, Sahoo Kshira Sagar, Naik Debadatt, Rath Santanu Kumar, and Sahoo Bibhudatta. 2021. Structural mining for link prediction using various machine learning algorithms. International Journal of Social Ecology and Sustainable Development 12, 3 (2021), 6678.Google ScholarGoogle ScholarCross RefCross Ref
  5. [5] Blei David and Lafferty John. 2006. Correlated topic models. In Proceedings of the 18th International Conference on Neural Information Processing Systems. 147.Google ScholarGoogle Scholar
  6. [6] Blei David M. and McAuliffe Jon D.. 2007. Supervised topic models. In Proceedings of the 20th International Conference on Neural Information Processing Systems. 121128.Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. [7] David M. Blei, Andrew Y. Ng, and Michael I. Jordan. 2003. Latent dirichlet allocation. J. Mach. Learn. Res. 3 (2003), 993–1022. http://jmlr.org/papers/v3/blei03a.html.Google ScholarGoogle Scholar
  8. [8] Cai Hongyun, Zheng Vincent W., and Chang Kevin Chen-Chuan. 2018. A comprehensive survey of graph embedding: Problems, techniques, and applications. IEEE Transactions on Knowledge and Data Engineering 30, 9 (2018), 16161637.Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. [9] Cheng Jiajun, Zhao Shenglin, Zhang Jiani, King Irwin, Zhang Xin, and Wang Hui. 2017. Aspect-level sentiment classification with heat (hierarchical attention) network. In Proceedings of the 2017 ACM on Conference on Information and Knowledge Management. 97106.Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. [10] Craven Mark, McCallum Andrew, PiPasquo Dan, Mitchell Tom, and Freitag Dayne. 1998. Learning to Extract Symbolic Knowledge from the World Wide Web. Technical Report. Carnegie-Mellon Univ Pittsburgh PA School of Computer Science.Google ScholarGoogle Scholar
  11. [11] Forman George. 2008. BNS feature scaling: An improved representation over TF-IDF for SVM text classification. In Proceedings of the 17th ACM Conference on Information and Knowledge Management. 263270.Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. [12] Huang Lianzhe, Ma Dehong, Li Sujian, Zhang Xiaodong, and Houfeng Wang. 2019. Text level graph neural network for text classification. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing. 34353441.Google ScholarGoogle ScholarCross RefCross Ref
  13. [13] Joachims Thorsten. 1998. Text categorization with support vector machines: Learning with many relevant features. In Proceedings of the 10th European Conference on Machine Learning. Springer, 137142.Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. [14] Kim Yoon. 2014. Convolutional neural networks for sentence classification. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing.Google ScholarGoogle Scholar
  15. [15] Kingma Diederik P. and Ba Jimmy. 2014. Adam: A method for stochastic optimization. InProceedings of the International Conference on Learning Representations.Google ScholarGoogle Scholar
  16. [16] Kipf Thomas N. and Welling Max. 2016. Semi-supervised classification with graph convolutional networks. In Proceedings of the International Conference on Learning Representations.Google ScholarGoogle Scholar
  17. [17] Li Shen, Zhao Zhe, Hu Renfen, Li Wensi, Liu Tao, and Du Xiaoyong. 2018. Analogical reasoning on Chinese morphological and semantic relations. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). 138143.Google ScholarGoogle ScholarCross RefCross Ref
  18. [18] Li Yujia, Tarlow Daniel, Brockschmidt Marc, and Zemel Richard. 2015. Gated graph sequence neural networks. In Proceedings of the International Conference on Learning Representations.Google ScholarGoogle Scholar
  19. [19] Li Zekun, Cui Zeyu, Wu Shu, Zhang Xiaoyu, and Wang Liang. 2019. Fi-GNN: Modeling feature interactions via graph neural networks for CTR prediction. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management. 539548.Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. [20] Li Zheng, Wei Ying, Zhang Yu, and Yang Qiang. 2018. Hierarchical attention transfer network for cross-domain sentiment classification. In Proceedings of the AAAI Conference on Artificial Intelligence. Vol. 32.Google ScholarGoogle ScholarCross RefCross Ref
  21. [21] Tingting Liang, Lei Zheng, Liang Chen, Yao Wan, Philip S. Yu, and Jian Wu. 2020. Multi-view factorization machines for mobile app recommendation based on hierarchical attention. Knowl. Based Syst. 187 (2020), 104821. Google ScholarGoogle ScholarCross RefCross Ref
  22. [22] Liefa Liao and Fugang Zhu Yalan Le. 2017. The application of LDA model in patent text classification. Journal of Modern Information 37, 3 (2017), 3539.Google ScholarGoogle Scholar
  23. [23] Lin Zhouhan, Feng Minwei, Santos Cicero Nogueira dos, Yu Mo, Xiang Bing, Zhou Bowen, and Bengio Yoshua. 2017. A structured self-attentive sentence embedding. InProceedings of the International Conference on Learning Representations.Google ScholarGoogle Scholar
  24. [24] Liu Pengfei, Qiu Xipeng, and Huang Xuanjing. 2016. Recurrent neural network for text classification with multi-task learning. In Proceedings of the 25th International Joint Conference on Artificial Intelligence. 28732879.Google ScholarGoogle Scholar
  25. [25] Mikolov Tomáš, Karafiát Martin, Burget Lukáš, Černockỳ Jan, and Khudanpur Sanjeev. 2010. Recurrent neural network based language model. In Proceedings of the 11th Annual Conference of the International Speech Communication Association.Google ScholarGoogle ScholarCross RefCross Ref
  26. [26] Nikolentzos Giannis, Tixier Antoine, and Vazirgiannis Michalis. 2020. Message passing attention networks for document understanding. In Proceedings of the AAAI Conference on Artificial Intelligence. Vol. 34, 85448551.Google ScholarGoogle ScholarCross RefCross Ref
  27. [27] Pavlinek Miha and Podgorelec Vili. 2017. Text classification method based on self-training and LDA topic models. Expert Systems with Applications 80, C (2017), 8393.Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. [28] Pennington Jeffrey, Socher Richard, and Manning Christopher D.. 2014. Glove: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing. 15321543.Google ScholarGoogle ScholarCross RefCross Ref
  29. [29] Rousseau François, Kiagias Emmanouil, and Vazirgiannis Michalis. 2015. Text categorization as a graph classification problem. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). 17021712.Google ScholarGoogle ScholarCross RefCross Ref
  30. [30] Md. Shahriare Satu, Md. Imran Khan, Mufti Mahmud, Shahadat Uddin, Matthew A. Summers, Julian M. W. Quinn, and Mohammad Ali Moni. 2021. TClustVID: A novel machine learning classification model to investigate topics and sentiment in COVID-19 tweets. Knowl. Based Syst. 226 (2021), 107126. Google ScholarGoogle ScholarCross RefCross Ref
  31. [31] Sun Qingying, Wang Zhongqing, Zhu Qiaoming, and Zhou Guodong. 2018. Stance detection with hierarchical attention network. In Proceedings of the 27th International Conference on Computational Linguistics. 23992409.Google ScholarGoogle Scholar
  32. [32] Tan Songbo. 2006. An effective refinement strategy for KNN text classifier. Expert Systems with Applications 30, 2 (2006), 290298.Google ScholarGoogle ScholarDigital LibraryDigital Library
  33. [33] Xu Nuo, Wang Pinghui, Chen Long, Pan Li, Wang Xiaoyan, and Zhao Junzhou. 2020. Distinguish confusing law articles for legal judgment prediction. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 30863095.Google ScholarGoogle ScholarCross RefCross Ref
  34. [34] Yadati Naganand. 2020. Neural message passing for multi-relational ordered and recursive hypergraphs. In Proceedings of the 34th International Conference on Neural Information Processing Systems. 33.Google ScholarGoogle Scholar
  35. [35] Yang Zichao, Yang Diyi, Dyer Chris, He Xiaodong, Smola Alex, and Hovy Eduard. 2016. Hierarchical attention networks for document classification. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 14801489.Google ScholarGoogle ScholarCross RefCross Ref
  36. [36] Yao Liang, Mao Chengsheng, and Luo Yuan. 2019. Graph convolutional networks for text classification. In Proceedings of the AAAI Conference on Artificial Intelligence. Vol. 33, 73707377.Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. [37] Ying Haochao, Zhuang Fuzhen, Zhang Fuzheng, Liu Yanchi, Xu Guandong, Xie Xing, Xiong Hui, and Wu Jian. 2018. Sequential recommender system based on hierarchical attention network. In Proceedings of the 27th International Joint Conference on Artificial Intelligence.Google ScholarGoogle ScholarDigital LibraryDigital Library
  38. [38] Zhang Tianyang, Huang Minlie, and Zhao Li. 2018. Learning structured representation for text classification via reinforcement learning. In Proceedings of the AAAI Conference on Artificial Intelligence. Vol. 32.Google ScholarGoogle ScholarCross RefCross Ref
  39. [39] Zhang Yufeng, Yu Xueli, Cui Zeyu, Wu Shu, Wen Zhongzhen, and Wang Liang. 2020. Every document Owns its structure: Inductive text classification via graph neural networks. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 334339.Google ScholarGoogle ScholarCross RefCross Ref
  40. [40] Zhang Zhao, Zhuang Fuzhen, Zhu Hengshu, Shi Zhiping, Xiong Hui, and He Qing. 2020. Relational graph neural network with hierarchical attention for knowledge graph completion. In Proceedings of the AAAI Conference on Artificial Intelligence. Vol. 34, 96129619.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Meta-Information Fusion of Hierarchical Semantics Dependency and Graph Structure for Structured Text Classification

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in

    Full Access

    • Published in

      cover image ACM Transactions on Knowledge Discovery from Data
      ACM Transactions on Knowledge Discovery from Data  Volume 17, Issue 2
      February 2023
      355 pages
      ISSN:1556-4681
      EISSN:1556-472X
      DOI:10.1145/3572847
      Issue’s Table of Contents

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 20 February 2023
      • Online AM: 17 May 2022
      • Accepted: 8 May 2022
      • Received: 7 August 2021
      Published in tkdd Volume 17, Issue 2

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Full Text

    View this article in Full Text.

    View Full Text

    HTML Format

    View this article in HTML Format .

    View HTML Format