skip to main content
research-article

An Integrated Topic Modelling and Graph Neural Network for Improving Cross-lingual Text Classification

Published:25 November 2022Publication History
Skip Abstract Section

Abstract

In recent years, along with the dramatic developments of deep learning in the natural language processing (NLP) domain, notable multilingual pre-trained language techniques have been proposed. These recent multilingual text analysis and mining models have demonstrated state-of-the-art performance in several primitive NLP tasks, including cross-lingual text classification (CLC). However, these recent multilingual pre-trained language models still suffer limitations regarding their adaptation for specific task-driven fine-tuning in the context of low-resource languages. Moreover, they also encounter problems related to the capability of preserving the global semantic (e.g., topic, etc.) and long-range relationships between words to better fine-tune and effectively handle the cross-lingual text classification task. To meet these challenges, in this article, we propose a novel topic-driven multi-typed text graph attention–based representation learning method for dealing with the cross-lingual text classification problem called TG-CTC. In the proposed TG-CTC model, we utilize a novel fused topic-driven multi-typed text graph representation to jointly learn the rich-schematic structural and global semantic information of texts to effectively handle the CLC task. More specifically, we integrate the heterogeneous text graph attention network with the neural topic modelling approach to enrich the semantic information of learned textual representations in the context of multiple languages. Extensive experiments in benchmark multilingual datasets showed the effectiveness of the proposed TG-CTC model compared with the contemporary state-of-the-art baselines.

REFERENCES

  1. [1] Nakamura T., Shirakawa M., Hara T., and Nishio S.. 2018. Wikipedia-based relatedness measurements for multilingual short text clustering. ACM Trans. Asian Low-Resour. Lang. Inf. Process. 18, 2 (2018), 125.Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. [2] Xiang Y., Yu Z., Guo J., Huang Y., and Xian Y.. 2021. Event graph neural network for opinion target classification of microblog comments. Trans. Asian Low-Resour. Lang. Inf. Process. 21, 1 (2021), 113.Google ScholarGoogle Scholar
  3. [3] De A., Bandyopadhyay D., Gain B., and Ekbal A.. 2021. A transformer-based approach to multilingual fake news detection in low-resource languages. Trans. Asian Low-Resour. Lang. Inf. Process. 21, 1 (2021), 120.Google ScholarGoogle Scholar
  4. [4] Wan X.. Co-training for cross-lingual sentiment classification. In Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP.Google ScholarGoogle Scholar
  5. [5] Duh K., Fujino A., and Nagata M.. 2011. Is machine translation ripe for cross-lingual sentiment classification? In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies.Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. [6] Shi L., Mihalcea R., and Tian M.. 2010. Cross language text classification by model translation and semi-supervised learning. In Proceedings of the Conference on Empirical Methods in Natural Language Processing.Google ScholarGoogle Scholar
  7. [7] Andrade D., Sadamasa K., Tamura A., and Tsuchida M.. 2015. Cross-lingual text classification using topic-dependent word probabilities. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies.Google ScholarGoogle ScholarCross RefCross Ref
  8. [8] Xu R., Yang Y., Liu H., and Hsi A.. 2016. Cross-lingual text classification via model translation with limited dictionaries. In Proceedings of the 25th ACM International on Conference on Information and Knowledge Management.Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. [9] Ziser Y. and Reichart R.. 2018. Deep pivot-based modeling for cross-language cross-domain transfer with minimal guidance. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing.Google ScholarGoogle ScholarCross RefCross Ref
  10. [10] Chen X., Sun Y., Athiwaratkun B., Cardie C., and Weinberger K.. 2018. Adversarial deep averaging networks for cross-lingual sentiment classification. Trans. Assoc. Comput. Ling. 6 (2018), 557570.Google ScholarGoogle ScholarCross RefCross Ref
  11. [11] Peters M. E., Neumann M., Iyyer M., Gardner M., Clark C., Lee K., and Zettlemoyer L.. 2018. Deep contextualized word representations. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies.Google ScholarGoogle ScholarCross RefCross Ref
  12. [12] Radford A., Narasimhan K., Salimans T., and Sutskever I.. 2018. Improving language understanding by generative pre-training. OpenAI.Google ScholarGoogle Scholar
  13. [13] Devlin J., Chang M. W., Lee K., and Toutanova K.. 2019. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies.Google ScholarGoogle Scholar
  14. [14] Hamilton W. L., Ying R., and Leskovec J.. 2017. Inductive representation learning on large graphs. In Proceedings of the 31st International Conference on Neural Information Processing Systems.Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. [15] Kipf T. N. and Welling M.. 2016. Semi-supervised classification with graph convolutional networks. In 5th International Conference on Learning Representations (ICLR'17). Google ScholarGoogle Scholar
  16. [16] Veličković P., Cucurull G., Casanova A., Romero A., Lio P., and Bengio Y.. 2018. Graph attention networks. In Proceedings of the International Conference on Learning Representations (ICLR’18).Google ScholarGoogle Scholar
  17. [17] Yao L., Mao C., and Luo Y.. 2019. Graph convolutional networks for text classification. In Proceedings of the AAAI Conference on Artificial Intelligence.Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. [18] Liu X., You X., Zhang X., Wu J., and Lv P.. 2020. Tensor graph convolutional networks for text classification. In Proceedings of the AAAI Conference on Artificial Intelligence.Google ScholarGoogle ScholarCross RefCross Ref
  19. [19] Zhang H. and Zhang J.. 2020. Text graph transformer for document classification. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP’20).Google ScholarGoogle ScholarCross RefCross Ref
  20. [20] Kherwa P. and Bansal P.. 2020. Topic modeling: a comprehensive review. EAI Endors. Trans. Scal. Inf. Syst. 7, 24, (2020).Google ScholarGoogle Scholar
  21. [21] Lin L., Jiang H., and Rao Y.. 2020. Copula guided neural topic modelling for short texts. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval.Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. [22] Wang R., Hu X., Zhou D., He Y., Xiong Y., Ye C., and Xu H.. 2020. Neural topic modeling with bidirectional adversarial training. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.Google ScholarGoogle ScholarCross RefCross Ref
  23. [23] Peinelt N., Nguyen D., and Liakata M.. 2020. tBERT: Topic models and BERT joining forces for semantic similarity detection. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.Google ScholarGoogle ScholarCross RefCross Ref
  24. [24] Song X., Petrak J., Jiang Y., Singh I., Maynard D., and Bontcheva K.. 2021. Classification aware neural topic model for COVID-19 disinformation categorisation. PloS One 16, 2 (2021), e0247086.Google ScholarGoogle ScholarCross RefCross Ref
  25. [25] Miao Y., Grefenstette E., and Blunsom P.. 2017. Discovering discrete latent topics with neural variational inference. In Proceedings of the International Conference on Machine Learning (PMLR’17).Google ScholarGoogle Scholar
  26. [26] Srivastava A., and Sutton C.. 2017. Autoencoding variational inference for topic models. In Proceedings of the 5th International Conference on Learning Representations (ICLR).Google ScholarGoogle Scholar
  27. [27] Conneau A., Khandelwal K., Goyal N., Chaudhary V., Wenzek G., Guzmán F., and Stoyanov V.. 2020. Unsupervised cross-lingual representation learning at scale. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.Google ScholarGoogle ScholarCross RefCross Ref
  28. [28] Wang Z., Liu X., Yang P., Liu S., and Wang Z.. 2021. Cross-lingual text classification with heterogeneous graph neural network. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 2: Short Papers). 612--620.Google ScholarGoogle ScholarCross RefCross Ref
  29. [29] Fei H. and Li P.. 2020. Cross-lingual unsupervised sentiment classification with multi-view transfer learning. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.Google ScholarGoogle ScholarCross RefCross Ref
  30. [30] Qin L., Ni M., Zhang Y., and Che W.. 2020. Cosda-ml: Multi-lingual code-switching data augmentation for zero-shot cross-lingual nlp. In Proceedings of the Twenty-Ninth International Conference on International Joint Conferences on Artificial Intelligence. 3853--3860.Google ScholarGoogle Scholar
  31. [31] Ruder S., Vulić I., and Søgaard A.. 2019. A survey of cross-lingual word embedding models. J. Artif. Intell. Res. 65 (2019) 569631.Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. [32] Karamanolakis G., Hsu D., and Gravano L.. 2020. Cross-lingual text classification with minimal resources by transferring a sparse teacher. In Findings of the Association for Computational Linguistics: EMNLP 2020. 3604--3622.Google ScholarGoogle ScholarCross RefCross Ref
  33. [33] Zhang M., Fujinuma Y., and Boyd-Graber J.. 2020. Exploiting cross-lingual subword similarities in low-resource document classification. In Proceedings of the AAAI Conference on Artificial Intelligence.Google ScholarGoogle ScholarCross RefCross Ref
  34. [34] Yuan M., Zhang M., Van Durme B., Findlater L., and Boyd-Graber J.. 2020. Interactive refinement of cross-lingual word embeddings. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP’20).Google ScholarGoogle ScholarCross RefCross Ref
  35. [35] Zhang M., Fujinuma Y., Paul M. J., and Boyd-Graber J.. 2020. Why overfitting isn't always bad: Retrofitting cross-lingual word embeddings to dictionaries. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 2214--2220.Google ScholarGoogle ScholarCross RefCross Ref
  36. [36] Conneau A. and Lample G.. 2019. Cross-lingual language model pretraining. In Advances in Neural Information Processing Systems.Google ScholarGoogle Scholar
  37. [37] Hu J., Ruder S., Siddhant A., Neubig G., Firat O., and Johnson M.. 2020. Xtreme: A massively multilingual multi-task benchmark for evaluating cross-lingual generalisation. In Proceedings of the International Conference on Machine Learning (PMLR).Google ScholarGoogle Scholar
  38. [38] Mimno D., Wallach H., Naradowsky J., Smith D. A., and McCallum A.. 2009. Polylingual topic models. In Proceedings of the Conference on Empirical Methods in Natural Language Processing.Google ScholarGoogle ScholarCross RefCross Ref
  39. [39] Yuan M., Van Durme D., and Ying J. L.. 2018. Multilingual anchoring: interactive topic modeling and alignment across languages. In Proceedings of the Conference and Workshop on Neural Information Processing Systems (NeurIPS’18), 86678677.Google ScholarGoogle Scholar
  40. [40] Yang W., Boyd-Graber J., and Resnik P.. 2019. A multilingual topic model for learning weighted topic links across corpora with low comparability. In Proceedings of the Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP’19).Google ScholarGoogle ScholarCross RefCross Ref
  41. [41] Li Z., Kumar M., Headden W., Yin B., Wei Y., Zhang Y., and Yang Q.. 2020. Learn to cross-lingual transfer with meta graph learning across heterogeneous languages. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP’20).Google ScholarGoogle ScholarCross RefCross Ref
  42. [42] Hochreiter S. and Schmidhuber J.. 1997. Long short-term memory. Neur. Comput. 9, 8 (1997), 17351780.Google ScholarGoogle ScholarDigital LibraryDigital Library
  43. [43] Graves A., Fernández S., and Schmidhuber J.. 2005. Bidirectional LSTM networks for improved phoneme classification and recognition. In Proceedings of the International Conference on Artificial Neural Networks.Google ScholarGoogle ScholarCross RefCross Ref
  44. [44] Manning C. D., Surdeanu M., Bauer J., Finkel J. R., Bethard S., and McClosky D.. 2014. The stanford CORENLP natural language processing toolkit. In Proceedings of 52nd Annual Meeting of the Association for Computational Linguistics: System Demonstrations.Google ScholarGoogle ScholarCross RefCross Ref
  45. [45] Linmei H., Yang T., Shi C., Ji H., and Li X.. 2019. Heterogeneous graph attention networks for semi-supervised short text classification. In Proceedings of the Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP’19).Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. An Integrated Topic Modelling and Graph Neural Network for Improving Cross-lingual Text Classification

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in

    Full Access

    • Published in

      cover image ACM Transactions on Asian and Low-Resource Language Information Processing
      ACM Transactions on Asian and Low-Resource Language Information Processing  Volume 22, Issue 1
      January 2023
      340 pages
      ISSN:2375-4699
      EISSN:2375-4702
      DOI:10.1145/3572718
      Issue’s Table of Contents

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 25 November 2022
      • Online AM: 14 April 2022
      • Accepted: 3 April 2022
      • Revised: 16 December 2021
      • Received: 12 July 2021
      Published in tallip Volume 22, Issue 1

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article
      • Refereed

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Full Text

    View this article in Full Text.

    View Full Text

    HTML Format

    View this article in HTML Format .

    View HTML Format