skip to main content
10.1145/3581807.3581853acmotherconferencesArticle/Chapter ViewAbstractPublication PagesiccprConference Proceedingsconference-collections
research-article

Multi-Channel Text Classification Model Based on ERNIE

Published:22 May 2023Publication History

ABSTRACT

Aiming at the large amount of news and review text data, sparse features, and the inability of traditional text feature representation to dynamically obtain grammatical structure, semantic information, and multi-dimensional rich feature representation of entity phrases. This paper proposes to obtain more generalized knowledge semantic feature information such as rich context phrases, entity words and so on by integrating knowledge enhanced semantic representation (Enhanced Representation Through Knowledge Integration, ERNIE). The pre-trained language model ERNIE hides words and entities by random Semantic unit prediction context realizes word vector language representation, and the output vector representation of ERNIE is input to BiLSTM, Attention mechanism and DPCNN network model to generate high-order text feature vectors, and each channel vector is processed by BatchNormalization and ReLU activation functions respectively.Thus, the semantic description information of the multi-channel word vector is fused. The model proposed in this paper can not only improve the training speed and prevent overfitting, but also enhance the feature information such as semantics and grammatical structure, thereby improving the text classification effect. By comparing the two datasets with other improved ERNIE models in terms of accuracy, precision, recall, and F1 value, the experimental results show that the model proposed in this paper can obtain multi-dimensional rich semantic grammatical structure features for text classification, and then improve the Text classification effect.

References

  1. Li M, Wang L. A survey on personalized news recommendation technology[J]. IEEE Access, 2019, 7: 145861-145879.Google ScholarGoogle ScholarCross RefCross Ref
  2. Mitra B, Craswell N. Neural models for information retrieval[J]. arXiv preprint arXiv:1705.01509, 2017.Google ScholarGoogle Scholar
  3. Shrawan Kumar Trivedi,Prabin Kumar Panigrahi. Spam classification: a comparative analysis of different boosted decision tree approaches[J]. Journal of Systems and Information Technology,2018,20(3):Google ScholarGoogle Scholar
  4. Mukherjee S. Sentiment analysis[M]//ML. NET Revealed. Apress, Berkeley, CA, 2021: 113-127.Google ScholarGoogle Scholar
  5. Shah K, Patel H, Sanghvi D, A comparative analysis of logistic regression, random forest and KNN models for the text classification[J]. Augmented Human Research, 2020, 5(1): 1-16.Google ScholarGoogle ScholarCross RefCross Ref
  6. Ruan S, Li H, Li C, Class-specific deep feature weighting for Naïve Bayes text classifiers[J]. IEEE Access, 2020, 8: 20151-20159.Google ScholarGoogle ScholarCross RefCross Ref
  7. Y. Bengio,R. Ducharme,P. Vincent. A Neural Probabilistic Language Model[J] Journal of Machine Learning Research,2003,3:1137- 1155.Google ScholarGoogle Scholar
  8. Mikolov T, Chen K, Corrado G, et al.Efficient estimation of word representations in vector space [EB/OL] ( 2013-09-07) [2017-05-08] https: / /arxiv.org /abs/1301.3781.Google ScholarGoogle Scholar
  9. Mikolov T, Sutskever I, Chen K, et al. Distributed representations of words and phrases and their compositionality [C ]/ /MIT. Advances in Neural Information Processing Systems. Massachusetts: MIT Press, 2013: 3111-3119.Google ScholarGoogle Scholar
  10. PENNINGTON J, SOCHER R, MANNING C D. Glove: global vectors for word representation[C]//Proceedings of the 2014 conference on empirical methods in natural language processing, Stroudsburg, PA: Association for Computational Linguistics, 2014.Google ScholarGoogle Scholar
  11. PETERS M, NEUMANN M, IYYER M, Deep contextualized word representations[C]//Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Stroudsburg, PA: Association for Computational Linguistics, 2018.Google ScholarGoogle Scholar
  12. VASWANI A, SHAZEER N, PARMAR N, Attention is all you need[C]//Proceedings of the 2017 conference on neural information processing systems, Stroudsburg, PA: Association for Computational Linguistics,2017, 5998-6008.Google ScholarGoogle Scholar
  13. DEVLIN J, CHANG M W, LEE K, BERT: pre-training of deep bidirectional transformers for language understanding[C]//Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Stroudsburg, PA: Association for Computational Linguistics,2019.Google ScholarGoogle Scholar
  14. Sun Y, Wang S, Li Y, Ernie: Enhanced representation through knowledge integration[J]. arXiv preprint arXiv:1904.09223, 2019.Google ScholarGoogle Scholar
  15. KIM Y.Convolutional neural networks for sentence classification [J]. arXiv preprint arXiv: 1408.5882, 2014.Google ScholarGoogle Scholar
  16. Jordan M I. Serial order: A parallel distributed processing approach[M]//Advances in psychology. North-Holland, 1997, 121: 471-495.Google ScholarGoogle Scholar
  17. Liu P, Qiu X, Huang X. Recurrent neural network for text classification with multi-task learning[J]. arXiv preprint arXiv:1605.05101, 2016.Google ScholarGoogle Scholar
  18. Chen Jie, Ma Jing, Li Xiaofeng. A short text classification method incorporating text features of pre-training models [J]. Data Analysis and Knowledge Discovery, 2021,5(09):21-30.Google ScholarGoogle Scholar
  19. Shuwei X, Xuyang G, Ying W. A Study on the EDA-based Classification of News Text[C]//Journal of Physics: Conference Series. IOP Publishing, 2021, 1792(1): 012080.Google ScholarGoogle ScholarCross RefCross Ref
  20. Guo H, Chi C, Zhan X. ERNIE-BiLSTM Based Chinese Text Sentiment Classification Method[C]//2021 International Conference on Computer Engineering and Application (ICCEA). IEEE, 2021: 84-88.Google ScholarGoogle Scholar
  21. Li J, Cao H. Research on Dual Channel News Headline Classification Based on ERNIE Pre-training Model[J]. arXiv preprint arXiv:2202.06600, 2022.Google ScholarGoogle Scholar
  22. Lai S, Xu L, Liu K, Recurrent convolutional neural networks for text classification[C]//Twenty-ninth AAAI conference on artificial intelligence. 2015.Google ScholarGoogle Scholar
  23. Johnson R, Zhang T. Deep pyramid convolutional neural networks for text categorization[C]//Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2017: 562-570.Google ScholarGoogle Scholar

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Other conferences
    ICCPR '22: Proceedings of the 2022 11th International Conference on Computing and Pattern Recognition
    November 2022
    683 pages
    ISBN:9781450397056
    DOI:10.1145/3581807

    Copyright © 2022 ACM

    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 22 May 2023

    Permissions

    Request permissions about this article.

    Request Permissions

    Check for updates

    Qualifiers

    • research-article
    • Research
    • Refereed limited
  • Article Metrics

    • Downloads (Last 12 months)34
    • Downloads (Last 6 weeks)7

    Other Metrics

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format .

View HTML Format