Abstract
Text categorization, or text classification, is one of key tasks for representing the semantic information of documents. Traditional deep leaning models for text categorization are generally time-consuming with large-sized datasets due to slow convergence rate. In this paper, we propose a character-level model for short text classification with a combination of convolutional neural network (CNN), gated recurrent unit (GRU) and highway network (HN), which can capture both the global and the local textual semantics while having a tractable computational complexity. In addition, error minimization extreme learning machine (EM-ELM) is incorporated into the proposed model to improve the classification accuracy further. Extensive experiments show that our approach achieves the state-of-the-art performance when the hybrid model based on EM-ELM is trained using large-sized datasets.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
- 2.
- 3.
- 4.
- 5.
- 6.
Yahoo! Webscope program’s Yahoo! Answers Comprehensive Questions and Answers version 1.0 dataset.
References
Gong, Y., Ke, Q., Isard, M., Lazebnik, S.: A multi-view embedding space for modeling internet images, tags, and their semantics. Int. J. Comput. Vis. 106(2), 210–233 (2014)
Poria, S., Cambria, E., Bajpai, R., Hussain, A.: A review of affective computing: from unimodal analysis to multimodal fusion. Inf. Fusion. 37, 98–125 (2017)
Cambria, E.: Affective computing and sentiment analysis. IEEE Intell. Syst. 31(2), 102–107 (2016)
Zhang, Y., Wang, S., Phillips, P.: Binary PSO with mutation operator for feature selection using decision tree applied to spam detection. Knowl. Based Syst. 64(1), 22–31 (2014)
Poria, S., Cambria, E., Howard, N., Huang, G.-B., Hussain, A.: Fusing audio, visual and textual clues for sentiment analysis from multimodal content. Neurocomputing 174, 50–59 (2016)
Cambria, E., White, B.: Jumping NLP curves: a review of natural language processing research. IEEE Comput. Intell. Mag. 9(2), 48–57 (2014)
Collobert, R., Weston, J.: A unified architecture for natural language processing: deep neural networks with multitask learning. In: Proceedings of the 25th International Conference on Machine Learning, pp. 160–167. ACM (2008)
Kim, Y.: Convolutional neural networks for sentence classification. arXiv preprint arXiv:1408.5882 (2014)
Cambria, E., Fu, J., Bisio, F., Poria, S.: Affective Space 2: enabling affective intuition for concept-level sentiment analysis. In: AAAI, Austin, pp. 508–514 (2015)
He, T., Huang, W., Qiao, Y.: Text-attentional convolutional neural network for scene text detection. IEEE Trans. Image Process. 25(6), 25–29 (2016)
Ren, X., Zhou, Y., He, J.: A convolutional neural network based chinese text detection algorithm via text structure modeling. IEEE Trans. Multimedia 19, 506–518 (2017)
Chaturvedi, I., Ong, Y.-S., Tsang, I., Welsch, R., Cambria, E.: Learning word dependencies in text by means of a deep recurrent belief network. Knowl. Based Syst. 108, 144–154 (2016)
Cho, K., Van Merriënboer, B., Gulcehre, C., Bahdanau, D., Bougares, F., Schwenk, H., Bengio, Y.: Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv preprint arXiv:1406.1078 (2014)
Feng, G., Huang, G.B., Lin, Q.: Error minimized extreme learning machine with growth of hidden nodes and incremental learning. IEEE Trans. Neural Netw. 20(8), 1352–1357 (2009)
Majumder, N., Poria, S., Gelbukh, A., Cambria, E.: Deep learning based document modeling for personality detection from text. IEEE Intell. Syst. 32(2), 74–79 (2017)
Tang, D., Qin, B., Liu, T.: Document modeling with gated recurrent neural network for sentiment classification. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pp. 1422–1432 (2015)
Tai, K.S., Socher, R., Manning, C.D.: Improved semantic representations from tree-structured long short-term memory networks. arXiv preprint arXiv:1503.00075 (2015)
Zhang, M.-L., Zhou, Z.-H.: Multilabel neural networks with applications to functional genomics and text categorization. IEEE Trans. Knowl. Data Eng. 18(10), 1338–1351 (2006)
Nam, J., Kim, J., Mencía, E.L., Gurevych, I., Fürnkranz, J.: Largescale multi-label text classification—revisiting neural networks. In: Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pp. 437–452. Springer (2014)
Benites, F., Sapozhnikova, E.: HARAM: a hierarchical ARAM neural network for large-scale text classification. In: IEEE International Conference on Data Mining Workshop (ICDMW), pp. 847–854. IEEE (2015)
Nair, V., Hinton, G.E: Rectified linear units improve restricted Boltzmann machines. In: Proceedings of the 27th International Conference on Machine Learning (ICML 2010), pp. 807–814 (2010)
Dai, A.M., Le, Q.V.: Semi-supervised sequence learning. In: Advances in Neural Information Processing Systems, pp. 3079–3087 (2015)
Santos, C., Zadrozny, B.: Learning character-level representations for part-of-speech tagging. In: Proceedings of the 31st International Conference on Machine Learning, pp. 1818–1826 (2014)
Johnson, R., Zhang, T.: Effective use of word order for text categorization with convolutional neural networks. In: Proceeding of North American Chapter of the Association for Computational Linguistics, pp. 103–112 (2014)
Zhang, X, LeCun, Y.: Text Understanding from Scratch. Computation and Language, arXiv preprint arXiv:1502.01710 (2015)
Ling, W., Luís, T., Marujo, L.: Finding function in form: compositional character models for open vocabulary word representation. Computer Science, pp. 1899–1907 (2015)
Huang, Z., Xu, W., Yu, K.: Bidirectional LSTM-CRF models for sequence tagging. Computer Science (2015)
Srivastava, R.K., Greff, K., Schmidhuber, J.: Training very deep networks. Computer Science, pp. 1–5 (2015)
Kingma, D., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)
Acknowledgments
This work is supported by “the Fundamental Research Funds for the Central Universities” (NO. 2017XKQY082).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Liu, B., Zhou, Y., Sun, W. (2020). Character-Level Hybrid Convolutional and Recurrent Neural Network for Fast Text Categorization. In: Cao, J., Vong, C., Miche, Y., Lendasse, A. (eds) Proceedings of ELM 2018. ELM 2018. Proceedings in Adaptation, Learning and Optimization, vol 11. Springer, Cham. https://doi.org/10.1007/978-3-030-23307-5_12
Download citation
DOI: https://doi.org/10.1007/978-3-030-23307-5_12
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-23306-8
Online ISBN: 978-3-030-23307-5
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)