Abstract
Human experts can often easily write a set of approximate rules based on their domain knowledge for supporting automatic text classification. While such approximate rules are able to conduct classification at a general level, they are not effective for handling diverse and specific situations for a particular category. Given a set of approximate rules and a moderate amount of labeled data, existing incremental text classification learning models can be employed for tackling this problem by continuous rule refinement. However, these models lack the consideration of context information, which inherently exists in data. We propose a framework comprising rule embeddings and context embeddings derived from data to enhance the adaptability of approximate rules via considering the context information. We conduct extensive experiments and the results demonstrate that our proposed framework performs better than existing models in some benchmarking datasets, indicating that learning the context of rules is constructive for improving text classification performance.
Keywords
The work described in this paper is substantially supported by a grant from the Direct Grant of the Faculty of Engineering, CUHK (Project Code: 4055093).
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Bostanci, B., Bostanci, E.: An evaluation of classification algorithms using Mc Nemar’s test. In: Bansal, J., Singh, P., Deep, K., Pant, M., Nagar, A. (eds.) BIC-TA 2012. AISC, vol. 201, pp. 15–26. Springer, India (2013). https://doi.org/10.1007/978-81-322-1038-2_2
Cohen, W.W.: Fast effective rule induction. In: Twelfth International Conference on Machine Learning (ICML), pp. 115–123 (1995)
Cohen, W.W., Singer, Y.: Context-sensitive learning methods for text categorization. ACM Trans. Inf. Syst. (TOIS) 17(2), 141–173 (1999)
Elwell, R., Polikar, R.: Incremental learning of concept drift in nonstationary environments. IEEE Trans. Neural Netw. 22(10), 1517–1531 (2011)
Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)
Muhlbaier, M.D., Topalis, A., Polikar, R.: Learn\({++}\). NC: combining ensemble of classifiers with dynamically weighted consult-and-vote for efficient incremental learning of new classes. IEEE Trans. Neural Netw. 20(1), 152–168 (2009)
Paszke, A., et al.: Automatic differentiation in PyTorch. In: Advances in Neural Information Processing Systems (NIPS) - Workshop (2017)
Polikar, R., Upda, L., Upda, S.S., Honavar, V.: Learn++: an incremental learning algorithm for supervised neural networks. IEEE Trans. Syst. Man Cybern. part C (Appl. Rev.) 31(4), 497–508 (2001)
Zhang, X., Zhao, J., LeCun, Y.: Character-level convolutional networks for text classification. In: Advances in Neural Information Processing Systems (NIPS), pp. 649–657 (2015)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Switzerland AG
About this paper
Cite this paper
Wong, W.C., Lai, S., Lam, W., Leung, K.S. (2018). Guiding Approximate Text Classification Rules via Context Information. In: Tseng, YH., et al. Information Retrieval Technology. AIRS 2018. Lecture Notes in Computer Science(), vol 11292. Springer, Cham. https://doi.org/10.1007/978-3-030-03520-4_13
Download citation
DOI: https://doi.org/10.1007/978-3-030-03520-4_13
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-03519-8
Online ISBN: 978-3-030-03520-4
eBook Packages: Computer ScienceComputer Science (R0)