Abstract
The rapid development of social media has brought the prosperity of online economy. Recently, product promotion in social networks has become an essential way of online marketing. As one of the most common marketing means, Content Marketing (CM) inserts advertisements into regular articles in a roundabout and covert way. However, the values and characteristics of products are often exaggerated to attract users’ attention. It could cause severe economic losses to users and influence the creditworthiness of the platforms. In this paper, we model the problem of advertisement extraction from CM articles as a sentence classification task. We propose a topic-enhanced deep neural network to encode the semantic information of a sentence for classification. Motivated by the characteristics of CM articles, we develop a segment-aware optimization method that considers the label transitions of sentences in different segments of an article to improve the performance of the classifier. Experimental results based on real-world datasets demonstrate the superiority of the proposed method over state-of-the-art approaches.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Most popular social networks worldwide as of January (2021). https://www.statista.com/statistics/272014/global-social-networks-ranked-by-number-of-users/
Zhang, X., Zhao, J., LeCun, Y.: Character-level convolutional networks for text classification. Adv. Neural Inf. Process. Syst. 28, 649–657 (2015)
Hofmann, T.: Probabilistic latent semantic indexing. In: Proceedings of the 22nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 50–57 (1999)
Blei, D.M., Ng, A.Y., Jordan, M.I.: Latent dirichlet allocation. J. Mach. Learn. Res. 3, 993–1022 (2003)
Hong, L., Davison, B.D.: Empirical study of topic modeling in twitter. In: Proceedings of the First Workshop on Social Media Analytics, pp. 80–88 (2010)
Yan, X., Guo, J., Lan, Y., Cheng, X.: A biterm topic model for short texts, In: Proceedings of the 22nd International Conference on World Wide Web, pp. 1445–1456 (2013)
Yao, L., Mao, C., Luo, Y.: Graph convolutional networks for text classification. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, pp. 7370–7377 (2019)
Wieting, J., Kiela, D.: No training required: Exploring random encoders for sentence classification. arXiv:1901.10444 (2019)
Liang, X., Wang, C., Zhao, G.: Enhancing content marketing article detection with graph analysis. IEEE Access 7, 94869–94881 (2019)
Hao, Y., et al.: An end-to-end model for question answering over knowledge base with cross-attention combining global knowledge. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Long Papers), vol. 1, pp. 221–231 (2017)
Luong, M.-T., Pham, H., Manning, C. D.: Effective approaches to attention-based neural machine translation. arXiv:1508.04025 (2015)
Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv:1412.6980 (2014)
Sinha, K., Dong, Y., Cheung, J.C.K., Ruths, D.: A hierarchical neural attention-based text classifier, In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 817–823 (2018)
Liu, P., Qiu, X., Huang, X.: Recurrent neural network for text classification with multi-task learning, 2873–2879 (2016)
Joulin, A., Grave, E., Bojanowski, P., Mikolov, T.: Bag of tricks for efficient text classification. arXiv preprint arXiv:1607.01759 (2016)
Luo, Y.: Recurrent neural networks for classifying relations in clinical notes. J. Biomed. Inf. 72, 85–95 (2017)
Wang, J., Wang, Z., Zhang, D., Yan, J.: Combining knowledge with deep convolutional neural networks for short text classification., In: IJCAI, vol. 350, pp. 2915–2921 (2017)
Fan, X., Wang, C., Liang, X.: Extracting advertisements from content marketing articles based on topiccnn. In: 2020 IEEE International Conference on Dependable, Autonomic and Secure Computing, IEEE International Conferenceon Pervasive Intelligence and Computing, IEEE International Conferenceon Cloud and Big Data Computing, IEEE International Conferenceon Cyber Science and Technology Congress (DASC/PiCom/CBDCom/CyberSciTech), pp. 355–360. IEEE (2020)
Linmei, H., Yang, T., Shi, C., Ji, H., Li, X.: Heterogeneous graph attention networks for semi-supervised short text classification. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 4823–4832 (2019)
Lai, S., Xu, L., Liu, K., Zhao, J.: Recurrent convolutional neural networks for text classification. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 29, pp. 2267–2273 (2015)
Angelov, D.: Top2Vec: distributed representations of topics. arXiv:2008.09470 (2020)
Kim, Y.: Convolutional neural networks for sentence classification. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, pp. 1746–1751 (2014)
Acknowledgment
The research presented in this paper is supported in part by National Natural Science Foundation of China (No. 61602370, U1736205, 61922067, 61902305), Shenzhen Basic Research Grant (JCYJ20170816100819428), Natural Science Basic Research Plan in Shaanxi Province (2021JM-018).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Fan, X., Wang, C. (2021). Advertisement Extraction from Content Marketing Articles via Segment-Aware Sentence Classification. In: Wang, L., Feng, Y., Hong, Y., He, R. (eds) Natural Language Processing and Chinese Computing. NLPCC 2021. Lecture Notes in Computer Science(), vol 13028. Springer, Cham. https://doi.org/10.1007/978-3-030-88480-2_50
Download citation
DOI: https://doi.org/10.1007/978-3-030-88480-2_50
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-88479-6
Online ISBN: 978-3-030-88480-2
eBook Packages: Computer ScienceComputer Science (R0)