Abstract
Incorporating personalization into document-level sentiment classification has gained considerable attention due to its better performance on diverse domains. Current progress in this field is attributed to the developed mechanisms of effectively modeling the interaction among the three fundamental factors: users, items, and words. However, how to improve the representation learning of the three factors themselves is largely unexplored. To bridge this gap, we propose to enrich users, items, and words representations in the state-of-the-art personalized sentiment classification model with an end-to-end training fashion. Specifically, relations between users and items are respectively modeled by graph neural networks to enhance original user and item representations. We further promote word representation by utilizing powerful pre-trained language models. Comprehensive experiments on several public and widely-used datasets demonstrate the superiority of the proposed approach, validating the contribution of the improved representations.
This work was supported in part by the National Key Research and Development Program (2019YFB2102600), NSFC (61702190), Shanghai Sailing Program (17YF1404500), and Zhejiang Lab (2019KB0AB04).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Amplayo, R.K.: Rethinking attribute representation and injection for sentiment classification. In: EMNLP, pp. 5601–5612 (2019)
Amplayo, R.K., Kim, J., Sung, S., Hwang, S.: Cold-start aware user and product attention for sentiment classification. In: ACL, pp. 2535–2544 (2018)
Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate. In: ICLR (2015)
Chen, H., Sun, M., Tu, C., Lin, Y., Liu, Z.: Neural sentiment classification with user and product attention. In: EMNLP, pp. 1650–1659 (2016)
Devlin, J., Chang, M., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: NAACL, pp. 4171–4186 (2019)
Dou, Z.: Capturing user and product information for document level sentiment analysis with deep memory network. In: EMNLP, pp. 521–526 (2017)
Feng, F., He, X., Wang, X., Luo, C., Liu, Y., Chua, T.: Temporal relational ranking for stock prediction. ACM Trans. Inf. Syst. 37(2), 27:1–27:30 (2019)
Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)
Hu, G., Dai, X., Song, Y., Huang, S., Chen, J.: A synthetic approach for recommendation: Combining ratings, social relations, and reviews. In: IJCAI, pp. 1756–1762 (2015)
Jiang, L., Yu, M., Zhou, M., Liu, X., Zhao, T.: Target-dependent Twitter sentiment classification. In: ACL, pp. 151–160 (2011)
Kim, Y.: Convolutional neural networks for sentence classification. In: EMNLP, pp. 1746–1751 (2014)
Kipf, T.N., Welling, M.: Semi-supervised classification with graph convolutional networks. In: ICLR (2017)
Li, F., Wang, S., Liu, S., Zhang, M.: SUIT: a supervised user-item based topic model for sentiment analysis. In: AAAI, pp. 1636–1642 (2014)
Li, Y., Cai, Y., Leung, H., Li, Q.: Improving short text modeling by two-level attention networks for sentiment classification. In: DASFAA, pp. 878–890 (2018)
Liu, B.: Sentiment Analysis - Mining Opinions, Sentiments, and Emotions. Cambridge University Press, Cambridge (2015)
McAuley, J.J., Leskovec, J.: Hidden factors and hidden topics: understanding rating dimensions with review text. In: RecSys, pp. 165–172 (2013)
Mikolov, T., Sutskever, I., Chen, K., Corrado, G.S., Dean, J.: Distributed representations of words and phrases and their compositionality. In: NIPS, pp. 3111–3119 (2013)
Mukherjee, S., Basu, G., Joshi, S.: Joint author sentiment topic model. In: SDM, pp. 370–378 (2014)
Ni, J., Li, J., McAuley, J.J.: Justifying recommendations using distantly-labeled reviews and fine-grained aspects. In: EMNLP, pp. 188–197 (2019)
Pennington, J., Socher, R., Manning, C.D.: Glove: global vectors for word representation. In: EMNLP, pp. 1532–1543 (2014)
Peters, M.E., et al.: Deep contextualized word representations. In: NAACL, pp. 2227–2237 (2018)
Taboada, M., Brooke, J., Tofiloski, M., Voll, K., Stede, M.: Lexicon-based methods for sentiment analysis. Comput. Linguist. 37(2), 267–307 (2011)
Tang, D., Qin, B., Liu, T.: Document modeling with gated recurrent neural network for sentiment classification. In: EMNLP, pp. 1422–1432 (2015)
Tang, D., Qin, B., Liu, T.: Learning semantic representations of users and products for document level sentiment classification. In: ACL, pp. 1014–1023 (2015)
Vaswani, A., et al.: Attention is all you need. In: NIPS, pp. 5998–6008 (2017)
Velickovic, P., Cucurull, G., Casanova, A., Romero, A., Liò, P., Bengio, Y.: Graph attention networks. In: ICLR (2018)
Wang, X., He, X., Wang, M., Feng, F., Chua, T.: Neural graph collaborative filtering. In: SIGIR, pp. 165–174 (2019)
Wu, Z., Dai, X., Yin, C., Huang, S., Chen, J.: Improving review representations with user attention and product attention for sentiment classification. In: AAAI, pp. 5989–5996 (2018)
Yang, Z., Yang, D., Dyer, C., He, X., Smola, A.J., Hovy, E.H.: Hierarchical attention networks for document classification. In: NAACL, pp. 1480–1489 (2016)
Yao, L., Mao, C., Luo, Y.: Graph convolutional networks for text classification. In: AAAI, pp. 7370–7377 (2019)
Zhang, W., Yuan, Q., Han, J., Wang, J.: Collaborative multi-level embedding learning from reviews for rating prediction. In: IJCAI, pp. 2986–2992 (2016)
Zhao, K., Zhang, Y., Zhang, Y., Xing, C., Li, C.: Learning from user social relation for document sentiment classification. In: DASFAA, pp. 86–103 (2019)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Zhang, Y., Zhang, W. (2020). Improved Representations for Personalized Document-Level Sentiment Classification. In: Nah, Y., Cui, B., Lee, SW., Yu, J.X., Moon, YS., Whang, S.E. (eds) Database Systems for Advanced Applications. DASFAA 2020. Lecture Notes in Computer Science(), vol 12112. Springer, Cham. https://doi.org/10.1007/978-3-030-59410-7_53
Download citation
DOI: https://doi.org/10.1007/978-3-030-59410-7_53
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-59409-1
Online ISBN: 978-3-030-59410-7
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)