Skip to main content

Net2Text: An Edge Labelling Language Model for Personalized Review Generation

  • Conference paper
  • First Online:

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 11446))

Abstract

Writing an item review for online shopping or sharing the dining experience of a restaurant has become major Internet activities of young people. This kind of review system could not only help users express and exchange experience but also prompt business to improve service quality. Instead of taking time to type in the review, we would like to make the review process more automated. In this work, we study an edge labelling language model for personalized review generation, e.g., the problem of generating text (e.g., a review) on the edges of the network (e.g., online shopping). It is related to both network structure and rich text semantic information. Previously, link prediction models have been applied to recommender system and event prediction. However, they could not migrate to text generation on the edges of networks since most of them are numerical prediction or tag labelling tasks. To bridge the gap between link prediction and natural language generation, in this paper, we propose a model called Net2Text, which can simultaneously learn the structural information in the network and build a language model over text on the edges. The performance of Net2Text is demonstrated in our experiments, showing that our model performs better than other baselines, and is able to produce reasonable reviews between users and items.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Notes

  1. 1.

    https://www.amazon.com/.

  2. 2.

    https://www.yelp.com/.

  3. 3.

    https://www.yelp.com/dataset/challenge.

References

  1. Aiello, L.M., Barrat, A., Schifanella, R., Cattuto, C., Markines, B., Menczer, F.: Friendship prediction and homophily in social media. ACM Trans. Web 6(2), 9 (2012)

    Article  Google Scholar 

  2. Almansoori, W., et al.: Link prediction and classification in social networks and its application in healthcare and systems biology. Netw. Model. Anal. Health Inf. Bioinf. 1(1–2), 27–36 (2012)

    Article  Google Scholar 

  3. Anderson, P., et al.: Bottom-up and top-down attention for image captioning and visual question answering. In: CVPR 2018, p. 6 (2018)

    Google Scholar 

  4. Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate. In: ICLR (2015)

    Google Scholar 

  5. Bengio, Y., Ducharme, R., Vincent, P., Jauvin, C.: A neural probabilistic language model. J. Mach. Learn. Res. 3(Feb), 1137–1155 (2003)

    MATH  Google Scholar 

  6. Berger, A.L., Pietra, V.J.D., Pietra, S.A.D.: A maximum entropy approach to natural language processing. Comput. Linguist. 22(1), 39–71 (1996)

    Google Scholar 

  7. Brown, P.F., Desouza, P.V., Mercer, R.L., Pietra, V.J.D., Lai, J.C.: Class-based n-gram models of natural language. Comput. Linguist. 18(4), 467–479 (1992)

    Google Scholar 

  8. Chung, J., Gulcehre, C., Cho, K., Bengio, Y.: Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:1412.3555 (2014)

  9. Graves, A., Mohamed, A.r., Hinton, G.: Speech recognition with deep recurrent neural networks. In: ICASSP 2013, pp. 6645–6649. IEEE (2013)

    Google Scholar 

  10. Grover, A., Leskovec, J.: node2vec: Scalable feature learning for networks. In: KDD 2016, pp. 855–864. ACM (2016)

    Google Scholar 

  11. He, R., McAuley, J.: Ups and downs: modeling the visual evolution of fashion trends with one-class collaborative filtering. In: WWW 2016, pp. 507–517. International World Wide Web Conferences Steering Committee (2016)

    Google Scholar 

  12. Herlocker, J.L., Konstan, J.A., Terveen, L.G., Riedl, J.T.: Evaluating collaborative filtering recommender systems. ACM Trans. Inf. Syst. 22(1), 5–53 (2004)

    Article  Google Scholar 

  13. Karpathy, A., Fei-Fei, L.: Deep visual-semantic alignments for generating image descriptions. In: CVPR 2015, pp. 3128–3137 (2015)

    Google Scholar 

  14. Li, X., Chen, H.: Recommendation as link prediction in bipartite graphs: a graph kernel-based machine learning approach. Decis. Support Syst. 54(2), 880–890 (2013)

    Article  Google Scholar 

  15. Marchette, D.J., Priebe, C.E.: Predicting unobserved links in incompletely observed networks. Comput. Stat. Data Anal. 52(3), 1373–1386 (2008)

    Article  MathSciNet  Google Scholar 

  16. Mikolov, T., Karafiát, M., Burget, L., Cernockỳ, J., Khudanpur, S.: Recurrent neural network based language model. In: InterSpeech 2010, vol. 2, p. 3 (2010)

    Google Scholar 

  17. Mikolov, T., Sutskever, I., Chen, K., Corrado, G.S., Dean, J.: Distributed representations of words and phrases and their compositionality. In: NIPS 2013, pp. 3111–3119 (2013)

    Google Scholar 

  18. Papineni, K., Roukos, S., Ward, T., Zhu, W.J.: Bleu: a method for automatic evaluation of machine translation. In: ACL 2002, pp. 311–318. Association for Computational Linguistics (2002)

    Google Scholar 

  19. Pavlov, M., Ichise, R.: Finding experts by link prediction in co-authorship networks. In: FEWS 2007, pp. 42–55 (2007)

    Google Scholar 

  20. Perozzi, B., Al-Rfou, R., Skiena, S.: DeepWalk: online learning of social representations. In: KDD 2014, pp. 701–710. ACM (2014)

    Google Scholar 

  21. Potamianos, G., Jelinek, F.: A study of n-gram and decision tree letter language modeling methods. Speech Commun. 24(3), 171–192 (1998)

    Article  Google Scholar 

  22. Schwenk, H., Gauvain, J.L.: Training neural network language models on very large corpora. In: HLT/EMNLP 2005, pp. 201–208. Association for Computational Linguistics (2005)

    Google Scholar 

  23. Shore, J., Johnson, R.: Axiomatic derivation of the principle of maximum entropy and the principle of minimum cross-entropy. TIT 26(1), 26–37 (1980)

    MathSciNet  MATH  Google Scholar 

  24. Sundermeyer, M., Schlüter, R., Ney, H.: LSTM neural networks for language modeling. In: InterSpeech (2012)

    Google Scholar 

  25. Sutskever, I., Martens, J., Hinton, G.E.: Generating text with recurrent neural networks. In: ICML 2011, pp. 1017–1024 (2011)

    Google Scholar 

  26. Wang, H., Zhang, F., Hou, M., Xie, X., Guo, M., Liu, Q.: SHINE: signed heterogeneous information network embedding for sentiment link prediction. In: WSDM 2018, pp. 592–600. ACM (2018)

    Google Scholar 

  27. Xiong, Y., Zhang, Y., Kong, X., Zhu, Y.: NetCycle+: a framework for collective evolution inference in dynamic heterogeneous networks. IEEE Trans. Knowl. Data Eng. 30(8), 1547–1560 (2018)

    Article  Google Scholar 

  28. Yao, Y., Viswanath, B., Cryan, J., Zheng, H., Zhao, B.Y.: Automated crowdturfing attacks and defenses in online review systems. In: CCS 2017, pp. 1143–1158. ACM (2017)

    Google Scholar 

  29. Zamora-Martinez, F., Frinken, V., España-Boquera, S., Castro-Bleda, M.J., Fischer, A., Bunke, H.: Neural network language models for off-line handwriting recognition. Pattern Recogn. 47(4), 1642–1652 (2014)

    Article  Google Scholar 

  30. Zhang, Y., Xiong, Y., Kong, X., Li, S., Mi, J., Zhu, Y.: Deep collective classification in heterogeneous information networks. In: WWW 2018, pp. 399–408. International World Wide Web Conferences Steering Committee (2018)

    Google Scholar 

  31. Zhang, Y., Xiong, Y., Kong, X., Zhu, Y.: NetCycle: collective evolution inference in heterogeneous information networks. In: KDD 2016, pp. 1365–1374 (2016)

    Google Scholar 

Download references

Acknowledgment

This work is supported in part by the National Natural Science Foundation of China Projects No. U1636207, No. 91546105, the Shanghai Science and Technology Development Fund No. 16JC1400801, No. 17511105502, No. 17511101702.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yun Xiong .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2019 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Xu, S., Xiong, Y., Kong, X., Zhu, Y. (2019). Net2Text: An Edge Labelling Language Model for Personalized Review Generation. In: Li, G., Yang, J., Gama, J., Natwichai, J., Tong, Y. (eds) Database Systems for Advanced Applications. DASFAA 2019. Lecture Notes in Computer Science(), vol 11446. Springer, Cham. https://doi.org/10.1007/978-3-030-18576-3_29

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-18576-3_29

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-18575-6

  • Online ISBN: 978-3-030-18576-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics