Skip to main content

Aspect-Level Sentiment Difference Feature Interaction Matching Model Based on Multi-round Decision Mechanism

  • Conference paper
  • First Online:
Algorithms and Architectures for Parallel Processing (ICA3PP 2020)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 12453))

  • 2012 Accesses

Abstract

Sentence matching is a key problem in natural language understanding, so the research on sentence matching can be applied to a large number of known natural language processing tasks, such as information retrieval, automatic question and answer, machine translation, dialogue system, paraphrase identification etc. In a series of natural language processing tasks, we need to rely on the participation and collaboration of the sentence matching model. The performance of the sentence matching model can greatly affect the final performance of these natural language processing tasks. We propose the Al-SFIM model, which improves the matching model from the perspective of word interaction. First, we propose sentiment attention mechanism based on the distribution of aspect-level sentiment difference to improve the interaction between cross-sentence words, and use the sentiment space position perception vector to improve the interaction between intra-sentence words, so that the model has the ability to perceive the subjective sentiment difference in the process of intra-sentence word interaction and cross-sentence word interaction. Then, we introduce a multi-round decision mechanism based on the accumulation of memory state, which iteratively updates the working memory state to make matching decisions in multiple rounds, so that the model can better understand the semantic of complex sentence. Experiment results show that the AL-SFIM model has made progress in sentence matching and has better matching performance for complex, long and incomprehensible sentences.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

Notes

  1. 1.

    https://nlp.stanford.edu/projects/glove/.

References

  1. Xue, X., Jeon. J., Croft, W.B.: Retrieval models for question and answer archives. In: Proceedings of the 31st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 475–482 (2008)

    Google Scholar 

  2. Surdeanu, M., Ciaramita, M., Zaragoza, H.: Learning to rank answers to non-factoid questions from web collections. Comput. Linguist. 37(2), 351–383 (2011)

    Article  Google Scholar 

  3. Yih, W.-T., et al.: Question answering using enhanced lexical semantic models. In: Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics, pp. 1744–1753 (2013)

    Google Scholar 

  4. Tymoshenko, K, Moschitti, A.: Assessing the impact of syntactic and semantic structures for answer passages reranking. In: Proceedings of the 24th ACM International on Conference on Information and Knowledge Management, pp. 1451–1460 (2015)

    Google Scholar 

  5. Huang, P.-S., et al.: Learning deep structured semantic models for web search using clickthrough data. In: Proceedings of the 22nd ACM International Conference on Information & Knowledge Management, pp. 2333–2338 (2013)

    Google Scholar 

  6. Shen, Y., et al.: A latent semantic model with convolutional-pooling structure for information retrieval. In: Proceedings of the 23rd ACM International Conference on Conference on Information and Knowledge Management, pp. 101–110 (2014)

    Google Scholar 

  7. Hu. B., et al.: Convolutional neural network architectures for matching natural language sentences. In: Advances in Neural Information Processing Systems, pp. 2042–2050 (2014)

    Google Scholar 

  8. Qiu, X., Huang, X.: Convolutional neural tensor network architecture for community-based question answering. In: Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence, pp. 1305–1311 (2015)

    Google Scholar 

  9. Palangi, H., et al.: Deep sentence embedding using long short-term memory networks: analysis and application to information retrieval. IEEE/ACM Trans. Audio Speech Lang. Process. 24(4), 694–707 (2016)

    Article  Google Scholar 

  10. Wan, S., et al.; A deep architecture for semantic matching with multiple positional sentence representations. In: Thirtieth AAAI Conference on Artificial Intelligence, pp. 2835–2841 (2016)

    Google Scholar 

  11. Yin, W., Schütze, H.: MultiGranCNN: an architecture for general matching of text chunks on multiple levels of Granularity. In: Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing, pp. 63–73 (2015)

    Google Scholar 

  12. Socher, R., et al.: Dynamic pooling and unfolding recursive autoencoders for paraphrase detection. In: Advances in Neural Information Processing Systems, pp. 801–809 (2011)

    Google Scholar 

  13. Pang, L., et al.: Text matching as image recognition. In: National Conference on Artificial Intelligence, pp. 2793–2799 (2016)

    Google Scholar 

  14. Lu, Z., Li, H.: A deep architecture for matching short texts. In: Advances in Neural Information Processing Systems, pp. 1367–1375 (2013)

    Google Scholar 

  15. Wan, S., et al.: Match-SRNN: modeling the recursive matching structure with spatial RNN. In: Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, pp. 2922–2928 (2016)

    Google Scholar 

  16. Wang, Z., Hamza, W., Florian, R.: Bilateral multi-perspective matching for natural language sentences. In: Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, pp. 4144–4150 (2017)

    Google Scholar 

  17. Gong, Y., Luo, H., Zhang, J.: Natural language inference over interaction space. In: 6th International Conference on Learning Representations (2018)

    Google Scholar 

  18. Kim, S., Kang, I., Kwak, N.: Semantic sentence matching with densely-connected recurrent and co-attentive information. In: The Thirty-Third AAAI Conference on Artificial Intelligence, pp. 6586–6593 (2019)

    Google Scholar 

  19. Luong, T., Pham, H., Manning, C.D.: Effective approaches to attention-based neural machine translation. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pp. 1412–1421 (2015)

    Google Scholar 

  20. Guo, M., Zhang, Y., Liu, T.: Gaussian transformer: a lightweight approach for natural language inference. In: The Thirty-Third AAAI Conference on Artificial Intelligence (2019)

    Google Scholar 

  21. Im, J., Cho, S.: Distance-based self-attention network for natural language inference. arXiv Comput. Lang. (2017)

    Google Scholar 

  22. Hill, F., et al.: The Goldilocks principle: reading children’s books with explicit memory representations. In: International Conference on Learning Representations (2016)

    Google Scholar 

  23. Shen, Y., et al.: An empirical analysis of multiple-turn reasoning strategies in reading comprehension tasks. In: International Joint Conference on Natural Language Processing, pp. 957–966 (2017)

    Google Scholar 

  24. Xu, Y., et al.: Multi-task Learning with sample re-weighting for machine reading comprehension. In: North American Chapter of the Association For Computational Linguistics, pp. 2644–2655 (2019)

    Google Scholar 

  25. Liu, X., Duh, K., Gao, J.: Stochastic answer networks for natural language inference. arXiv Comput. Lang. (2018)

    Google Scholar 

  26. Pennington, J., Socher, R., Manning, C.: Glove: global vectors for word representation. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 1532–1543 (2014)

    Google Scholar 

  27. Salimans, T., Kingma, D.P.: Weight normalization: a simple reparameterization to accelerate training of deep neural networks. In: Neural Information Processing Systems, pp. 901–909 (2016)

    Google Scholar 

  28. Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. In: International Conference on Learning Representations (2015)

    Google Scholar 

Download references

Acknowledgements

This research is supported by the Scientific Research Platforms and Projects in Universities in Guangdong Province under Grants 2019KTSCX204.

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to Yanzhi Wei or Xianghua Fu .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Wei, Y., Fu, X., Wang, S., Xie, W., He, J., Zhao, Y. (2020). Aspect-Level Sentiment Difference Feature Interaction Matching Model Based on Multi-round Decision Mechanism. In: Qiu, M. (eds) Algorithms and Architectures for Parallel Processing. ICA3PP 2020. Lecture Notes in Computer Science(), vol 12453. Springer, Cham. https://doi.org/10.1007/978-3-030-60239-0_32

Download citation

Publish with us

Policies and ethics