Abstract
Semantic text similarity (STS), which measures the semantic similarity of sentences, is an important task in the field of NLP. It has a wide range of applications, such as machine translation (MT), semantic search, and summarization. In recent years, with the development of deep neural networks, the existing semantic similarity measurement has made great progress. In particular, pretraining models, such as BERT-based models, which have been good representations of sentence features, have set a new state-of-the-art on STS tasks. Although a large amount of corpus data are used in the pretraining stage, there is no fine-grained semantic analysis. We observe that many sentences, such as user reviews and the QA corpus, can be abstractly regarded as including two core parts: a) this sentence states a certain attribute; and b) this attribute is described by descriptive words. This feature is particularly prominent in the corpus of reviews. Motivated by the above observations, in this paper, we propose a feature separation network (FSN) model, which can further separate and extract attribute features and description features and then measure the semantic similarity according to the separated features. To better verify the effectiveness of our model, we propose an unsupervised approach to construct the semantic similarity dataset in the review domain. Experimental results demonstrate that our method outperforms the general semantic similarity measurement method.






Similar content being viewed by others
Explore related subjects
Discover the latest articles and news from researchers in related subjects, suggested using machine learning.References
Agirre E, Banea C, Cer D, Diab M, Gonzalez Agirre A, Mihalcea R, Rigau Claramunt G, Wiebe J (2016) Semeval-2016 task 1: semantic textual similarity, monolingual and cross-lingual evaluation. In: SemEval-2016. 10th international workshop on semantic evaluation; 2016 Jun 16–17; San Diego, CA. Stroudsburg (PA): ACL; 2016. ACL (Association for Computational Linguistics), pp 497–511
Xiao Y, Liu L, Huang G, Cui Q, Huang S, Shi S, Chen J (2022) BiTIIMT: a bilingual text-infilling method for interactive machine translation. In: Proceedings of the 60th annual meeting of the association for computational linguistics (vol 1: Long Papers), pp 1958–1969
See A, Liu PJ, Manning CD (2017) Get to the point: summarization with pointer-generator networks. In: Proceedings of the 55th annual meeting of the association for computational linguistics (vol 1: Long Papers), pp 1073–1083
Lei X, Qian X, Zhao G (2016) Rating prediction based on social sentiment from textual reviews. IEEE trans multimed 18(9):1910–1921
Kuo Y, Fu M, Tsai W, Lee K, Chen L (2016) Integrated microblog sentiment analysis from users’ social interaction patterns and textual opinions. Appl Intell 44(2):399–413
Kim HK, Kim M (2016) Model-induced term-weighting schemes for text classification. Appl Intell 45(1):30–43
Zhao G, Liu Z, Chao Y, Qian X (2021) CAPER: context-aware personalized emoji recommendation. IEEE Trans Knowl Data Eng 33(9):3160–3172
Lopez-Gazpio I, Maritxalar M, Gonzalez-Agirre A, Rigau G, Uria L, Agirre E (2017) Interpretable semantic textual similarity: finding and explaining differences between sentences. Knowl-Based Syst 119:186–199
Yuan Z, Sun S, Duan L, Li C, Wu X, Xu C (2020) Adversarial multimodal network for movie story question answering. IEEE Trans Multimed 23:1744–1756
Janda HK, Pawar A, Du S, Mago V (2019) Syntactic, semantic and sentiment analysis: the joint effect on automated essay evaluation. IEEE Access 7:108486–108503
Dong J, Li X, Xu D (2018) Cross-media similarity evaluation for web image retrieval in the wild. IEEE Trans Multimed 20(9):2371–2384
Zhou Y, Zhao G, Qian X (2022) Dialogue state tracking based on hierarchical slot attention and contrastive learning. In: Hasan MA, Xiong L (eds) Proceedings of the 31st ACM international conference on information & knowledge management, pp 4737–4741
Wu Y, Liao L, Zhang G, Lei W, Zhao G, Qian X, Chua T-S (2022) State graph reasoning for multimodal conversational recommendation. IEEE Trans Multimed, 1–1. https://doi.org/10.1109/TMM.2022.3155900
Kenton JDM-WC, Toutanova LK (2019) Bert: pre-training of deep bidirectional transformers for language understanding, 4171– 4186
Chen K, Zhang Z, Long J, Zhang H (2016) Turning from TF-IDF to TF-IGM for term weighting in text classification. Expert Syst Appl 66:245–260
Wang T, Guo J, Wu Z, Xu T (2021) IFTA: iterative filtering by using TF-AICL algorithm for Chinese encyclopedia knowledge refinement. Appl Intell 51(8):6265–6293
Yang Y, Yuan S, Cer D, Kong S-y, Constant N, Pilar P, Ge H, Sung Y-H, Strope B, Kurzweil R (2018) Learning semantic textual similarity from conversations, 164–174
Lan W, Xu W (2018) Neural network models for paraphrase identification, semantic textual similarity, natural language inference, and question answering. In: Proceedings of the 27th international conference on computational linguistics, pp 3890–3902
Wang T, Li J, Guo J (2021) A scalable parallel chinese online encyclopedia knowledge denoising method based on entry tags and spark cluster. Appl Intell 51(10):7573–7599
Zhu J, He Y, Zhao G, Bo X, Qian X (2022) Joint reason generation and rating prediction for explainable recommendation. IEEE Trans Knowl Data Eng
Liu Y, Ott M, Goyal N, Du J, Joshi M, Chen D, Levy O, Lewis M, Zettlemoyer L, Stoyanov V (2019) Roberta: a robustly optimized bert pretraining approach. arXiv:1907.11692
Yang Z, Dai Z, Yang Y, Carbonell J, Salakhutdinov RR, Le QV (2019) Xlnet: generalized autoregressive pretraining for language understanding. Adv Neural Inf Process Syst, 32
Sun Y, Wang S, Li Y, Feng S, Tian H, Wu H, Wang H. (2020) Ernie 2.0: a continual pre-training framework for language understanding. In: Proceedings of the AAAI conference on artificial intelligence, vol 34, pp 8968–8975
Lan Z, Chen M, Goodman S, Gimpel K, Sharma P, Soricut R (2019) Albert: a lite bert for self-supervised learning of language representations
May C, Wang A, Bordia S, Bowman SR, Rudinger R (2019) On measuring social biases in sentence encoders, 622–628
Zhang T, Kishore V, Wu F, Weinberger KQ, Artzi Y (2019) Bertscore: evaluating text generation with Bert, 1–43
Qiao Y, Xiong C, Liu Z, Liu Z (2019) Understanding the behaviors of bert in ranking. arXiv:1904.07531
Zheng X, Zhao G, Zhu L, Qian X (2022) PERD: personalized emoji recommendation with dynamic user preference. In: Proceedings of the 45th international ACM SIGIR conference on research and development in information retrieval, pp 1922–1926
Reimers N, Gurevych I (2019) Sentence-bert: sentence embeddings using siamese bert-networks, 3982–3992
Das A, Yenala H, Chinnakotla MK, Shrivastava M (2016) Together we stand: Siamese networks for similar question retrieval. In: Proceedings of the 54th annual meeting of the association for computational linguistics, ACL 2016, August 7–12, 2016, Berlin, Germany, vol 1: long papers. The Association for Computer Linguistics
Ranasinghe T, Orǎsan C, Mitkov R (2019) Semantic textual similarity with siamese neural networks. In: Proceedings of the international conference on recent advances in natural language processing (RANLP 2019), pp 1004–1011
Pontes EL, Huet S, Linhares AC, Torres-Moreno J-M (2018) Predicting the semantic textual similarity with siamese CNN and LSTM. arXiv:1810.10641
Cranias L, Papageorgiou H, Piperidis S (1994) A matching technique in example-based machine translation. In: COLING 1994 vol 1: the 15th international conference on computational linguistics
Salton G, Wong A, Yang C-S (1975) A vector space model for automatic indexing. Commun ACM 18(11):613–620
Qiang J, Li Y, Yuan Y, Wu X (2018) Short text clustering based on Pitman-Yor process mixture model. Appl Intell 48(7):1802–1812
Zhao G, Lou P, Qian X, Hou X (2020) Personalized location recommendation by fusing sentimental and spatial context. Knowl-Based Syst 196:105849
Zhang D, Zhu Z, Kang S, Zhang G, Liu P (2021) Syntactic and semantic analysis network for aspect-level sentiment classification. Appl Intell 51(8):6136–6147
Majumder G, Pakray P, Das R, Pinto D (2021) Interpretable semantic textual similarity of sentences using alignment of chunks with classification and regression. Appl Intell 51(10):7322– 7349
Wang T, Liu L, Liu N, Zhang H, Zhang L, Feng S (2020) A multi-label text classification method via dynamic semantic representation model and deep neural network. Appl Intell 50(8):2339–2351
Zheng X, Zhao G, Zhu L, Zhu J, Qian X (2022) What you like, what I am: online dating recommendation via matching individual preferences with features. IEEE Trans Knowl Data Eng
Vo A, Nguyen Q, Ock C (2020) Semantic and syntactic analysis in learning representation based on a sentiment analysis model. Appl Intell 50(3):663–680
Kim Y (2014) Convolutional neural networks for sentence classification. In: Moschitti A, Pang B, Daelemans W (eds) Proceedings of the 2014 conference on empirical methods in natural language processing, EMNLP 2014, October 25–29, 2014, Doha, Qatar, a meeting of SIGDAT, a special interest group of the ACL. ACL, pp 1746–1751
Hu B, Lu Z, Li H, Chen Q (2014) Convolutional neural network architectures for matching natural language sentences. In: Advances in neural information processing systems, pp 2042–2050
Mueller J, Thyagarajan A (2016) Siamese recurrent architectures for learning sentence similarity. In: Proceedings of the AAAI conference on artificial intelligence, vol 30
Zhao G, Lei X, Qian X, Mei T (2018) Exploring users’ internal influence from reviews for social recommendation. IEEE Trans Multimed 21(3):771–781
Bahdanau D, Cho KH, Bengio Y (2015) Neural machine translation by jointly learning to align and translate
Wang S, Jiang J (2016) Learning natural language inference with LSTM. In: Proceedings of the 2016 conference of the north american chapter of the association for computational linguistics: human language technologies, pp 1442–1451
Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser L, Polosukhin I (2017) Attention is all you need. In: NIPS
Peters M, Neumann M, Iyyer M, Gardner M, Clark C, Lee K, Zettlemoyer L (2018) Deep contextualized word representations. In: Proceedings of the 2018 conference of the North American chapter of the association for computational linguistics: human language technologies, vol 1 (Long Papers), pp 2227–2237
Wang B, Kuo C-CJ (2020) Sbert-wk: a sentence embedding method by dissecting bert-based word models. IEEE/ACM Trans Audio Speech Lang Process 28:2146–2157. https://doi.org/10.1109/TASLP.2020.3008390
Arora S, Liang Y, Ma T (2017) A simple but tough-to-beat baseline for sentence embeddings. In: 5th international conference on learning representations, ICLR 2017
Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780
Liu X, Chen Q, Deng C, Zeng H, Chen J, Li D, Tang B (2018) Lcqmc: A large-scale chinese question matching corpus. In: Proceedings of the 27th international conference on computational linguistics, pp 1952–1962
Kingma DP, Ba J (2014) Adam: a method for stochastic optimization. arXiv:1412.6980
Mitra B, Diaz F, Craswell N (2017) Learning to match using local and distributed representations of text for web search. In: Proceedings of the 26th international conference on World Wide Web, pp 1291–1299
Agirre E, Banea C, Cardie C, Cer D, Diab M, Gonzalez-Agirre A, Guo W, Mihalcea R, Rigau G, Wiebe J (2014) Sem Eval-2014 task 10: Multilingual semantic textual similarity. In: Proceedings of the 8th international workshop on semantic evaluation (SemEval 2014). https://doi.org/10.3115/v1/S14-2010, https://www.aclweb.org/anthology/S14-2010. Association for Computational Linguistics, pp 81–91
Acknowledgements
This work is funded in part by the NSFC, China under Grant 61902309; in part by the Fundamental Research Funds for the Central Universities, China (xxj022019003, xzd012022006); in part by the China Postdoctoral Science Foundation (2020M683496); and in part by the National Postdoctoral Innovative Talents Support Program, China (BX20190273); in part by the Humanities and Social Sciences Foundation of Ministry of Education, China under Grant 16XJAZH003; and in part by the Science and Technology Program of Xi’an, China under Grant 21RGZN0017.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of Interests
The authors declare that they have no conflicts of interest.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Chen, Q., Zhao, G., Wu, Y. et al. Fine-grained semantic textual similarity measurement via a feature separation network. Appl Intell 53, 18205–18218 (2023). https://doi.org/10.1007/s10489-022-04448-6
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-022-04448-6