Skip to main content

Y-X-Y Encoding for Identifying Types of Sentence Similarity

  • Conference paper
  • First Online:
Data Science and Emerging Technologies (DaSET 2022)

Abstract

Determining the semantic similarity of any two arbitrary sentences requires two steps, i.e. sentence encoding and semantic similarity measure. The most important step is to encode a set of sentences into a set of equal-length vectors for similarity measure in forms of classification. Two practical encoding schemes had been proposed, statistical-based direct encoding and pretrained encoding. The first approach lacks considering word correlation and the dimension of encoded vector is very large. For the second approach, it requires an extra training time prior to the classification process. This study compromises the previous approaches by considering shallow neural networks for encoding sentences and classifying entailment relations between two sentences. A set of y-x-y encoder models is proposed where y can be greater or less than x depending on given dataset. Neither encoder models nor their corresponding classifiers are built upon big and complex structure, and hence is suitable for carrying out such task. The encoding scheme is tested with SICK 2014 dataset [1], specially designed for neutral, entailment, and contradiction sentence pairs. Comparison results (neutral 97.1%, entailment 91.1%, contradiction 94.6%) support the possibility of the proposed scheme to sentence similarity measure.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 189.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 249.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Free shipping worldwide - see info
Hardcover Book
USD 249.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Marco, M., et al.: SemEval-2014 task 1: evaluation of compositional distributional semantic models on full sentences through semantic relatedness and textual entailment. In: Proceedings of the SemEval (2014)

    Google Scholar 

  2. Pennington, J., Socher, R., Manning, C.D.: GloVe: global vectors for word representation. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP) (2014)

    Google Scholar 

  3. Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)

    Article  Google Scholar 

  4. Devlin, J., et al.: BERT: pre-training of deep bidirectional transformers for language understanding. Association for Computational Linguistics, Minneapolis, Minnesota (2019)

    Google Scholar 

  5. Vaswani, A., et al.: Attention is all you need. Adv. Neural Inf. Process. Syst. 30 (2017)

    Google Scholar 

  6. Jiao, X., et al.: TinyBERT: distilling BERT for natural language understanding. arXiv preprint arXiv:1909.10351 (2019)

  7. Sanh, V., et al.: DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108 (2019)

  8. Li, B., et al.: On the sentence embeddings from pre-trained language models. arXiv preprint arXiv:2011.05864 (2020)

  9. Mikolov, T., et al.: Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781 (2013)

  10. Peters, M.E., et al.: Deep contextualized word representations. Association for Computational Linguistics, New Orleans, Louisiana (2018)

    Google Scholar 

  11. Mikolov, T., et al.: Advances in pre-training distributed word representations. arXiv preprint arXiv:1712.09405 (2017)

  12. Bowman, S.R., et al.: A large annotated corpus for learning natural language inference. Association for Computational Linguistics, Portugal (2015)

    Google Scholar 

  13. Harris, Z.S.: Distributional structure. Word 10(2–3), 146–162 (1954)

    Article  Google Scholar 

  14. Chen, Z., Gao, Q., Moss, L.S.: NeuralLog: natural language inference with joint neural and logical reasoning. Association for Computational Linguistics (2021)

    Google Scholar 

  15. Lai, A., Hockenmaier, J.: Illinois-LH: a denotational and distributional approach to semantics. Association for Computational Linguistics, Dublin, Ireland (2014)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Chidchanok Lursinsap .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Jinnovart, T., Lursinsap, C. (2023). Y-X-Y Encoding for Identifying Types of Sentence Similarity. In: Wah, Y.B., Berry, M.W., Mohamed, A., Al-Jumeily, D. (eds) Data Science and Emerging Technologies. DaSET 2022. Lecture Notes on Data Engineering and Communications Technologies, vol 165. Springer, Singapore. https://doi.org/10.1007/978-981-99-0741-0_37

Download citation

Publish with us

Policies and ethics