skip to main content
10.1145/3585967.3585994acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicwcsnConference Proceedingsconference-collections
research-article

Short-Text Semantic Similarity Model of BERT-Based Siamese Network

Authors Info & Claims
Published:19 April 2023Publication History

ABSTRACT

People convey their emotions and thoughts through words, the medium of human thoughts. Up against the vigorous development of streaming media, the calculation of text similarity is imperative in the field of natural language processing. Any text-related field is inseparable from text semantic similarity. The calculation of text semantic similarity plays a key role in document management, document classification, and document relevance. Besides, popular natural language processing tasks in some trendy fields, such as artificial intelligence, human-machine translation, problem system, intelligent chat system, and nomenclature recognition, are intertwined with text semantic similarity calculation. In recent years, many excellent researchers have studied the algorithms and models of text semantic similarity from different dimensions. In this paper, a new short-text cosine similarity calculation model of the BERT-based Siamese network is proposed.

References

  1. Chengcheng H A N, Lei L I, Tingting L I U, Approaches for semantic textual similarity[J]. Journal of East China Normal University (Natural Science), 2020, 2020(5): 95Google ScholarGoogle Scholar
  2. Mikolov T, Chen K, Corrado G, Efficient estimation of word representations in vector space[J]. arXiv preprint arXiv:1301.3781, 2013.Google ScholarGoogle Scholar
  3. Pennington J, Socher R, Manning C D. Glove: Global vectors for word representation[C]//Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP). 2014: 1532-1543.Google ScholarGoogle Scholar
  4. Joulin A, Grave E, Bojanowski P, Bag of tricks for efficient text classification[J]. arXiv preprint arXiv:1607.01759, 2016.Google ScholarGoogle Scholar
  5. Peters, M., Neumann, M., Iyyer, M. .Deep contextualized word representations[J]. arXiv preprint arXiv:1802.05365, 2018.Google ScholarGoogle Scholar
  6. Devlin J, Chang M W, Lee K, Bert: Pre-training of deep bidirectional transformers for language understanding[J]. arXiv preprint arXiv:1810.04805, 2018.Google ScholarGoogle Scholar
  7. Radford A, Narasimhan K, Salimans T, Improving language understanding by generative pre-training[J]. 2018.Google ScholarGoogle Scholar
  8. Huang P S, He X, Gao J, Learning deep structured semantic models for web search using clickthrough data[C]//Proceedings of the 22nd ACM international conference on Information & Knowledge Management. 2013: 2333-2338.Google ScholarGoogle Scholar
  9. Yin W, Schütze H, Xiang B, Abcnn: Attention-based convolutional neural network for modeling sentence pairs[J]. Transactions of the Association for Computational Linguistics, 2016, 4: 259-272.Google ScholarGoogle ScholarCross RefCross Ref
  10. Vaswani A, Shazeer N, Parmar N, Attention is all you need[J]. Advances in neural information processing systems, 2017, 30.Google ScholarGoogle Scholar
  11. Reimers N, Gurevych I. Sentence-bert: Sentence embeddings using siamese bert-networks[J]. arXiv preprint arXiv:1908.10084, 2019.Google ScholarGoogle Scholar

Index Terms

  1. Short-Text Semantic Similarity Model of BERT-Based Siamese Network
      Index terms have been assigned to the content through auto-classification.

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Other conferences
        icWCSN '23: Proceedings of the 2023 10th International Conference on Wireless Communication and Sensor Networks
        January 2023
        162 pages
        ISBN:9781450398466
        DOI:10.1145/3585967

        Copyright © 2023 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 19 April 2023

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article
        • Research
        • Refereed limited
      • Article Metrics

        • Downloads (Last 12 months)63
        • Downloads (Last 6 weeks)4

        Other Metrics

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      HTML Format

      View this article in HTML Format .

      View HTML Format