skip to main content
10.1145/3313991.3314019acmotherconferencesArticle/Chapter ViewAbstractPublication PagesiccaeConference Proceedingsconference-collections
research-article

Enhancing Semantic Word Representations by Embedding Deep Word Relationships

Authors Info & Claims
Published:23 February 2019Publication History

ABSTRACT

Word representations are created using analogy context-based statistics and lexical relations on words. Word representations are inputs for the learning models in Natural Language Understanding (NLU) tasks. However, to understand language, knowing only the context is not sufficient. Reading between the lines is a key component of NLU. Embedding deeper word relationships which are not represented in the context enhances the word representation. This paper presents a word embedding which combines an analogy, context-based statistics using Word2Vec, and deeper word relationships using Conceptnet, to create an expanded word representation. In order to fine-tune the word representation, Self-Organizing Map is used to optimize it. The proposed word representation is compared with semantic word representations using Simlex 999. Furthermore, the use of 3D visual representations has shown to be capable of representing the similarity and association between words. The proposed word representation shows a Spearman correlation score of 0.886 and provided the best results when compared to the current state-of-the-art methods, and exceed the human performance of 0.78.

References

  1. T. Mikolov, I. Sutskever, K. Chen, G. S. Corrado, and J. Dean, "Distributed representations of words and phrases and their compositionality," in Advances in neural information processing systems, 2013, pp. 3111--3119. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. G. Recski, E. Iklódi, K. Pajkossy, and A. Kornai, "Measuring semantic similarity of words using concept networks," in Proceedings of the 1st Workshop on Representation Learning for NLP, 2016, pp. 193--200.Google ScholarGoogle Scholar
  3. B. Dhingra, H. Liu, R. Salakhutdinov, and W. W. Cohen, "A comparative study of word embeddings for reading comprehension," arXiv preprint arXiv:1703.00993, 2017.Google ScholarGoogle Scholar
  4. R. Collobert and J. Weston, "A unified architecture for natural language processing: Deep neural networks with multitask learning," in Proceedings of the 25th international conference on Machine learning, 2008, pp. 160--167: ACM. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. A. Kumar et al., "Ask me anything: Dynamic memory networks for natural language processing," in International Conference on Machine Learning, 2016, pp. 1378--1387. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. A. Nugaliyadde, K. W. Wong, F. Sohel, and H. Xie, "Reinforced memory network for question answering," in International Conference on Neural Information Processing, 2017, pp. 482--490: Springer.Google ScholarGoogle Scholar
  7. D. Bahdanau, K. Cho, and Y. Bengio, "Neural machine translation by jointly learning to align and translate," arXiv preprint arXiv:1409.0473, 2014.Google ScholarGoogle Scholar
  8. T.-H. Wen, M. Gasic, N. Mrksic, P.-H. Su, D. Vandyke, and S. Young, "Semantically conditioned lstm-based natural language generation for spoken dialogue systems," arXiv preprint arXiv:1508.01745, 2015.Google ScholarGoogle Scholar
  9. X. Zhang and Y. LeCun, "Text understanding from scratch," arXiv preprint arXiv:1502.01710, 2015.Google ScholarGoogle Scholar
  10. J. P. Chiu and E. Nichols, "Named entity recognition with bidirectional LSTM-CNNs," arXiv preprint arXiv:1511.08308, 2015.Google ScholarGoogle Scholar
  11. T. Mikolov, K. Chen, G. Corrado, and J. Dean, "Efficient estimation of word representations in vector space," arXiv preprint arXiv:1301.3781, 2013.Google ScholarGoogle Scholar
  12. J. Pennington, R. Socher, and C. Manning, "Glove: Global vectors for word representation," in Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), 2014, pp. 1532--1543.Google ScholarGoogle Scholar
  13. C. Potts, D. Lassiter, R. Levy, and M. C. Frank, "Embedded implicatures as pragmatic inferences under compositional lexical uncertainty," Journal of Semantics, vol. 33, no. 4, pp. 755--802, 2016.Google ScholarGoogle Scholar
  14. A. Nugaliyadde, K. W. Wong, F. Sohel, and H. Xie, "Multi-level Search of a Knowledgebase for Semantic Parsing," in International Workshop on Multi-disciplinary Trends in Artificial Intelligence, 2017, pp. 44--53: Springer.Google ScholarGoogle Scholar
  15. H. Liu and P. Singh, "ConceptNet---a practical commonsense reasoning tool-kit," BT technology journal, vol. 22, no. 4, pp. 211--226, 2004. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. F. Hill, R. Reichart, and A. Korhonen, "Simlex-999: Evaluating semantic models with (genuine) similarity estimation," Computational Linguistics, vol. 41, no. 4, pp. 665--695, 2015. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. J. A. Bullinaria and J. P. Levy, "Extracting semantic representations from word co-occurrence statistics: A computational study," Behavior research methods, vol. 39, no. 3, pp. 510--526, 2007.Google ScholarGoogle Scholar
  18. R. Lebret and R. Collobert, "Word emdeddings through hellinger PCA," arXiv preprint arXiv:1312.5542, 2013.Google ScholarGoogle Scholar
  19. T. Mikolov, W.-t. Yih, and G. Zweig, "Linguistic regularities in continuous space word representations," in Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2013, pp. 746--751.Google ScholarGoogle Scholar
  20. Y. Goldberg and O. Levy, "word2vec Explained: deriving Mikolov et al.'s negative-sampling word-embedding method," arXiv preprint arXiv:1402.3722, 2014.Google ScholarGoogle Scholar
  21. R. Speer, J. Chin, and C. Havasi, "ConceptNet 5.5: An Open Multilingual Graph of General Knowledge," in AAAI, 2017, pp. 4444--4451. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. R. Speer and J. Chin, "An ensemble method to produce high-quality word embeddings," arXiv preprint arXiv:1604.01692, 2016.Google ScholarGoogle Scholar
  23. G. N. Leech, Principles of pragmatics. Routledge, 2016.Google ScholarGoogle ScholarCross RefCross Ref
  24. T. Kohonen, "The self-organizing map," Neurocomputing, vol. 21, no. 1-3, pp. 1--6, 1998.Google ScholarGoogle Scholar
  25. L. Finkelstein et al., "Placing search in context: The concept revisited," in Proceedings of the 10th international conference on World Wide Web, 2001, pp. 406--414: ACM. Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. J. Weston et al., "Towards ai-complete question answering: A set of prerequisite toy tasks," arXiv preprint arXiv:1502.05698, 2015.Google ScholarGoogle Scholar

Index Terms

  1. Enhancing Semantic Word Representations by Embedding Deep Word Relationships

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Other conferences
      ICCAE 2019: Proceedings of the 2019 11th International Conference on Computer and Automation Engineering
      February 2019
      160 pages
      ISBN:9781450362870
      DOI:10.1145/3313991

      Copyright © 2019 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 23 February 2019

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article
      • Research
      • Refereed limited

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader