Lexical Granularity for Automatic Indexing and Means to Achieve It: The Case of Swedish MeSH®

Lexical Granularity for Automatic Indexing and Means to Achieve It: The Case of Swedish MeSH®

Dimitrios Kokkinakis
ISBN13: 9781605662749|ISBN10: 1605662747|ISBN13 Softcover: 9781616925284|EISBN13: 9781605662756
DOI: 10.4018/978-1-60566-274-9.ch002
Cite Chapter Cite Chapter

MLA

Kokkinakis, Dimitrios. "Lexical Granularity for Automatic Indexing and Means to Achieve It: The Case of Swedish MeSH®." Information Retrieval in Biomedicine: Natural Language Processing for Knowledge Integration, edited by Violaine Prince and Mathieu Roche, IGI Global, 2009, pp. 11-37. https://doi.org/10.4018/978-1-60566-274-9.ch002

APA

Kokkinakis, D. (2009). Lexical Granularity for Automatic Indexing and Means to Achieve It: The Case of Swedish MeSH®. In V. Prince & M. Roche (Eds.), Information Retrieval in Biomedicine: Natural Language Processing for Knowledge Integration (pp. 11-37). IGI Global. https://doi.org/10.4018/978-1-60566-274-9.ch002

Chicago

Kokkinakis, Dimitrios. "Lexical Granularity for Automatic Indexing and Means to Achieve It: The Case of Swedish MeSH®." In Information Retrieval in Biomedicine: Natural Language Processing for Knowledge Integration, edited by Violaine Prince and Mathieu Roche, 11-37. Hershey, PA: IGI Global, 2009. https://doi.org/10.4018/978-1-60566-274-9.ch002

Export Reference

Mendeley
Favorite

Abstract

The identification and mapping of terminology from large repositories of life science data onto concept hierarchies constitute an important initial step for a deeper semantic exploration of unstructured textual content. Accurate and efficient mapping of this kind is likely to provide better means of enhancing indexing and retrieval of text, uncovering subtle differences, similarities and useful patterns, and hopefully new knowledge, among complex surface realisations, overlooked by shallow techniques based on various forms of lexicon look-up approaches. However, a finer-grained level of mapping between terms as they occur in natural language and domain concepts is a cumbersome enterprise that requires various levels of processing in order to make explicit relevant linguistic structures. This chapter highlights some of the challenges encountered in the process of bridging free text to controlled vocabularies and thesauri and vice versa. The author investigates how the extensive variability of lexical terms in authentic data can be efficiently projected to hierarchically structured codes, while means to increase the coverage of the underlying lexical resources are also investigated.

Request Access

You do not own this content. Please login to recommend this title to your institution's librarian or purchase it from the IGI Global bookstore.