Med­i­cal Con­cept Em­bed­dings via La­beled Back­ground Cor­po­ra

This entry contains the resources used in and resulting from

Eneldo Loza Mencía, Gerard de Melo and Jinseok Nam, Medical Concept Embeddings via Labeled Background Corpora, in: Proceedings of the 10th International Conference on Language Resources and Evaluation (LREC 2016), 2016

In recent years, we have seen an increasing amount of interest in low-dimensional vector representations of words. Among other things, these facilitate computing word similarity and relatedness scores. The most well-known example of algorithms to produce representations of this sort are the word2vec approaches. In this paper, we investigate a new model to induce such vector spaces for medical concepts, based on a joint objective that exploits not only word co-occurrences but also manually labeled documents, as available from sources such as PubMed. Our extensive experimental analysis shows that our embeddings lead to significantly higher correlations with human similarity and relatedness assessments than previous work. Due to the simplicity and versatility of vector representations, these findings suggest that our resource can easily be used as a drop-in replacement to improve any systems relying on medical concept similarity measures.

Related Identifier
Metadata Access
Creator Loza Mencia, Eneldo; de Melo, Gerard; Nam, Jinseok
Publisher TU Darmstadt
Contributor TU Darmstadt
Publication Year 2016
Rights info:eu-repo/semantics/openAccess
OpenAccess true
Language English
Resource Type Dataset
Format application/octet-stream; text/plain; application/zip; application/gzip
Version Version 1.0
Discipline Linguistics