Skip to yearly menu bar Skip to main content


Poster
in
Workshop: Bridging the Gap: from Machine Learning Research to Clinical Practice

Contextualized Representation Learning in Biomedical Word Sense Disambiguation

Mozhgan saeidi


Abstract:

Representation learning is an important component in solving most Natural Language Processing(NLP) problems, including Word Sense Disambiguation(WSD). The WSD task tries to find the best meaning in a knowledge base for a word with multiple meanings(ambiguous word). WSD methods choose this best meaning based on the context, i.e., the words around the ambiguous word in the input text document. Thus, word representations may improve the effectiveness of the disambiguation models if they carry useful information from the context and the knowledge base. Most of the current representation learning approaches are that they are mostly trained on the general English text and are not domain specified. In this paper, we present a novel contextual-knowledge base aware sense representation method in the biomedical domain. The novelty in our representation is the integration of the knowledge base and the context. This representation lies in a space comparable to that of contextualized word vectors, thus allowing a word occurrence to be easily linked to its meaning by applying a simple nearest neighbor approach. Comparing our approach with state-of-the-art methods shows the effectiveness of our method in terms of text coherence.

Chat is not available.