/Wordembedding-and-semantics

Evaluation and improvement of wordembeding and semantics

Primary LanguagePythonApache License 2.0Apache-2.0

Wordembedding-and-semantics

Abstract

Neural network based word embedding has demonstrated outstanding results in variety task, and become a standard input for NLP related deep learning research. These representations are capable to catch semantic regularities in language, e.g. analogy relation. While a general question "what kind of semantic relation does the embedding represent and how the semantic relation could be retrieved using the embedding model?" is not clear and rare relevant work was explored. In this study, we proposed a new approach to explore the semantic relation represented in neural-embedding based on WordNet and UMLS. Our study demonstrated neural embedding did prefer some semantic relation as well as the neural embedding also represented diverse semantic relations. Our study also found out the NER based phrase composition outperformed Word2phrase and the word variants did not affect the performance on analogy and semantic relation tasks.

Method

Presentation Link

Result

Top 10 nearest neighbors

top 10 nearest neighbos

Evaluation term and its relation term

relation word

Analogy tasks @ Word2vec

analogy

Analogy task @ top 5

analogy top 5

Semantic relation task @ Word2vec

relation @ Word2vec

Semantic relation @ top 5

relation top 5

Semantic weighted relation @ top5

relation top 5

Analogy phrase composition

phrase analogy top 5

Semantic relation phrase composition

phrase relation top 5