This repository contains the implementation of our approach proposed in the paper
Wiedemann, G., Remus, S., Chawla, A., Biemann, C. (2019): Does BERT Make Any Sense? Interpretable Word Sense Disambiguation with Contextualized Embeddings. Proceedings of KONVENS 2019, Erlangen, Germany.
Word Sense Disambiguation (WSD) is the task to identify the correct sense of the usage of a word from a (usually) fixed inventory of sense identifiers. We propose a simple approach that scans through the training data to learn the Contextualized Word Embeddings(CWE) of sense labels and classifies the ambiguous words on the basis Cosine Similarity with the learnt CWEs for that word.
- Python3.5 or above
- PyTorch 0.4.0
- Huggingface's Pytorch-Pretrained-BERT
- Allennlp ELMo
- Flair
- Scikit-learn
- UFSAC
- Java (version 8 or higher)
- Maven
- Get UFSAC datasets
- Run commands ...