palashmoon/word_embedding_using_keras
A word embedding is a learned representation for text where words that have the same meaning have a similar representation.Word embeddings are in fact a class of techniques where individual words are represented as real-valued vectors in a predefined vector space. Each word is mapped to one vector and the vector values are learned in a way that resembles a neural network
Jupyter Notebook
No issues in this repository yet.