Information_retrieval

  • word2vec.py
    word2vec implementation, You can use Hierarchical softmax, Negative Sampling, Subsampling (Skip_gram, CBOW)

  • fast.py
    fast_text implementation, You can use Negative Sampling, Subsampling (Skip_gram, CBOW)

  • fastclassification.py
    fast text based classification the result of AG datasets is 0.90 almost same as paper presented 92.5(h = 10, bigram). The difference of accuracy is from batch or lr setting (In my opinion).

  • textCNN.ipynb
    textCNN-rand(2014, kim) implementation in tensorflow2.0. The accuracy is 74.7 which is similar to 76.1 in paper. The accuracy is increasing when I use another initializers. In this implementation, I use default one.