Pinned Repositories
acl2017
Code to train and use models from "Revisiting Recurrent Networks for Paraphrastic Sentence Embeddings"
beyond-bleu
Python code for training models in the ACL paper, "Beyond BLEU:Training Neural Machine Translation with Semantic Similarity".
bilingual-generative-transformer
Code for "A Bilingual Generative Transformer for Semantic Sentence Embedding" published at EMNLP 2020.
charagram
Code to train and use models from "Charagram: Embedding Words and Sentences via Character n-grams".
emnlp2017
Code and data to train and use models from "Learning Paraphrastic Sentence Embeddings from Back-Translated Bitext"
iclr2016
Python code for training all models in the ICLR paper, "Towards Universal Paraphrastic Sentence Embeddings". These models achieve strong performance on semantic similarity tasks without any training or tuning on the training data for those tasks. They also can produce features that are at least as discriminative as skip-thought vectors for semantic similarity tasks at a minimum. Moreover, this code can achieve state-of-the-art results on entailment and sentiment tasks.
para-nmt-50m
Pre-trained models and code and data to train and use models from "Pushing the Limits of Paraphrastic Sentence Embeddings with Millions of Machine Translations"
paragram-word
Python code for training Paragram word embeddings. These achieve human-level performance on some word similiarty tasks including SimLex-999.This code was used to obtain results in the appendix of our 2015 TACL paper "From Paraphrase Database to Compositional Paraphrase Model and Back".
paraphrastic-representations-at-scale
simple-and-effective-paraphrastic-similarity
Python code for training models in the ACL paper, "Simple and Effective Paraphrastic Similarity from Parallel Translations".
jwieting's Repositories
jwieting/iclr2016
Python code for training all models in the ICLR paper, "Towards Universal Paraphrastic Sentence Embeddings". These models achieve strong performance on semantic similarity tasks without any training or tuning on the training data for those tasks. They also can produce features that are at least as discriminative as skip-thought vectors for semantic similarity tasks at a minimum. Moreover, this code can achieve state-of-the-art results on entailment and sentiment tasks.
jwieting/charagram
Code to train and use models from "Charagram: Embedding Words and Sentences via Character n-grams".
jwieting/para-nmt-50m
Pre-trained models and code and data to train and use models from "Pushing the Limits of Paraphrastic Sentence Embeddings with Millions of Machine Translations"
jwieting/paraphrastic-representations-at-scale
jwieting/beyond-bleu
Python code for training models in the ACL paper, "Beyond BLEU:Training Neural Machine Translation with Semantic Similarity".
jwieting/acl2017
Code to train and use models from "Revisiting Recurrent Networks for Paraphrastic Sentence Embeddings"
jwieting/paragram-word
Python code for training Paragram word embeddings. These achieve human-level performance on some word similiarty tasks including SimLex-999.This code was used to obtain results in the appendix of our 2015 TACL paper "From Paraphrase Database to Compositional Paraphrase Model and Back".
jwieting/simple-and-effective-paraphrastic-similarity
Python code for training models in the ACL paper, "Simple and Effective Paraphrastic Similarity from Parallel Translations".
jwieting/bilingual-generative-transformer
Code for "A Bilingual Generative Transformer for Semantic Sentence Embedding" published at EMNLP 2020.
jwieting/emnlp2017
Code and data to train and use models from "Learning Paraphrastic Sentence Embeddings from Back-Translated Bitext"
jwieting/tacl2015
Matlab code for training a recursive neural network to learn a paraphrase model from PPBD. Also includes matlab code for training paragram word embeddings. This code was used to obtain results in our 2015 TACL paper "From Paraphrase Database to Compositional Paraphrase Model and Back."