Pinned Repositories
gpt-neo
An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.
arabert
Pre-trained Transformers for the Arabic Language Understanding and Generation (Arabic BERT, Arabic GPT2, Arabic Electra)
BERTopic
Leveraging BERT and c-TF-IDF to create easily interpretable topics.
Deep-Learning
gpt-neo
An implementation of model parallel GPT2& GPT3-like models, with the ability to scale up to full GPT3 sizes (and possibly more!), using the mesh-tensorflow library.
gpt-neox
An implementation of model parallel GPT-3-like models on GPUs, based on the DeepSpeed library. Designed to be able to train models in the hundreds of billions of parameters or larger.
huggingtweets
Tweet Generation with Huggingface
mesh-transformer-jax
Model parallel transformers in JAX and Haiku
promptsource
Toolkit for collecting and applying prompts
Stanza-FastAPI-docker
Stanza parser as a service.
srulikbd's Repositories
srulikbd/Deep-Learning
srulikbd/arabert
Pre-trained Transformers for the Arabic Language Understanding and Generation (Arabic BERT, Arabic GPT2, Arabic Electra)
srulikbd/BERTopic
Leveraging BERT and c-TF-IDF to create easily interpretable topics.
srulikbd/gpt-neo
An implementation of model parallel GPT2& GPT3-like models, with the ability to scale up to full GPT3 sizes (and possibly more!), using the mesh-tensorflow library.
srulikbd/gpt-neox
An implementation of model parallel GPT-3-like models on GPUs, based on the DeepSpeed library. Designed to be able to train models in the hundreds of billions of parameters or larger.
srulikbd/huggingtweets
Tweet Generation with Huggingface
srulikbd/mesh-transformer-jax
Model parallel transformers in JAX and Haiku
srulikbd/promptsource
Toolkit for collecting and applying prompts
srulikbd/Stanza-FastAPI-docker
Stanza parser as a service.