BaileyWei's Stars
yuanzhoulvpi2017/zero_nlp
中文nlp解决方案(大模型、数据、模型、训练、推理)
xlang-ai/UnifiedSKG
[EMNLP 2022] Unifying and multi-tasking structured knowledge grounding with language models
THUDM/ChatGLM-6B
ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型
NVIDIA/Megatron-LM
Ongoing research training transformer models at scale
tloen/llama-int8
Quantized inference code for LLaMA models
tatsu-lab/stanford_alpaca
Code and documentation to train Stanford's Alpaca models, and generate the data.
pyg-team/pytorch_geometric
Graph Neural Network Library for PyTorch
shawwn/llama-dl
High-speed download of LLaMA, Facebook's 65B parameter GPT model
THUDM/GLM-130B
GLM-130B: An Open Bilingual Pre-Trained Model (ICLR 2023)
anthropics/hh-rlhf
Human preference data for "Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback"
meta-llama/llama
Inference code for Llama models
microsoft/torchscale
Foundation Architecture for (M)LLMs
openai/gpt-2
Code for the paper "Language Models are Unsupervised Multitask Learners"
microsoft/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
codemayq/chinese-chatbot-corpus
中文公开聊天语料库
huggingface/transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
OpenBMB/BMTrain
Efficient Training (including pre-training and fine-tuning) for Big Models
JonasGeiping/cramming
Cramming the training of a (BERT-type) language model into limited compute.
hpcaitech/ColossalAI
Making large AI models cheaper, faster and more accessible
lucidrains/electra-pytorch
A simple and working implementation of Electra, the fastest way to pretrain language models from scratch, in Pytorch
andy-yangz/easy_bert_pretrain
The very easy BERT pretrain process by using tokenizers and transformers repos
cosmoquester/transformers-bart-pretrain
Script to pre-train hugginface transformers BART with Tensorflow 2
Tencent/TencentPretrain
Tencent Pre-training framework in PyTorch & Pre-trained Model Zoo
BlinkDL/ChatRWKV
ChatRWKV is like ChatGPT but powered by RWKV (100% RNN) language model, and open source.
Spico197/DocEE
🕹️ A toolkit for document-level event extraction, containing some SOTA model implementations.
f/awesome-chatgpt-prompts
This repo includes ChatGPT prompt curation to use ChatGPT better.
dyweb/awesome-resume-for-chinese
:page_facing_up: 适合中文的简历模板收集(LaTeX,HTML/JS and so on)由 @hoochanlon 维护
arasgungore/arasgungore-CV
My curriculum vitae (CV) written using LaTeX.
zhanglj37/Tutorial-on-PhD-Application
Tutorial on PhD Application
BlinkDL/RWKV-LM
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.