D0z1ngShark's Stars
jaywcjlove/awesome-mac
Now we have become very big, Different from the original idea. Collect premium software in various categories.
karpathy/nanoGPT
The simplest, fastest repository for training/finetuning medium-sized GPTs.
hiyouga/LLaMA-Factory
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
rasbt/LLMs-from-scratch
Implementing a ChatGPT-like LLM in PyTorch from scratch, step by step
linexjlin/GPTs
leaked prompts of GPTs
ymcui/Chinese-LLaMA-Alpaca
中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
LlamaFamily/Llama-Chinese
Llama中文社区,Llama3在线体验和微调模型已开放,实时汇总最新Llama3学习资料,已将所有代码更新适配Llama3,构建最好的中文Llama大模型,完全开源可商用
naklecha/llama3-from-scratch
llama3 implementation one matrix multiplication at a time
NVIDIA/Megatron-LM
Ongoing research training transformer models at scale
dair-ai/ML-Papers-of-the-Week
🔥Highlighting the top ML papers every week.
mistralai/mistral-inference
Official inference library for Mistral models
huggingface/tokenizers
💥 Fast State-of-the-Art Tokenizers optimized for Research and Production
facebookresearch/xformers
Hackable and optimized Transformers building blocks, supporting a composable construction.
Morizeyao/GPT2-Chinese
Chinese version of GPT2 training code, using BERT tokenizer.
ymcui/Chinese-LLaMA-Alpaca-2
中文LLaMA-2 & Alpaca-2大模型二期项目 + 64K超长上下文模型 (Chinese LLaMA-2 & Alpaca-2 LLMs with 64K long context models)
lyogavin/airllm
AirLLM 70B inference with single 4GB GPU
microsoft/LLMLingua
[EMNLP'23, ACL'24] To speed up LLMs' inference and enhance LLM's perceive of key information, compress the prompt and KV-Cache, which achieves up to 20x compression with minimal performance loss.
stanford-futuredata/ColBERT
ColBERT: state-of-the-art neural search (SIGIR'20, TACL'21, NeurIPS'21, NAACL'22, CIKM'22, ACL'23, EMNLP'23)
OpenRLHF/OpenRLHF
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention)
Tele-AI/Telechat
XueFuzhao/OpenMoE
A family of open-sourced Mixture-of-Experts (MoE) Large Language Models
OpenLMLab/MOSS-RLHF
MOSS-RLHF
charent/ChatLM-mini-Chinese
中文对话0.2B小模型(ChatLM-Chinese-0.2B),开源所有数据集来源、数据清洗、tokenizer训练、模型预训练、SFT指令微调、RLHF优化等流程的全部代码。支持下游任务sft微调,给出三元组信息抽取微调示例。
dandelionsllm/pandallm
Panda项目是于2023年5月启动的开源海外中文大语言模型项目,致力于大模型时代探索整个技术栈,旨在推动中文自然语言处理领域的创新和合作。
deepseek-ai/DeepSeek-MoE
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models
HIT-SCIR/Chinese-Mixtral-8x7B
中文Mixtral-8x7B(Chinese-Mixtral-8x7B)
AviSoori1x/makeMoE
From scratch implementation of a sparse mixture of experts language model inspired by Andrej Karpathy's makemore :)
charent/Phi2-mini-Chinese
Phi2-Chinese-0.2B 从0开始训练自己的Phi2中文小模型,支持接入langchain加载本地知识库做检索增强生成RAG。Training your own Phi2 small chat model from scratch.
GAIR-NLP/MathPile
[NeurlPS D&B 2024] Generative AI for Math: MathPile
yanqiangmiffy/how-to-train-tokenizer
怎么训练一个LLM分词器