yaof20's Stars
dair-ai/Prompt-Engineering-Guide
🐙 Guides, papers, lecture, notebooks and resources for prompt engineering
geekan/MetaGPT
🌟 The Multi-Agent Framework: First AI Software Company, Towards Natural Language Programming
microsoft/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
openai/chatgpt-retrieval-plugin
The ChatGPT Retrieval Plugin lets you easily find personal or work documents by asking questions in natural language.
huggingface/peft
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
meta-llama/llama-recipes
Scripts for fine-tuning Meta Llama with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment. Demo apps to showcase Meta Llama for WhatsApp & Messenger.
Stability-AI/StableLM
StableLM: Stability AI Language Models
huggingface/trl
Train transformer language models with reinforcement learning.
artidoro/qlora
QLoRA: Efficient Finetuning of Quantized LLMs
huggingface/text-generation-inference
Large Language Model Text Generation Inference
NVIDIA/cutlass
CUDA Templates for Linear Algebra Subroutines
arcee-ai/mergekit
Tools for merging pretrained large language models.
Facico/Chinese-Vicuna
Chinese-Vicuna: A Chinese Instruction-following LLaMA-based Model —— 一个中文低资源的llama+lora方案,结构参考alpaca
yuanzhoulvpi2017/zero_nlp
中文nlp解决方案(大模型、数据、模型、训练、推理)
zjunlp/LLMAgentPapers
Must-read Papers on LLM Agents.
microsoft/DeepSpeed-MII
MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.
gpu-mode/resource-stream
GPU programming related news and material links
AGI-Edgerunners/LLM-Agents-Papers
A repo lists papers related to LLM based agent
srush/awesome-o1
A bibliography and survey of the papers surrounding o1
hollobit/GenAI_LLM_timeline
ChatGPT, GenerativeAI and LLMs Timeline
bigcode-project/bigcode-evaluation-harness
A framework for the evaluation of autoregressive code generation language models.
SinclairCoder/Instruction-Tuning-Papers
Reading list of Instruction-tuning. A trend starts from Natrural-Instruction (ACL 2022), FLAN (ICLR 2022) and T0 (ICLR 2022).
alibaba/Megatron-LLaMA
Best practice for training LLaMA models in Megatron-LM
Jack47/hack-SysML
The road to hack SysML and become an system expert
microsoft/GRIN-MoE
GRadient-INformed MoE
Hannibal046/nanoRWKV
The nanoGPT-style implementation of RWKV Language Model - an RNN with GPT-level LLM performance.
luban-agi/Awesome-Tool-Learning
A curated list of papers and applications on tool learning.
RZFan525/Awesome-ScalingLaws
A curated list of awesome resources dedicated to Scaling Laws for LLMs
THUlawtech/MUSER
kyegomez/FlashLora
FlashAttention2.0 with Lora