llm-training
There are 380 repositories under llm-training topic.
gitleaks/gitleaks
Find secrets with Gitleaks 🔑
liguodongiot/llm-action
本项目旨在分享大模型相关技术原理以及实战经验(大模型工程化、大模型应用落地)
ludwig-ai/ludwig
Low-code framework for building custom LLMs, neural networks, and other AI models
skypilot-org/skypilot
Run, manage, and scale AI workloads on any AI infrastructure. Use one system to access & manage all AI compute (Kubernetes, 17+ clouds, or on-prem).
linkedin/Liger-Kernel
Efficient Triton Kernels for LLM Training
InternLM/xtuner
A Next-Generation Training Engine Built for Ultra-Large MoE Models
h2oai/h2o-llmstudio
H2O LLM Studio - a framework and no-code GUI for fine-tuning LLMs. Documentation: https://docs.h2o.ai/h2o-llmstudio/
databricks/dbrx
Code examples and resources for DBRX, a large language model developed by Databricks
MoonshotAI/MoBA
MoBA: Mixture of Block Attention for Long-Context LLMs
dstackai/dstack
dstack is an open-source control plane for running development, training, and inference jobs on GPUs—across hyperscalers, neoclouds, or on-prem.
intelligent-machine-learning/dlrover
DLRover: An Automatic Distributed Deep Learning System
utkuozdemir/nvidia_gpu_exporter
Nvidia GPU exporter for prometheus using nvidia-smi binary
volcengine/veScale
A PyTorch Native LLM Training Framework
sail-sg/Adan
Adan: Adaptive Nesterov Momentum Algorithm for Faster Optimizing Deep Models
ghimiresunil/LLM-PowerHouse-A-Curated-Guide-for-Large-Language-Models-with-Custom-Training-and-Inferencing
LLM-PowerHouse: Unleash LLMs' potential through curated tutorials, best practices, and ready-to-use code for custom training and inferencing.
feifeibear/long-context-attention
USP: Unified (a.k.a. Hybrid, 2D) Sequence Parallel Attention for Long Context Transformers Model Training and Inference
rohan-paul/LLM-FineTuning-Large-Language-Models
LLM (Large Language Model) FineTuning
anarchy-ai/LLM-VM
irresponsible innovation. Try now at https://chat.dev/
mallorbc/Finetune_LLMs
Repo for fine-tuning Casual LLMs
FlagAI-Open/Aquila2
The official repo of Aquila2 series proposed by BAAI, including pretrained & chat large language models.
yinizhilian/ICLR2025-Papers-with-Code
历年ICLR论文和开源项目合集,包含ICLR2021、ICLR2022、ICLR2023、ICLR2024、ICLR2025.
InternLM/InternEvo
InternEvo is an open-sourced lightweight training framework aims to support model pre-training without the need for extensive dependencies.
tigerlab-ai/tiger
Open Source LLM toolkit to build trustworthy LLM applications. TigerArmor (AI safety), TigerRAG (embedding, RAG), TigerTune (fine-tuning)
aws-samples/awsome-distributed-training
Collection of best practices, reference architectures, model training examples and utilities to train large models on AWS.
openpsi-project/ReaLHF
Super-Efficient RLHF Training of LLMs with Parameter Reallocation
armbues/SiLLM
SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.
MLSys-Learner-Resources/Awesome-MLSys-Blogger
The repository has collected a batch of noteworthy MLSys bloggers (Algorithms/Systems)
zhuhanqing/APOLLO
APOLLO: SGD-like Memory, AdamW-level Performance; MLSys'25 Oustanding Paper Honorable Mention
promptslab/LLMtuner
FineTune LLMs in few lines of code (Text2Text, Text2Speech, Speech2Text)
neo4j-labs/text2cypher
collection of text2cypher datasets, evaluations, and finetuning instructions
bd4sur/Nano
电子鹦鹉 / Toy Language Model
Laz4rz/GPT-2
Following master Karpathy with GPT-2 implementation and training, writing lots of comments cause I have memory of a goldfish
substratusai/runbooks
Finetune LLMs on K8s by using Runbooks
MatX-inc/seqax
seqax = sequence modeling + JAX
shivendrra/SmallLanguageModel
a LLM cookbook, for building your own from scratch, all the way from gathering data to training a model
fiddlecube/fiddlecube-sdk
Generate ideal question-answers for testing RAG