longxudou
LLM Researcher @sail-sg. Maintainer ⚓️Sailor | 🔱Sailor2 | 🚢 SailCraft | 🧭 SailCompass
Research Scientist @ Sea AI LabHarbin
longxudou's Stars
hiyouga/LLaMA-Factory
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
triton-lang/triton
Development repository for the Triton language and compiler
MathFoundationRL/Book-Mathematical-Foundation-of-Reinforcement-Learning
This is the homepage of a new book entitled "Mathematical Foundations of Reinforcement Learning."
arcee-ai/mergekit
Tools for merging pretrained large language models.
kvcache-ai/Mooncake
Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.
togethercomputer/MoA
Together Mixture-Of-Agents (MoA) – 65.1% on AlpacaEval with OSS models
NVIDIA/TransformerEngine
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
mit-han-lab/once-for-all
[ICLR 2020] Once for All: Train One Network and Specialize it for Efficient Deployment
huggingface/lighteval
Lighteval is your all-in-one toolkit for evaluating LLMs across multiple backends
RLHFlow/RLHF-Reward-Modeling
Recipes to train reward model for RLHF.
horseee/LLM-Pruner
[NeurIPS 2023] LLM-Pruner: On the Structural Pruning of Large Language Models. Support Llama-3/3.1, Llama-2, LLaMA, BLOOM, Vicuna, Baichuan, TinyLlama, etc.
NVIDIA/NeMo-Curator
Scalable data pre processing and curation toolkit for LLMs
locuslab/wanda
A simple and effective LLM pruning approach.
allenai/OLMoE
OLMoE: Open Mixture-of-Experts Language Models
princeton-nlp/LLM-Shearing
[ICLR 2024] Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruning
TencentARC/LLaMA-Pro
[ACL 2024] Progressive LLaMA with Block Expansion.
NVIDIA/NeMo-Framework-Launcher
Provides end-to-end model development pipelines for LLMs and Multimodal models that can be launched on-prem or cloud-native.
NVlabs/Minitron
A family of compressed models obtained via pruning and knowledge distillation
arcee-ai/EvolKit
EvolKit is an innovative framework designed to automatically enhance the complexity of instructions used for fine-tuning Large Language Models (LLMs).
Outsider565/LoRA-GA
sail-sg/regmix
[ICLR 2025] 🧬 RegMix: Data Mixture as Regression for Language Model Pre-training (Spotlight)
pprp/Pruner-Zero
Pruner-Zero: Evolving Symbolic Pruning Metric from scratch for LLMs
sail-sg/scaling-with-vocab
[NeurIPS-2024] 📈 Scaling Laws with Vocabulary: Larger Models Deserve Larger Vocabularies https://arxiv.org/abs/2407.13623
Qichuzyy/POA
Official implementation of ECCV24 paper: POA
zirui-HIT/DAC
zhxlia/FLEXTAF
zirui-HIT/Fused
zirui-HIT/Encore
OpenDFM/EST