ericxian1997's Stars
microsoft/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
geekan/HowToLiveLonger
程序员延寿指南 | A programmer's guide to live longer
meta-llama/llama3
The official Meta Llama 3 GitHub site
facebookresearch/fastText
Library for fast text representation and classification.
spotify/luigi
Luigi is a Python module that helps you build complex pipelines of batch jobs. It handles dependency resolution, workflow management, visualization etc. It also comes with Hadoop support built in.
Lightning-AI/litgpt
20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.
jzhang38/TinyLlama
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
openai/transformer-debugger
atfortes/Awesome-LLM-Reasoning
Reasoning in Large Language Models: Papers and Resources, including Chain-of-Thought and OpenAI o1 🍓
opendilab/PPOxFamily
PPO x Family DRL Tutorial Course(决策智能入门级公开课:8节课帮你盘清算法理论,理顺代码逻辑,玩转决策AI应用实践 )
GAIR-NLP/O1-Journey
O1 Replication Journey: A Strategic Progress Report – Part I
jquesnelle/yarn
YaRN: Efficient Context Window Extension of Large Language Models
mlfoundations/dclm
DataComp for Language Models
yaodongC/awesome-instruction-dataset
A collection of open-source dataset to train instruction-following LLMs (ChatGPT,LLaMA,Alpaca)
trotsky1997/MathBlackBox
SimpleBerry/LLaMA-O1
Large Reasoning Models
NVIDIA/NeMo-Curator
Scalable data pre processing and curation toolkit for LLMs
NVIDIA/NeMo-Aligner
Scalable toolkit for efficient model alignment
madaan/self-refine
LLMs can generate feedback on their work, use it to improve the output, and repeat this process iteratively.
FranxYao/Long-Context-Data-Engineering
Implementation of paper Data Engineering for Scaling Language Models to 128K Context
MARIO-Math-Reasoning/Super_MARIO
facebookresearch/MetaICL
An original implementation of "MetaICL Learning to Learn In Context" by Sewon Min, Mike Lewis, Luke Zettlemoyer and Hannaneh Hajishirzi
OFA-Sys/InsTag
InsTag: A Tool for Data Analysis in LLM Supervised Fine-tuning
withinmiaov/A-Survey-on-Mixture-of-Experts
The official GitHub page for the survey paper "A Survey on Mixture of Experts".
src-d/minhashcuda
Weighted MinHash implementation on CUDA (multi-gpu).
TIGER-AI-Lab/LongICLBench
Code and Data for "Long-context LLMs Struggle with Long In-context Learning"
allenai/allennlp-reading-comprehension
OpenLMLab/scaling-rope
code for Scaling Laws of RoPE-based Extrapolation
hughbzhang/o1_inference_scaling_laws
Replicating O1 inference-time scaling laws
tml-epfl/icl-alignment
Is In-Context Learning Sufficient for Instruction Following in LLMs?