proceduralia's Stars
donnemartin/system-design-primer
Learn how to design large-scale systems. Prep for the system design interview. Includes Anki flashcards.
OpenInterpreter/open-interpreter
A natural language interface for computers
charlax/professional-programming
A collection of learning resources for curious software engineers
mlabonne/llm-course
Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.
maybe-finance/maybe
The OS for your personal finances
bloomberg/memray
Memray is a memory profiler for Python
stas00/ml-engineering
Machine Learning Engineering Open Book
TencentARC/PhotoMaker
PhotoMaker
huggingface/accelerate
🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support
eureka-research/Eureka
Official Repository for "Eureka: Human-Level Reward Design via Coding Large Language Models" (ICLR 2024)
sgl-project/sglang
SGLang is a structured generation language designed for large language models (LLMs). It makes your interaction with models faster and more controllable.
webdataset/webdataset
A high-performance Python-based I/O system for large (and small) deep learning problems, with strong support for PyTorch.
ysymyth/ReAct
[ICLR 2023] ReAct: Synergizing Reasoning and Acting in Language Models
google/maxtext
A simple, performant and scalable Jax LLM!
huggingface/nanotron
Minimalistic large language model 3D-parallelism training
mistralai/megablocks-public
tatsu-lab/alpaca_farm
A simulation framework for RLHF and alternatives. Develop your RLHF method without collecting human data.
ezelikman/quiet-star
Code for Quiet-STaR
RLHFlow/RLHF-Reward-Modeling
Recipes to train reward model for RLHF.
carlosferrazza/humanoid-bench
allenai/reward-bench
RewardBench: the first evaluation tool for reward models.
facebookincubator/dynolog
Dynolog is a telemetry daemon for performance monitoring and tracing. It exports metrics from different components in the system like the linux kernel, CPU, disks, Intel PT, GPUs etc. Dynolog also integrates with pytorch and can trigger traces for distributed training applications.
SalesforceAIResearch/DiffusionDPO
Code for "Diffusion Model Alignment Using Direct Preference Optimization"
YifeiZhou02/ArCHer
Research Code for "ArCHer: Training Language Model Agents via Hierarchical Multi-Turn RL"
facebookresearch/RLCD
Reproduction of "RLCD Reinforcement Learning from Contrast Distillation for Language Model Alignment
roeehendel/icl_task_vectors
facebookresearch/rlfh-gen-div
This is code for most of the experiments in the paper Understanding the Effects of RLHF on LLM Generalisation and Diversity
deeplearning-wisc/args
Jiuzhouh/Uncertainty-Aware-Language-Agent
This is the official repo for Towards Uncertainty-Aware Language Agent.
spikedoanz/weenygrad
Minimalist vector AD