will-thompson-k's Stars
vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
janhq/jan
Jan is an open source alternative to ChatGPT that runs 100% offline on your computer. Multiple engine support (llama.cpp, TensorRT-LLM)
spotify/luigi
Luigi is a Python module that helps you build complex pipelines of batch jobs. It handles dependency resolution, workflow management, visualization etc. It also comes with Hadoop support built in.
microsoft/LoRA
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
Lightning-AI/litgpt
20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.
karpathy/micrograd
A tiny scalar-valued autograd engine and a neural net library on top of it with PyTorch-like API
mistralai/mistral-inference
Official inference library for Mistral models
facebookresearch/xformers
Hackable and optimized Transformers building blocks, supporting a composable construction.
axolotl-ai-cloud/axolotl
Go ahead and axolotl questions
WooooDyy/LLM-Agent-Paper-List
The paper list of the 86-page paper "The Rise and Potential of Large Language Model Based Agents: A Survey" by Zhiheng Xi et al.
langchain-ai/langgraph
Build resilient language agents as graphs.
huggingface/alignment-handbook
Robust recipes to align language models with human and AI preferences
google-deepmind/open_spiel
OpenSpiel is a collection of environments and algorithms for research in general reinforcement learning and search/planning in games.
NVIDIA/NeMo-Guardrails
NeMo Guardrails is an open-source toolkit for easily adding programmable guardrails to LLM-based conversational systems.
openai/transformer-debugger
yandex/YaLM-100B
Pretrained language model with 100B parameters
pytorch/torchtitan
A native PyTorch Library for large model training
openai/weak-to-strong
EdinburghNLP/awesome-hallucination-detection
List of papers on hallucination detection in LLMs.
gordicaleksa/get-started-with-JAX
The purpose of this repo is to make it easy to get started with JAX, Flax, and Haiku. It contains my "Machine Learning with JAX" series of tutorials (YouTube videos and Jupyter Notebooks) as well as the content I found useful while learning about the JAX ecosystem.
google/paxml
Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimentation and parallelization, and has demonstrated industry leading model flop utilization rates.
lucidrains/triton-transformer
Implementation of a Transformer, but completely in Triton
huggingface/llm-swarm
Manage scalable open LLM inference endpoints in Slurm clusters
kingoflolz/swarm-jax
Swarm training framework using Haiku + JAX + Ray for layer parallel transformer language models on unreliable, heterogeneous nodes
lucidrains/speculative-decoding
Explorations into some recent techniques surrounding speculative decoding
Mihaiii/llm_steer
Steer LLM outputs towards a certain topic/subject and enhance response capabilities using activation engineering by adding steering vectors
teknium1/LLM-Benchmark-Logs
Just a bunch of benchmark logs for different LLMs
XZhang97666/AlpaCare
hwchase17/chain-of-verification
yisding/litllm