tokenizer-decode's Stars
huggingface/peft
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
AGI-Edgerunners/LLM-Adapters
Code for our EMNLP 2023 Paper: "LLM-Adapters: An Adapter Family for Parameter-Efficient Fine-Tuning of Large Language Models"
superagent-ai/superagent
🥷 Run AI-agents with an API
a-real-ai/pywinassistant
The first open source Large Action Model generalist Artificial Narrow Intelligence that controls completely human user interfaces by only using natural language. PyWinAssistant utilizes Visualization-of-Thought Elicits Spatial Reasoning in Large Language Models.
leloykun/flash-hyperbolic-attention-minimal
Flash Hyperbolic Attention in ~[...] lines of CUDA
tspeterkim/flash-attention-minimal
Flash Attention in ~100 lines of CUDA (forward pass only)
Antlera/nanoGPT-moe
Enable moe for nanogpt.
karpathy/llm.c
LLM training in simple, raw C/CUDA
EleutherAI/gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries
moritztng/fltr
Like grep but for natural language questions. Based on Mistral 7B or Mixtral 8x7B.
labmlai/annotated_deep_learning_paper_implementations
🧑🏫 60 Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠
pytorch-labs/gpt-fast
Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.
flexflow/FlexFlow
FlexFlow Serve: Low-Latency, High-Performance LLM Serving
johnma2006/mamba-minimal
Simple, minimal implementation of the Mamba SSM in one file of PyTorch.
xue160709/Local-LLM-User-Guideline
lucidrains/x-transformers
A simple but complete full-attention transformer with a set of promising experimental features from various papers
mit-han-lab/llm-awq
[MLSys 2024 Best Paper Award] AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
stas00/ml-engineering
Machine Learning Engineering Open Book
pytorch/pytorch
Tensors and Dynamic neural networks in Python with strong GPU acceleration
numba/numba
NumPy aware dynamic Python compiler using LLVM