cangoksen's Stars
lm-sys/FastChat
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
karpathy/LLM101n
LLM101n: Let's build a Storyteller
Dao-AILab/flash-attention
Fast and memory-efficient exact attention
dair-ai/ML-Papers-of-the-Week
🔥Highlighting the top ML papers every week.
jadore801120/attention-is-all-you-need-pytorch
A PyTorch implementation of the Transformer model in "Attention is All You Need".
huggingface/accelerate
🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support
mit-han-lab/streaming-llm
[ICLR 2024] Efficient Streaming Language Models with Attention Sinks
NVIDIA/cuda-samples
Samples for CUDA Developers which demonstrates features in CUDA Toolkit
NVIDIA/cutlass
CUDA Templates for Linear Algebra Subroutines
linux-surface/linux-surface
Linux Kernel for Surface Devices
microsoft/Codex-CLI
CLI tool that uses Codex to turn natural language commands into their Bash/ZShell/PowerShell equivalents
openai/blocksparse
Efficient GPU kernels for block-sparse matrix multiplication and convolution
davidtvs/pytorch-lr-finder
A learning rate range test implementation in PyTorch
okankop/Efficient-3DCNNs
PyTorch Implementation of "Resource Efficient 3D Convolutional Neural Networks", codes and pretrained models.
lucidrains/linear-attention-transformer
Transformer based on a variant of attention that is linear complexity in respect to sequence length
tomaarsen/attention_sinks
Extend existing LLMs way beyond the original training length with constant memory usage, without retraining
LeapLabTHU/Agent-Attention
Official repository of Agent Attention (ECCV2024)
lucidrains/local-attention
An implementation of local windowed attention for language modeling
tmabraham/diffusion_reading_group
Diffusion Reading Group at EleutherAI
prprbr/awesome-lifelong-continual-learning
A list of papers, blogs, datasets and software in the field of lifelong/continual machine learning
facebookresearch/SpinQuant
Code repo for the paper "SpinQuant LLM quantization with learned rotations"
pprp/Awesome-LLM-Quantization
Awesome list for LLM quantization
lezcano/expRNN
Optimization with orthogonal constraints and on general manifolds
Dao-AILab/fast-hadamard-transform
Fast Hadamard transform in CUDA, with a PyTorch interface
AlexanderMath/fasth
Code for the article "What if Neural Networks had SVDs?", to be presented as a spotlight paper at NeurIPS 2020.
VaticanCameos99/knowledge-distillation-for-unet
An implementation of Knowledge distillation for segmentation, to train a small (student) UNet from a larger (teacher) UNet thereby reducing the size of the network while achieving performance similar to the heavier model.
MattShannon/bandmat
A banded matrix library for python.
hhb072/OrthogonalTransformer
cdluminate/cdluminate
Goutam-Kelam/LayerOut
A new regularization technique that freezes the layers of the deep neural networks stochastically.