yaodongyu's Stars
NX-AI/vision-lstm
xLSTM as Generic Vision Backbone
lucidrains/transfusion-pytorch
Pytorch implementation of Transfusion, "Predict the Next Token and Diffuse Images with One Multi-Modal Model", from MetaAI
tyshiwo1/DiM-DiffusionMamba
The official implementation of DiM: Diffusion Mamba for Efficient High-Resolution Image Synthesis
linkedin/Liger-Kernel
Efficient Triton Kernels for LLM Training
openai/consistency_models
Official repo for consistency models.
Lightning-AI/litgpt
20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.
hao-ai-lab/Consistency_LLM
[ICML 2024] CLLMs: Consistency Large Language Models
Cranial-XIX/longhorn
Official PyTorch Implementation of the Longhorn Deep State Space Model
gnobitab/RectifiedFlow
Official Implementation of Rectified Flow (ICLR2023 Spotlight)
facebookresearch/dpcap
Codebase for the ICML 2024 paper "Differentially Private Representation Learning via Image Captioning"
llm-attacks/llm-attacks
Universal and Transferable Attacks on Aligned Language Models
ToyotaResearchInstitute/gradient-estimation-sampler
Code for the paper "Interpreting and Improving Diffusion Models from an Optimization Perspective", appearing in ICML 2024
luping-liu/PNDM
The official implementation for Pseudo Numerical Methods for Diffusion Models on Manifolds (PNDM, PLMS | ICLR2022)
Ma-Lab-Berkeley/CRATE
Code for CRATE (Coding RAte reduction TransformEr).
LTH14/mar
PyTorch implementation of MAR+DiffLoss https://arxiv.org/abs/2406.11838
hao-ai-lab/LookaheadDecoding
[ICML 2024] Break the Sequential Dependency of LLM Inference Using Lookahead Decoding
openai/prm800k
800,000 step-level correctness labels on LLM solutions to MATH problems
awslabs/fast-differential-privacy
Fast, memory-efficient, scalable optimization of deep learning with differential privacy
lucidrains/denoising-diffusion-pytorch
Implementation of Denoising Diffusion Probabilistic Model in Pytorch
lucidrains/routing-transformer
Fully featured implementation of Routing Transformer
lucidrains/sinkhorn-transformer
Sinkhorn Transformer - Practical implementation of Sparse Sinkhorn Attention
lucidrains/linear-attention-transformer
Transformer based on a variant of attention that is linear complexity in respect to sequence length
twistedcubic/attention-rank-collapse
[ICML 2021 Oral] We show pure attention suffers rank collapse, and how different mechanisms combat it.
lucidrains/performer-pytorch
An implementation of Performer, a linear attention-based transformer, in Pytorch
lucidrains/reformer-pytorch
Reformer, the efficient Transformer, in Pytorch
lucidrains/x-transformers
A concise but complete full-attention transformer with a set of promising experimental features from various papers
jbloomAus/SAELens
Training Sparse Autoencoders on Language Models
frank-xwang/UnSAM
[NeurIPS 2024] Code release for "Segment Anything without Supervision"
ShishirPatil/gorilla
Gorilla: Training and Evaluating LLMs for Function Calls (Tool Calls)
xlang-ai/xlang-paper-reading
Paper collection on building and evaluating language model agents via executable language grounding