vasqu's Stars
lucidrains/minGRU-pytorch
Implementation of the proposed minGRU in Pytorch
Adibvafa/CodonTransformer
CodonTransformer: The ultimate tool for codon optimization, optimizing DNA sequences for heterologous protein expression across 164 species.
Modalities/modalities
Modalities, a PyTorch-native framework for distributed and reproducible foundation model training.
goombalab/phi-mamba
Official implementation of Phi-Mamba. A MOHAWK-distilled model (Transformers to SSMs: Distilling Quadratic Knowledge to Subquadratic Models)
kyutai-labs/moshi
sustcsonglin/flash-linear-attention
Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton
linkedin/Liger-Kernel
Efficient Triton Kernels for LLM Training
goombalab/hydra
Official implementation of "Hydra: Bidirectional State Space Models Through Generalized Matrix Mixers"
Hprairie/Bi-Mamba2
A Triton Kernel for incorporating Bi-Directionality in Mamba2
black-forest-labs/flux
Official inference repo for FLUX.1 models
facebookresearch/sam2
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
InternLM/xtuner
An efficient, flexible and full-featured toolkit for fine-tuning LLM (InternLM2, Llama3, Phi3, Qwen, Mistral, ...)
LSX-UniWue/SuperGLEBer
German Language Understanding Evaluation Benchmark @NAACL24
marzenakrp/nocha
hsiehjackson/RULER
This repo contains the source code for RULER: What’s the Real Context Size of Your Long-Context Language Models?
karpathy/LLM101n
LLM101n: Let's build a Storyteller
Dao-AILab/causal-conv1d
Causal depthwise conv1d in CUDA, with a PyTorch interface
HazyResearch/based
Code for exploring Based models from "Simple linear attention language models balance the recall-throughput tradeoff"
NX-AI/xlstm
Official repository of the xLSTM.
Dao-AILab/flash-attention
Fast and memory-efficient exact attention
yyyujintang/Awesome-Mamba-Papers
Awesome Papers related to Mamba.
microsoft/FILM
Official repo for "Make Your LLM Fully Utilize the Context"
kolinko/effort
An implementation of bucketMul LLM inference
kuleshov-group/caduceus
Bi-Directional Equivariant Long-Range DNA Sequence Modeling
flbbb/locost-summarization
xfactlab/orpo
Official repository for ORPO
facebookresearch/schedule_free
Schedule-Free Optimization in PyTorch
Yale-LILY/DYLE
Repository for ACL'22 paper: Dynamic Latent Extraction for Abstractive Long-Input Summarization
abertsch72/unlimiformer
Public repo for the NeurIPS 2023 paper "Unlimiformer: Long-Range Transformers with Unlimited Length Input"
psunlpgroup/Summ-N
Code for ACL 2022 Paper "SUMM^N: A Multi-Stage Summarization Framework for Long Input Dialogues and Documents"