versae's Stars
hpcaitech/ColossalAI
Making large AI models cheaper, faster and more accessible
karpathy/minGPT
A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training
haotian-liu/LLaVA
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
reflex-dev/reflex
🕸️ Web apps in pure Python 🐍
outlines-dev/outlines
Structured Text Generation
XPixelGroup/DiffBIR
Official codes of DiffBIR: Towards Blind Image Restoration with Generative Diffusion Prior
dvlab-research/LongLoRA
Code and documents of LongLoRA and LongAlpaca (ICLR 2024 Oral)
fudan-zvg/Semantic-Segment-Anything
Automated dense category annotation engine that serves as the initial semantic labeling for the Segment Anything dataset (SA-1B).
KoljaB/RealtimeTTS
Converts text to speech in realtime
mlabonne/llm-datasets
High-quality datasets, tools, and concepts for LLM fine-tuning.
facebookresearch/schedule_free
Schedule-Free Optimization in PyTorch
apple/ml-fastvit
This repository contains the official implementation of the research paper, "FastViT: A Fast Hybrid Vision Transformer using Structural Reparameterization" ICCV 2023
tomaarsen/attention_sinks
Extend existing LLMs way beyond the original training length with constant memory usage, without retraining
lhao499/ringattention
Transformers with Arbitrarily Large Context
HazyResearch/m2
Repo for "Monarch Mixer: A Simple Sub-Quadratic GEMM-Based Architecture"
henrikbostrom/crepes
Python package for conformal prediction
cisnlp/simalign
Obtain Word Alignments using Pretrained Language Models (e.g., mBERT)
ayaka14732/llama-2-jax
JAX implementation of the Llama 2 model
dwzhu-pku/PoSE
Positional Skip-wise Training for Efficient Context Window Extension of LLMs to Extremely Length (ICLR 2024)
imoneoi/multipack_sampler
Multipack distributed sampler for fast padding-free training of LLMs
ayaka14732/jax-smi
JAX Synergistic Memory Inspector
davisyoshida/lorax
LoRA for arbitrary JAX models and functions
luweigen/whisper_streaming
Whisper realtime streaming for long speech-to-text transcription and translation
hamishivi/EasyLM
Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetuning, evaluating and serving LLMs in JAX/Flax.
Beomi/Gemma-EasyLM
Train GEMMA on TPU/GPU! (Codebase for training Gemma-Ko Series)
lucaslingle/mu_transformer
Transformer with Mu-Parameterization, implemented in Jax/Flax. Supports FSDP on TPU pods.
rpuggaardrode/praatpicture
Make Praat Picture style plots of acoustic data
MeLeLBGU/SaGe
Code for SaGe subword tokenizer (EACL 2023)
qurator-spk/sbb_images
Image Annotation Tool and Image Search
bnagy/ruzicka