TiankaiHang's Stars
yt-dlp/yt-dlp
A feature-rich command-line audio/video downloader
2noise/ChatTTS
A generative speech model for daily dialogue.
karpathy/LLM101n
LLM101n: Let's build a Storyteller
black-forest-labs/flux
Official inference repo for FLUX.1 models
Zeyi-Lin/HivisionIDPhotos
⚡️HivisionIDPhotos: a lightweight and efficient AI ID photos tools. 一个轻量级的AI证件照制作算法。
OpenBMB/MiniCPM-V
MiniCPM-V 2.6: A GPT-4V Level MLLM for Single Image, Multi Image and Video on Your Phone
facebookresearch/segment-anything-2
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
facebookresearch/seamless_communication
Foundational Models for State-of-the-Art Speech and Text Translation
ShiArthur03/ShiArthur03
sgl-project/sglang
SGLang is a fast serving framework for large language models and vision language models.
linkedin/Liger-Kernel
Efficient Triton Kernels for LLM Training
jdepoix/youtube-transcript-api
This is a python API which allows you to get the transcript/subtitles for a given YouTube video. It also works for automatically generated subtitles and it does not require an API key nor a headless browser, like other selenium based solutions do!
sc0tfree/updog
Updog is a replacement for Python's SimpleHTTPServer. It allows uploading and downloading via HTTP/S, can set ad hoc SSL certificates and use http basic auth.
cuda-mode/lectures
Material for cuda-mode lectures
kvcache-ai/Mooncake
Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.
NVIDIA/TransformerEngine
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
bghira/SimpleTuner
A general fine-tuning kit geared toward diffusion models.
LTH14/mar
PyTorch implementation of MAR+DiffLoss https://arxiv.org/abs/2406.11838
xdit-project/xDiT
xDiT: A Scalable Inference Engine for Diffusion Transformers (DiTs) with Massive Parallelism
kvcache-ai/ktransformers
A Flexible Framework for Experiencing Cutting-edge LLM Inference Optimizations
efeslab/Nanoflow
A throughput-oriented high-performance serving framework for LLMs
buoyancy99/diffusion-forcing
code for "Diffusion Forcing: Next-token Prediction Meets Full-Sequence Diffusion"
pytorch-labs/attention-gym
Helpful tools and examples for working with flex-attention
EurekaLabsAI/mlp
The Multilayer Perceptron Language Model
feifeibear/long-context-attention
USP: Unified (a.k.a. Hybrid, 2D) Sequence Parallel Attention for Long Context Transformers Model Training and Inference
google-research/syn-rep-learn
Learning from synthetic data - code and models
aredden/flux-fp8-api
Flux diffusion model implementation using quantized fp8 matmul & remaining layers use faster half precision accumulate, which is ~2x faster on consumer devices.
zju-pi/diff-sampler
An open-source toolbox for fast sampling of diffusion models. Official implementations of our works published in ICML, NeurIPS, CVPR.
RockeyCoss/SPO
Aesthetic Post-Training Diffusion Models from Generic Preferences with Step-by-step Preference Optimization
BruceQFWang/ICML2024-PEG