ctlllll
ML PhD @Princeton. Life-long learner, hacker, and builder. Tech consultant & angel investor. Take a look at my recent research @FasterDecoding !
@PrincetonEarth
ctlllll's Stars
xai-org/grok-1
Grok open release
astral-sh/ruff
An extremely fast Python linter and code formatter, written in Rust.
myshell-ai/MeloTTS
High-quality multi-lingual text-to-speech library by MyShell.ai. Support English, Spanish, French, Chinese, Japanese and Korean.
kyegomez/BitNet
Implementation of "BitNet: Scaling 1-bit Transformers for Large Language Models" in pytorch
google/maxtext
A simple, performant and scalable Jax LLM!
jiaweizzhao/GaLore
GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
rcaloras/bashhub-client
:cloud: Bash history in the cloud. Indexed and searchable.
BAAI-DCAI/Bunny
A family of lightweight multimodal models.
carlini/yet-another-applied-llm-benchmark
A benchmark to evaluate language models on questions I've previously asked them to solve.
myshell-ai/AIlice
AIlice is a fully autonomous, general-purpose AI agent.
deepseek-ai/DeepSeek-Math
DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models
mit-han-lab/distrifuser
[CVPR 2024 Highlight] DistriFusion: Distributed Parallel Inference for High-Resolution Diffusion Models
zhuzilin/ring-flash-attention
Ring attention implementation with flash attention
lhao499/ringattention
Transformers with Arbitrarily Large Context
mlabonne/llm-autoeval
Automatically evaluate your LLMs in Google Colab
lucidrains/ring-attention-pytorch
Implementation of 💍 Ring Attention, from Liu et al. at Berkeley AI, in Pytorch
VIRL-Platform/VIRL
(ECCV 2024) Code for V-IRL: Grounding Virtual Intelligence in Real Life
catid/dora
Implementation of DoRA
erfanzar/EasyDeL
Accelerate, Optimize performance with streamlined training and serving options with JAX.
FasterDecoding/BitDelta
jxnl/n-levels-of-rag
jthickstun/anticipation
Anticipatory Autoregressive Models
felipemaiapolo/tinyBenchmarks
Evaluating LLMs with fewer examples
young-geng/scalax
A simple library for scaling up JAX programs
mit-han-lab/patch_conv
Patch convolution to avoid large GPU memory usage of Conv2D
luyug/magix
Supercharge huggingface transformers with model parallelism.
myshell-ai/MyShell-TTS-Subnet
euclaise/supertrainer2000
erfanzar/FJFormer
paralleled/unparalleled computational with FJFormer
sekstini/basedxl