Pinned Repositories
nanoRWKV
RWKV in nanoGPT style
nanoGPT
The simplest, fastest repository for training/finetuning medium-sized GPTs.
jax
Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more
flax
Flax is a neural network library for JAX that is designed for flexibility.
DeepSeek-V2
DeepSeek-V2: A Strong, Economical, and Efficient Mixture-of-Experts Language Model
equinox
Elegant easy-to-use neural networks + scientific computing in JAX. https://docs.kidger.site/equinox/
flash-linear-attention
Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton
nanoRWKV
RWKV in nanoGPT style
nanodo
gated_linear_attention
faresobeid's Repositories
faresobeid/nanoRWKV
RWKV in nanoGPT style