Pinned Repositories
flash_linear
Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton
lineargru
Implementation for MatMul-free LM.
pytorch_mamba
A simple and efficient Mamba implementation in PyTorch and MLX.
S5
sparse_mamba
uC-chip-interface-teensy41
neurobench
Benchmark harness and baseline results for the NeuroBench algorithm track.
Maxtimer97's Repositories
Maxtimer97/flash_linear
Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton
Maxtimer97/lineargru
Implementation for MatMul-free LM.
Maxtimer97/pytorch_mamba
A simple and efficient Mamba implementation in PyTorch and MLX.
Maxtimer97/S5
Maxtimer97/sparse_mamba
Maxtimer97/uC-chip-interface-teensy41