Pinned Repositories
chessformer
Chessformer
flash-attention
Fast and memory-efficient exact attention
flashT5
A fast implementation of T5/UL2 in PyTorch using Flash Attention
portfolio
Personal Portfolio in Machine Learning
flashT5
A fast implementation of T5/UL2 in PyTorch using Flash Attention
rustlm
RustLM : An efficient Rust CTC Decoder supporting external language models
triton-rust
An api for interfacing Nvidia Trition Inference Server with Rust
flash-attention
Fast and memory-efficient exact attention
lightly
A python library for self-supervised learning on images.
pytorch
Tensors and Dynamic neural networks in Python with strong GPU acceleration
b-albar's Repositories
b-albar/chessformer
Chessformer
b-albar/FlagAttention
A collection of memory efficient attention operators implemented in the Triton language.
b-albar/portfolio
Personal Portfolio in Machine Learning
b-albar/rebased
Official implementation of the paper "Linear Transformers with Learnable Kernel Functions are Better In-Context Models"