Pinned Repositories
autodiff
Introduction to autodiff and autograd
documents
dpu-utils
Utilities used by the Deep Program Understanding team
embeddings
Fast, DB Backed pretrained word embeddings for natural language processing.
fast-hadamard-transform
Fast Hadamard transform in CUDA, with a PyTorch interface
flash-attention
Fast and memory-efficient exact attention
optimization_examples
pashminacameron.github.io
py-repo-root
QuaRot
[Fork] Code for QuaRot, an end-to-end 4-bit inference of large language models.
pashminacameron's Repositories
pashminacameron/py-repo-root
pashminacameron/optimization_examples
pashminacameron/autodiff
Introduction to autodiff and autograd
pashminacameron/documents
pashminacameron/dpu-utils
Utilities used by the Deep Program Understanding team
pashminacameron/embeddings
Fast, DB Backed pretrained word embeddings for natural language processing.
pashminacameron/fast-hadamard-transform
Fast Hadamard transform in CUDA, with a PyTorch interface
pashminacameron/flash-attention
Fast and memory-efficient exact attention
pashminacameron/pashminacameron.github.io
pashminacameron/QuaRot
[Fork] Code for QuaRot, an end-to-end 4-bit inference of large language models.
pashminacameron/semantic-kernel
Integrate cutting-edge LLM technology quickly and easily into your apps
pashminacameron/transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.