Pinned Repositories
gptq
Code for the ICLR 2023 paper "GPTQ: Accurate Post-training Quantization of Generative Pretrained Transformers".
marlin
FP16xINT4 LLM inference kernel that can achieve near-ideal ~4x speedups up to medium batchsizes of 16-32 tokens.
OBC
Code for the NeurIPS 2022 paper "Optimal Brain Compression: A Framework for Accurate Post-Training Quantization and Pruning".
PanzaMail
qmoe
Code for the paper "QMoE: Practical Sub-1-Bit Compression of Trillion-Parameter Models".
Quartet
QUIK
Repository for the QUIK project, enabling the use of 4bit kernels for generative inference - EMNLP 2024
qutlass
QuTLASS: CUTLASS-Powered Quantized BLAS for Deep Learning
Sparse-Marlin
Boosting 4-bit inference kernels with 2:4 Sparsity
sparsegpt
Code for the ICML 2023 paper "SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot".
IST Austria Distributed Algorithms and Systems Lab's Repositories
IST Austria Distributed Algorithms and Systems Lab doesn’t have any repository yet.