Issues
- 2
Repetition with Llama3-70b and EETQ
#22 opened by mjsteele12 - 2
Does it support Vision Transformers?
#21 opened by PaulaDelgado-Santos - 1
Unsupported Arch Assertion fail
#30 opened by rahul3161 - 0
My system just updated CUDA to 12.6 and I can no longer compile EETQ. (Python 3.12)
#31 opened by michael-newsrx - 1
- 3
EETQ wheel not building
#27 opened by donjuanpond - 2
EETQ-quantized TrOCR gives nonsense output
#28 opened by donjuanpond - 2
Does it support Whisper model?
#26 opened by kadirnar - 6
- 1
Supports H100
#12 opened by mwbyeon - 4
Support CPU quantization
#19 opened by xgal - 2
ImportError: cannot import name 'EetqConfig' from 'transformers', despite using using 4.38.2 which satisfies >=4.27.0
#23 opened by moruga123 - 7
How to handle bfloat16?
#4 opened by vgoklani - 2
- 1
- 3
Qlora with eetq is quite slow
#17 opened by hjh0119 - 4
how to dequant a EETQ model?
#14 opened by mxjmtxrm - 3
Quantization takes a very long time
#10 opened by timohear - 3
- 1
Question on outlier handling
#1 opened by 0xymoro - 2
Why does EETQ take up all VRAM
#3 opened by RonanKMcGovern - 5
安装出错ERROR: Could not build wheels for EETQ, which is required to install pyproject.toml-based projects
#2 opened by linshuijin