qwopqwop200/GPTQ-for-LLaMa

Porting GPTQ to CPU?

yiliu30 opened this issue · 2 comments

Is it possible to run GPTQ on a machine that has only CPUs? If not, is there a plan for it?

You can use a GPTQ quantized model with llama.cpp by using this conversion script I believe.

Hiwyl commented

just quant model on CPU?