a questuon about the single GPU Inference
TitleZ99 opened this issue · 1 comments
TitleZ99 commented
Thanks for this great job and i'm wondering how to run inference in a 8GB single GPU,like your example showing in the readme. I tried it in my RTX2080ti with 11GB and the result is CUDA out of memory.
tanglaoya321 commented
Same problem, a single gpu, in the case of no quantization, it should be the need for 4*7B=28GB memory