How to estimate the maximum context length this repo can support for larger models?
Opened this issue · 0 comments
JingyangDeng commented
I want to fine-tune Qwen2-72B using my data. Could you please estimate how long a context this repo can support on a setup with 16 or 32 A100/A800 (80G) GPUs? Thanks!