jzhang38/EasyContext

How to estimate the maximum context length this repo can support for larger models?

Opened this issue · 0 comments

I want to fine-tune Qwen2-72B using my data. Could you please estimate how long a context this repo can support on a setup with 16 or 32 A100/A800 (80G) GPUs? Thanks!