Kav-K opened this issue 7 months ago · 0 comments
gpt-ask with older models seem to send a request where the max tokens are set to 4096, this is wrong, max tokens should either be omitted for these older models or just set to MODEL_MAX_TOKENS - CURRENT_TOKENS