twinnydotdev/twinny

Generation gets interrupted

Closed this issue · 6 comments

When generating it gets interrupted and stops.
CleanShot 2024-02-24 at 15 48 00@2x

Please check max token for chat setting and set to -1 for infinite. The model may be sending eos token or the request from the server was ended.

CleanShot 2024-02-24 at 16 19 14@2x
CleanShot 2024-02-24 at 16 19 33@2x
Sadly still ending abruptly. Tested with:
CleanShot 2024-02-24 at 16 20 15@2x

Hmm, not sure will have to check. Does it happen in terminal when calling Ollama with the same options? You can enable debug modem in settings to see the request and options.

Hey @MAKESEB I updated in 3.7.0 to the new version of the Ollama API which supports Open AI specification, it may help with this issue, please report back.

Updating twinny and setting all token limits to -1 solves the incompletion issue for me. Thanks

Thanks!