keldenl/gpt-llama.cpp

SERVER BUSY, REQUEST QUEUED

CyberRide opened this issue · 0 comments

===== CHAT COMPLETION REQUEST =====

===== LLAMA.CPP SPAWNED =====
/root/llama.cpp/main -m /root/llama.cpp/models/7B/ggml-model-q4_0.bin --temp 0.7 --n_predict 4000 --top_p 0.1 --top_k 40 -b 2000 -c 4096 --seed -1 --repeat_penalty 1.1764705882352942 --reverse-prompt user: --reverse-prompt
user --reverse-prompt system: --reverse-prompt
system --reverse-prompt ## --reverse-prompt

--reverse-prompt ### -i -p ### Instructions

Complete the following chat conversation between the user and the assistant. System messages should be strictly followed as additional instructions.

Inputs

system: You are a helpful assistant.
user: How are you?
assistant: Hi, how may I help you today?
system: You are ChatGPT, a helpful assistant developed by OpenAI.

Response

user: How are you doing today?
assistant:

===== REQUEST =====
user: How are you doing today?

===== RESPONSE =====

REQUEST RECEIVED
SERVER BUSY, REQUEST QUEUED