Issues
- 1
A question about the prefix of only SFT
#211 opened by RikkiXu - 1
use instructor in openchat
#213 opened by ramandada - 1
huggingface down: openchat can't be started
#214 opened by antioxidanz - 1
About performance of LLama (llama2, llama3) model
#215 opened by huazhenliu - 2
For đź’»Online Demo It has been broken out.
#207 opened by Schimasuperbra - 1
opchatdataset.estimate_num_batches returns 0 at the beginning of training, and training-stuck problem
#212 opened by syboomsy - 2
- 1
Cannot create data file
#175 opened by phuvinhnguyen - 1
Does the current ochat.training_deepspeed.train only train v.3.2? How to train v.3.5?
#173 opened by houghtonweihu - 3
- 2
Error running new model openchat-3.5-0106-gemma on RTX 4090 24GB machine, works with older mistral based model
#198 opened by vikrantrathore - 2
- 1
openchat.team is down for 3 days
#201 opened by Borschik69 - 0
about data
#206 opened by Luoqiu76 - 0
weighted_token_accuracy
#205 opened by bino282 - 0
- 0
Error when using openchat/openchat-3.5-0106-gemma in text-generation-inference
#202 opened by houghtonweihu - 1
llama_model_load: error loading model: create_tensor: tensor 'output.weight' not found
#199 opened by wac81 - 3
Exception in ASGI application
#176 opened by Favern - 4
Can not Reproduce benchmarks
#186 opened by zhang7346 - 0
- 0
in llama.cpp load gguf has some issue in it
#196 opened by wac81 - 0
- 0
- 1
- 0
- 1
- 1
Using anything > 2048 for batch_max_length during training results in cuda index errors
#190 opened by corey-lambda - 0
Why the trained model does not produce the answer provided in the training data?
#188 opened by houghtonweihu - 3
- 2
OpenChat API Non-Deterministic Behavior
#183 opened by SkanderHellal - 1
- 1
Question for https://huggingface.co/openchat/openchat-3.5-0106/blob/main/openchat.json
#181 opened by houghtonweihu - 1
Question about `--per-sequence-loss`
#182 opened by Sanster - 4
- 0
There is a 7B model, good at math: https://github.com/deepseek-ai/DeepSeek-Math.
#180 opened by houghtonweihu - 0
What is the meaning of padding-free in ReadMe?
#179 opened by houghtonweihu - 0
run multiple instances on a single gpu
#178 opened by amir264710 - 0
I tried to pre-tokeinzation but failed
#177 opened by userdsr - 0
How to finetune openchat for downstream apps?
#174 opened by houghtonweihu - 0
Extending your training section in ReadMe?
#163 opened by houghtonweihu - 0
In the Web UI: https://openchat.team/, how can we specify "condition": "Math Correct"?
#165 opened by houghtonweihu - 0
- 1
- 0
distill openchat
#172 opened by sujitvasanth - 2
the website requires an openAI API key to work
#169 opened by HMZElidrissi - 1
The AI on the site openchat.team is down
#168 opened by Gerrytheskull - 1
the model of Pre-tokenized dataset openchat_v3.2_super.train.parquet is Llama2 or Mistral?
#170 opened by alphanlp - 4
did ```calculate_auto_lr``` affect a lot?
#162 opened by fpcsong - 0