Issues
- 1
Is it commercially usable?
#20 opened by AayushSameerShah - 0
A new code editing dataset
#19 opened by likaixin2000 - 1
Cannot solve complex problems
#11 opened by haseeb-heaven - 0
- 0
Llama2 model with code instruction-tuning on a single RTX 3090 is available now
#17 opened by juyongjiang - 0
Difference between new_codealpaca.json, rosetta_alpaca.json and codealpaca-20k.json?
#16 opened by Symbolk - 0
Just want to say thank you.
#15 opened by nnnpa31 - 1
AssertionError: Check batch related parameters. train_batch_size is not equal to micro_batch_per_gpu * gradient_acc_step * world_size 256 != 8 * 1 * 8
#14 opened by huangd1999 - 0
Open LLaMA project
#13 opened by DevNullx64 - 0
- 0
Instructions for training 13b model
#10 opened by RevanthRameshkumar - 2
More training data
#2 opened by renau - 1
65b model possible?
#7 opened by alexconstant9108 - 0
How big is the finished model?
#9 opened by pmb2 - 2
Hosting your dataset on the Hugging Face Hub
#5 opened by lewtun - 1
bug: get empty state dict
#8 opened by Anditty - 7
How long did it take?
#3 opened by fareesh - 1
Private data
#6 opened by VedAustin - 1
Please publish weights?
#1 opened by OlegJakushkin