Issues
- 4
How can I use the finetuned model with text-generation-webui or KoboldAI?
#16 opened by IIIIIIIllllllllIIIII - 3
Slow generation speed: around 10 minutes / loading forever on rtx3090 with 64gb ram....
#17 opened by IIIIIIIllllllllIIIII - 2
- 0
How to work with vLLM such as LLAVA
#63 opened by dinhquangsonimip - 50
How should I prepare the dataset for generative question answering on the private documents?
#38 opened by AayushSameerShah - 3
Question: Native windows support
#23 opened by Paillat-dev - 0
will this work with quantized GGUF files?
#62 opened by sengiv - 0
[Request] Retrain adapter from checkpoint?
#61 opened by zaanind - 3
Performance after FineTuning
#47 opened by Datta0 - 0
About llama-2-70B fine-tuning
#60 opened by RickMeow - 0
- 2
How to use CPU instead of GPU
#33 opened by Shreyas-ITB - 0
AMD GPU compability or CPU
#58 opened by LeLaboDuGame - 3
- 0
M1/M2 Metal support?
#57 opened by itsPreto - 0
[Request] Mac ARM support
#56 opened by voidcenter - 3
RuntimeError: unscale_() has already been called on this optimizer since the last update().
#54 opened by junxu-ai - 0
[Request] QLoRA support
#55 opened by CoolOppo - 0
- 5
`LLaMATokenizer` vs `LlamaTokenizer` class names
#22 opened by vadi2 - 2
Question: Is fine tuning suitable for factual answers from custom data, or is it better to use vector databases and use only the relevant chunk in the prompt for factual answers?
#4 opened by petrbrzek - 0
Multi GPU running
#51 opened by Shashika007 - 7
Issue in train in colab
#42 opened by fermions75 - 2
Getting OOM
#46 opened by alior101 - 0
- 3
Suggestion to improve UX
#32 opened by ch3rn0v - 3
"The tokenizer class you load from this checkpoint is 'LLaMATokenizer'."
#40 opened by IIIIIIIllllllllIIIII - 1
- 2
- 3
- 5
"error" in training - AttributeError: 'CastOutputToFloat' object has no attribute 'weight', RuntimeError: Only Tensors of floating point and complex dtype can require gradients
#29 opened by GreenTeaBD - 3
- 1
how to finetune with 'system information'
#30 opened by mhyeonsoo - 2
Attempting to use 13B in the simple tuner -
#28 opened by Atlas3DSS - 2
Not a problem - but like people should know
#26 opened by Atlas3DSS - 2
Error during Training RuntimeError: mat1 and mat2 shapes cannot be multiplied (511x2 and 3x4096)
#25 opened by kasakh - 8
Traceback during inference.
#6 opened by Hello1024 - 4
Examples to get started with
#11 opened by vadi2 - 1
How the finetuning output looks like?
#27 opened by mhyeonsoo - 4
- 2
Inference doesn't work after training
#10 opened by vadi2 - 2
Finetuning in unsupported language
#15 opened by jumasheff - 4
Host on Hugging Face Spaces
#19 opened by osanseviero - 1
Inference output text keeps running on...
#1 opened by lxe - 12
Inference works just once
#12 opened by vadi2 - 6
(WSL2) - No GPU / Cuda detected....
#13 opened by IIIIIIIllllllllIIIII - 1
Collecting info on memory requirements
#2 opened by jmiskovic - 1
Is CUDA 12.0 supported?
#8 opened by vadi2 - 2