Issues
- 1
`.bin` vs `.pt` size discrepency
#342 opened - 5
New export code OOM with 7B model
#341 opened - 7
Possible issue in decode()
#340 opened - 2
Reasoning behind version1_export logic
#339 opened - 3
-n 0 makes no tokens (21/08/2023 pulls)
#334 opened - 5
- 3
File names discussion
#323 opened - 16
convert ckpt.pt to huggingface model
#321 opened - 0
- 3
Error in converting huggingface models
#314 opened - 22
- 4
Yet another RoPE rotation block
#302 opened - 1
why not use llama from transformers
#299 opened - 2
Train wont start for custom dataset
#296 opened - 1
- 2
- 1
k, v sizes change after multiquery
#291 opened - 6
unable to convert llama2 7b model
#288 opened - 1
The tiktoken dependency seems unnecessary
#287 opened - 2
TypeError("get_tokenizer_model_path() missing 1 required positional argument: 'vocab_size'")
#280 opened - 3
CuBLAS status not INITIALIZED ERROR
#279 opened - 23
Quantization Brainstorming
#277 opened - 4
i wanna run python train.py but get errors
#273 opened - 1
- 4
- 2
What is it `assert 0 <= 1 < ndim`?
#259 opened - 0
Print model info
#255 opened - 7
a new chatbot project derived from llama2.c
#252 opened - 5
Big speed regression with top-p sampling
#246 opened - 4
Seed 0 gives <unk>
#239 opened - 1
- 9
Segmentation fault with new models
#237 opened - 2
feature-request: tiny dialog dataset
#236 opened - 2
I just want to say thank you
#235 opened - 3
- 2
llama 2 result change frequently
#229 opened - 4
- 6
Training on a single GPU
#221 opened - 2
- 3
- 3
Support inference with LLAMA 2 70B model
#210 opened - 3
prompt doesn't handle all characters
#209 opened - 24
Question - Weird prompt results
#204 opened - 2
- 0
- 1
- 2
- 3
- 1
- 2
Use cblas for matrix multiplication
#182 opened