Issues
- 0
numpy llama2 for fun and learning
#450 opened - 0
- 1
- 0
How to train a chat model
#447 opened - 2
Prefill Processing
#445 opened - 4
- 0
Llama transformer walkthrough
#442 opened - 0
- 0
How to run interface on GPU
#437 opened - 1
- 10
runomp on Mac M1 Max is slower than runfast
#432 opened - 0
- 0
tok512.model adding a token at start.
#430 opened - 0
About runq.c
#428 opened - 5
- 2
- 7
Is this project still active?
#425 opened - 3
Question: Sliding window attention
#424 opened - 0
Questions about the matmul function in run.c
#423 opened - 0
I found that the dim parameter affects the learning loss and n_layers affects the training speed.
#422 opened - 1
How to convert to huggingface model format?
#421 opened - 0
The trained model will not be saved.
#419 opened - 2
Q: How to finetune?
#418 opened - 0
- 0
- 0
Is it possible to increase or decrease the size of only some of the layers of the model structure?
#406 opened - 1
How to save checkpoints at each step?
#405 opened - 0
-
#404 opened - 4
- 1
How does this part of the Train code work?
#401 opened - 2
Mojo version?
#396 opened - 3
What is a good pretrain dataset for llama2c?
#393 opened - 3
Evolution of tinystories. Open sourced.
#392 opened - 0
[Feature Request] Support InternLM Deploy
#390 opened - 0
- 0
Pure JavaScript port of llama2.c
#384 opened - 1
llama2_7b_chat have no any response
#382 opened - 3
- 4
Optimized code for matmul() works 3.5 faster (for Mac M1 Max with ARM NEON) ... and even more...
#377 opened - 4
Interpretability of models
#375 opened - 8
- 0
HF candle
#371 opened - 0
- 1
- 0
- 5
Chat functionality requires big 7B model
#357 opened - 2
Code Llama rope_theta parameter
#356 opened - 1
260K Model Parameter count not right?
#354 opened - 1
- 2