Issues
- 0
EOS
#544 opened by wendadawen - 4
- 0
runq.c quantization not symmetric
#542 opened by hafezmg48 - 0
how to use train.py to train from llama2-HF model.
#541 opened by bailuan - 0
llama2.c calculation in FP16
#539 opened by ztachip - 0
Convert GGUF file format to llam2.c 's bin file format
#538 opened by ztachip - 1
Adding support to Llama 3.1
#537 opened by asmit203 - 0
Everyone, I have implemented multi-token prediction of InfiniAttention and meta.
#517 opened by win10ogod - 1
[Suggestion] Enable Discussion
#492 opened by trholding - 3
Everyone, I have implemented multi-token prediction of InfiniAttention and meta.
#518 opened by win10ogod - 2
Runing llama2.c on a microcontroller
#480 opened by mahdi259 - 1
Error with torch not compiled with cuda enabled
#531 opened by berlinbrown - 0
Fail for execute ./run with meta llama2_7b.bin
#529 opened by sss28072637 - 2
- 0
Weight share of input and output embedding
#526 opened by fangzhangmnm - 1
Could llama2.c be adapted to BitNet?
#500 opened by izaxon - 3
- 2
Training Tiny Stories: 'CUDA' -vs- 'MPS'
#461 opened by dbl001 - 4
Simplified llama2.c.dll
#507 opened by JohnClaw - 2
Not an issue: Asking for help
#520 opened by Hjertesvikt - 1
- 0
mmap failed! ./run llama2_7b_q80.bin
#519 opened by codetown - 2
the export model and read_checkpoint is conflict
#501 opened by l1351868270 - 1
Tokenizer errors out when inferencing llama2
#502 opened by navidsam - 1
malloc failed! on stories260 model
#510 opened by vikramdattu - 1
How about Llama3?
#506 opened by franktaTian - 0
- 1
RuntimeError with CUDA assertion failure when resuming model training from checkpoint
#499 opened by fancling - 0
add feature: export (quantize) from Llama2.c format
#496 opened by hafezmg48 - 5
- 1
- 0
I'm doing an experiment with image generation, but my script outputs a binary file, how can I train a model using llama2.c?
#488 opened by win10ogod - 0
Could anyone port deepseek-moe to llama2.c?
#485 opened by win10ogod - 0
Please implement a project
#484 opened by win10ogod - 0
New Visual Walkthrough of Llama2.c
#483 opened by ZoroDerVonCodier - 0
Mobile React native Support Ported
#482 opened by hootan09 - 0
Understanding "multiple_of"
#481 opened by akbayt - 0
Train/val split
#478 opened by DavidHerel - 3
- 1
How to quantize stories15M.bin
#476 opened by forcekeng - 5
can i train on cpu
#456 opened by pure-water - 4
HOw to add different coropus ?
#462 opened by pure-water - 0
Keras based tiny llama implementations
#472 opened by Alexey234432 - 3
casual attention implementation
#469 opened by liecn - 1
- 0
Llama-shepherd-cli a small tool to keep track of implementations in various languages
#471 opened by mikepapadim - 0
NanoGPT in c for inference
#464 opened by liecn - 2
export does not seem to work?
#454 opened by pure-water - 0
mfu calculation
#460 opened by pure-water - 1
Is it possible to use Orca2 with this code ?
#452 opened by FiveTechSoft