Get up and running with Llama 3, Mistral, Gemma 2, and other large language models.
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
sfarzi doesn’t have any repository yet.