Lightning-AI/lit-llama
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
PythonApache-2.0
Stargazers
- BordaPrague
- bailooNew Delhi, India
- akihironittaLondon
- tchatonLondon
- alecmerdlerNew York City
- satani99
- VikramxD
- fer-gitSingapore
- msaroufimThe Matrix
- josephwinston
- pavelklymenkoSan Francisco Bay Area
- robmarkcoleLondon, UK
- daigo0927Tokyo, Japan
- laifi
- goxccchina
- towzeurParis, France
- DDanlov
- asifrNew York
- florianbaudLyon, France
- din0sAmsterdam, Netherlands
- sarisel
- senad96Rome, Italy
- mfranzonVerona, Italy
- bhaddow
- edenlightning
- thiagogalesi
- msrivastavaLos Angeles
- ElieAntoine
- batman-doHaNoi
- JugglingNumbers
- gyuro
- DonLeif
- bilelomrani1Paris, France
- Yannlecun
- faizwhb
- SamerW