Motsepe-Jr's Stars
joey00072/QuasiQ
quasiq is really simple quantum computer
Motsepe-Jr/bafoGPT
Series of Language Model for Zulu Langauge
Motsepe-Jr/mini_RLHF
Minimalist PPO with machine feedback
mikex86/LibreCuda
Motsepe-Jr/gpt2
changchencc/dreamerv2_pytorch
dreamerv2_pytorch
asg017/sqlite-vec
A vector search SQLite extension that runs anywhere!
aimerou/awesome-ai-papers
A curated list of the most impressive AI papers
ejmejm/CLAgent
Continual learning agent
danijar/dreamer
Dream to Control: Learning Behaviors by Latent Imagination
EleutherAI/lm-evaluation-harness
A framework for few-shot evaluation of language models.
karpathy/LLM101n
LLM101n: Let's build a Storyteller
unslothai/unsloth
Finetune Llama 3.1, Mistral, Phi & Gemma LLMs 2-5x faster with 80% less memory
allenai/OLMo
Modeling, training, eval, and inference code for OLMo
allenai/dolma
Data and tools for generating and inspecting OLMo pre-training data.
axolotl-ai-cloud/axolotl
Go ahead and axolotl questions
karpathy/llm.c
LLM training in simple, raw C/CUDA
Chinese-Tiny-LLM/Chinese-Tiny-LLM
nus-apr/auto-code-rover
A project structure aware autonomous software engineer aiming for autonomous program improvement. Resolved 30.67% tasks (pass@1) in SWE-bench lite and 38.40% tasks (pass@1) in SWE-bench verified with each task costs less than $0.7.
All-Hands-AI/OpenHands
🙌 OpenHands: Code Less, Make More
imoneoi/openchat
OpenChat: Advancing Open-source Language Models with Imperfect Data
valkey-io/valkey
A flexible distributed key-value datastore that supports both caching and beyond caching workloads.
joey00072/ohara
Collection of autoregressive model implementation
labmlai/annotated_deep_learning_paper_implementations
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠
gleam-lang/gleam
⭐️ A friendly language for building type-safe, scalable systems!
turboderp/exllama
A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.
Zjh-819/LLMDataHub
A quick guide (especially) for trending instruction finetuning datasets
jzhang38/TinyLlama
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
NVlabs/tiny-cuda-nn
Lightning fast C++/CUDA neural network framework
ray-project/ray
Ray is a unified framework for scaling AI and Python applications. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.