vgel's Stars
karpathy/LLM101n
LLM101n: Let's build a Storyteller
Mozilla-Ocho/llamafile
Distribute and run LLMs with a single file.
stanfordnlp/dspy
DSPy: The framework for programming—not prompting—language models
Lightning-AI/litgpt
20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.
srush/GPU-Puzzles
Solve puzzles. Learn CUDA.
InternLM/InternLM
Official release of InternLM2.5 base and chat models. 1M context support
asg017/sqlite-vec
A vector search SQLite extension that runs anywhere!
openai/transformer-debugger
eureka-research/Eureka
Official Repository for "Eureka: Human-Level Reward Design via Coding Large Language Models" (ICLR 2024)
TransformerLensOrg/TransformerLens
A library for mechanistic interpretability of GPT-style language models
astoilkov/main-thread-scheduling
Fast and consistently responsive apps using a single function call
sdan/vlite
fast vector database made in numpy
minosvasilias/godot-dodo
Finetuning large language models for GDScript generation.
google-deepmind/tracr
cgsdev0/bash-stack
modern web framework in bash
jbloomAus/SAELens
Training Sparse Autoencoders on Language Models
bsilverthorn/maccarone
AI-managed code blocks in Python ⏪⏩
EleutherAI/sae
Sparse autoencoders
Mihaiii/llm_steer
Steer LLM outputs towards a certain topic/subject and enhance response capabilities using activation engineering by adding steering vectors
mxbi/arckit
Tools for working with the Abstraction & Reasoning Corpus
xjdr-alt/simple_transformer
Simple Transformer in Jax
arXiv/arxiv-browse
Flask app for article abstract and listing pages
neelnanda-io/1L-Sparse-Autoencoder
abacaj/train-with-fsdp
lumpenspace/raft
RAFT, or Retrieval-Augmented Fine-Tuning, is a method comprising of a fine-tuning and a RAG-based retrieval phase. It is particularly suited for the creation of agents that realistically emulate a specific human target.
taylorai/onnx_embedding_models
utilities for loading and running text embeddings with onnx
shehper/sparse-dictionary-learning
An Open Source Implementation of Anthropic's Paper: "Towards Monosemanticity: Decomposing Language Models with Dictionary Learning"
teltam/twinkgrad
wjessup/handmade-transformer
making GPT2 transformer weights by hand
jhnwnstd/corpus_toolkit
Python toolkit for corpus analysis: tokenization, lexical diversity, vocabulary growth prediction, entropy measures, and Zipf/Heaps visualizations.