noprompt's Stars
labmlai/annotated_deep_learning_paper_implementations
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠
meta-llama/llama
Inference code for Llama models
lllyasviel/Fooocus
Focus on prompting and generating
tatsu-lab/stanford_alpaca
Code and documentation to train Stanford's Alpaca models, and generate the data.
ajeetdsouza/zoxide
A smarter cd command. Supports all major shells.
facebookresearch/audiocraft
Audiocraft is a library for audio processing and generation with deep learning. It features the state-of-the-art EnCodec audio compressor / tokenizer, along with MusicGen, a simple and controllable music generation LM with textual and melodic conditioning.
RustPython/RustPython
A Python Interpreter written in Rust
UKPLab/sentence-transformers
State-of-the-Art Text Embeddings
Lightning-AI/litgpt
20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.
huggingface/text-generation-inference
Large Language Model Text Generation Inference
karpathy/minbpe
Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.
openlm-research/open_llama
OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset
Lightning-AI/lit-llama
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
AutoGPTQ/AutoGPTQ
An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.
samuelmtimbo/unit
Next Generation Visual Programming System
huggingface/distil-whisper
Distilled variant of Whisper for speech recognition. 6x faster, 50% smaller, within 1% word error rate.
xinyu1205/recognize-anything
Open-source and strong foundation image recognition models.
zigzap/zap
blazingly fast backends in zig
turboderp/exllama
A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.
magefree/mage
XMage - Magic Another Game Engine
PygmalionAI/aphrodite-engine
Large-scale LLM inference engine
keredson/wordninja
Probabilistically split concatenated words using NLP based on English Wikipedia unigram frequencies.
Oneirocom/Magick
Magick is a cutting-edge toolkit for a new kind of AI builder. Make Magick with us!
lucidrains/MEGABYTE-pytorch
Implementation of MEGABYTE, Predicting Million-byte Sequences with Multiscale Transformers, in Pytorch
AlexanderDzhoganov/ComfyTextures
Unreal Engine ⚔️ ComfyUI - Automatic texturing using generative diffusion models
AnacletoLAB/grape
🍇 GRAPE is a Rust/Python Graph Representation Learning library for Predictions and Evaluations
neurallambda/neurallambda
Reasoning Computers. Lambda Calculus, Fully Differentiable. Also Neural Stacks, Queues, Arrays, Lists, Trees, and Latches.
gautierdag/bpeasy
Fast bare-bones BPE for modern tokenizer training
ryrobes/flowmaps
Construct & orchestrate stand-alone Clojure core.async pipelines with ease or use it to encapsulate & manage complex 'flows' within your application
facebookresearch/r-mae
PyTorch implementation of R-MAE https//arxiv.org/abs/2306.05411