antferdom's Stars
codecrafters-io/build-your-own-x
Master programming by recreating your favorite technologies from scratch.
karpathy/llm.c
LLM training in simple, raw C/CUDA
sharkdp/hyperfine
A command-line benchmarking tool
jart/cosmopolitan
build-once run-anywhere c library
Mozilla-Ocho/llamafile
Distribute and run LLMs with a single file.
gleam-lang/gleam
⭐️ A friendly language for building type-safe, scalable systems!
nats-io/nats-server
High-Performance server for NATS.io, the cloud and edge native messaging system.
outlines-dev/outlines
Structured Text Generation
turboderp/exllamav2
A fast inference library for running LLMs locally on modern consumer-class GPUs
flame/blis
BLAS-like Library Instantiation Software Framework
microsoft/aici
AICI: Prompts as (Wasm) Programs
Lightning-AI/lightning-thunder
Make PyTorch models up to 40% faster! Thunder is a source to source compiler for PyTorch. It enables using different hardware executors at once; across one or thousands of GPUs.
deepseek-ai/DeepSeek-MoE
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models
Vahe1994/AQLM
Official Pytorch repository for Extreme Compression of Large Language Models via Additive Quantization https://arxiv.org/pdf/2401.06118.pdf
nats-io/nats.py
Python3 client for NATS
theroyallab/tabbyAPI
An OAI compatible exllamav2 API that's both lightweight and fast
microsoft/onnxscript
ONNX Script enables developers to naturally author ONNX functions and models using a subset of Python.
antirez/gguf-tools
GGUF implementation in C as a library and a tools CLI program
dylibso/hermit
Actually Portable WebAssembly compiler toolchain for self-contained cross-platform binaries
locuslab/ect
Consistency Models Made Easy
d5h-foss/grpc-interceptor
Simplified Python gRPC interceptors
rapidsai/wholegraph
WholeGraph - large scale Graph Neural Networks
pytorch-labs/applied-ai
Applied AI experiments and examples for PyTorch
bethgelab/frequency_determines_performance
Code for the paper: "No Zero-Shot Without Exponential Data: Pretraining Concept Frequency Determines Multimodal Model Performance"
ezyang/torchdbg
PyTorch centric eager mode debugger
yixiaoer/mistral-v0.2-jax
JAX implementation of the Mistral 7b v0.2 model
lernapparat/torchhacks
Hacks for PyTorch
jart/matmul
joeyballentine/ESRGAN-candle-rs
ESRGAN implemented in rust with candle
3kwa/goingnats
a Python NATS client