Pinned Repositories
ai-artifacts
Hackable open-source version of Anthropic's AI Artifacts chat
candle
Minimalist ML framework for Rust
fasttensors
mistral.rs
Blazingly fast LLM inference.
text-generation-webui
A Gradio web UI for Large Language Models.
ro99's Repositories
ro99/ai-artifacts
Hackable open-source version of Anthropic's AI Artifacts chat
ro99/candle
Minimalist ML framework for Rust
ro99/exllamav2
A fast inference library for running LLMs locally on modern consumer-class GPUs
ro99/fasttensors
ro99/gpu-programming
Meta-GPU lesson covering general aspects of GPU programming as well as specific frameworks
ro99/llm.c
LLM training in simple, raw C/CUDA
ro99/mistral.rs
Blazingly fast LLM inference.
ro99/nbodysim
Realtime 3D N-Body-Simulation
ro99/NovelWithLLMs
ro99/open-gpu-kernel-modules
NVIDIA Linux open GPU kernel module source
ro99/text-generation-webui
A Gradio web UI for Large Language Models.
ro99/tract
Tiny, no-nonsense, self-contained, Tensorflow and ONNX inference
ro99/ZLUDA
CUDA on AMD GPUs
ro99/gallama
ro99/llm_client
The Easiest Rust Interface for Local LLMs and an Interface for Deterministic Signals from Probabilistic LLM Vibes
ro99/obsidian-copilot
THE Copilot in Obsidian
ro99/ollama
Get up and running with Llama 3.1, Mistral, Gemma 2, and other large language models.
ro99/open-webui
User-friendly WebUI for LLMs (Formerly Ollama WebUI)
ro99/supermemory
Build your own second brain with supermemory. It's a ChatGPT for your bookmarks. Import tweets or save websites and content using the chrome extension.
ro99/text-generation-inference
Large Language Model Text Generation Inference