localllm

There are 111 repositories under localllm topic.

  • page-assist

    n4ze3m/page-assist

    Use your locally running AI models to assist you in your web browsing

    Language:TypeScript7.1k52467637
  • mostlygeek/llama-swap

    Model swapping for llama.cpp (or any local OpenAI API compatible server)

    Language:Go1.5k1115996
  • sauravpanda/BrowserAI

    Run local LLMs like llama, deepseek-distill, kokoro and more inside your browser

    Language:TypeScript1.2k7102106
  • KwaiKEG/KwaiAgents

    A generalized information-seeking agent system with Large Language Models (LLMs).

    Language:Python1.2k2243116
  • SqueezeAILab/SqueezeLLM

    [ICML 2024] SqueezeLLM: Dense-and-Sparse Quantization

    Language:Python703172849
  • SqueezeAILab/KVQuant

    [NeurIPS 2024] KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization

    Language:Python381111837
  • PromptEngineer48/MemGPT-AutoGEN-LLM

    Run MemGPT-AutoGEN-Local LLM Together

    Language:Python31261587
  • lofcz/LlmTornado

    The .NET library to build AI systems with 100+ LLM APIs: Anthropic, Azure, Cohere, DeepInfra, DeepSeek, Google, Groq, Mistral, Ollama, OpenAI, OpenRouter, Perplexity, vLLM, Voyage, xAI, and many more!

    Language:C#27194838
  • ollama-ex

    lebrunel/ollama-ex

    A nifty little library for working with Ollama in Elixir.

    Language:Elixir128349
  • BodhiSearch/BodhiApp

    Run Open Source/Open Weight LLMs locally with OpenAI compatible APIs

    Language:Rust1165129
  • MDGrey33/pyvisionai

    The PyVisionAI Official Repo

    Language:Python10552712
  • tegridydev/dnd-llm-game

    MVP of an idea using multiple local LLM models to simulate and play D&D

    Language:Python843511
  • perk11/large-model-proxy

    Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on different ports and loading/unloading them on demand

    Language:Go822163
  • open_local_ui

    WilliamKarolDiCioccio/open_local_ui

    OpenLocalUI: Native desktop app for Windows, MacOS and Linux. Easily run Large Language Models locally, no complex setups required. Inspired by OpenWebUI's simplicity for LLM use.

    Language:Dart666393
  • wsmlby/homl

    The easiest & fastest way to run LLMs in your home lab

    Language:Python66
  • 3-ark/Cognito-AI_Sidekick

    Cognito: Supercharge your Chrome browser with AI. Guide, query, and control everything using natural language.

    Language:TypeScript55
  • Ganador1/FenixAI_tradingBot

    Fenix Ai Trading Bot with crew ai and ollama

    Language:Python54
  • docspedia/docspedia

    Chat with your pdf using your local LLM, OLLAMA client.(incomplete)

    Language:TypeScript39201
  • twinnydotdev/symmetry-cli

    The client for the Symmetry peer-to-peer inference network. Enabling users to connect with each other, share computational resources, and collect valuable machine learning data.

    Language:JavaScript28004
  • Darthph0enix7/DocPOI_repo

    A local chatbot for managing docs

    Language:Python24200
  • dokasto/Saidia

    Offline-first, desktop AI assistant tailored for educators, enabling them to generate questions directly from source materials.

    Language:TypeScript22
  • LLMProxy

    obirler/LLMProxy

    LLMProxy is an intelligent large language model backend routing proxy service.

    Language:C#191
  • Talnz007/VulkanIlm

    GPU-accelerated LLaMA inference wrapper for legacy Vulkan-capable systems a Pythonic way to run AI with knowledge (Ilm) on fire (Vulkan).

    Language:Python190
  • neodyland/entropix

    Unofficial entropix impl for Gemma2 and Llama and Qwen2 and Mistral

    Language:Python171
  • seyf1elislam/LocalLLM_OneClick_Colab

    Run gguf LLM models in Latest Version TextGen-webui and koboldcpp

    Language:Jupyter Notebook16100
  • imrightguy/CloudToLocalLLM

    Secure Flutter desktop app connecting Auth0 authentication with local Ollama AI models via encrypted tunneling. Access your private AI instances remotely while keeping data on your hardware.

    Language:Dart15
  • scouzi1966/vesta-mac-dist

    Vesta macOS Distribution - Official releases and downloads.Vesta AI Chat Assistant for macOS - Built with SwiftUI and Apple Intelligence using Apple's on device model on MacOs Tahoe (MacOS 26)

  • sujithhubpost/initialterm

    Local LLM enabled Human terminal interaction made easy.

    Language:Python15304
  • Wakoma/OfflineAI

    Local/Offline Machine Learning Resources

  • 10Nates/Humanlike-AI-Chat

    Humanlike AI Chat is a terminal-based LLM UI designed to study how to bypass AI text detection.

    Language:Python13101
  • akutishevsky/salesforce-local-llm-assistant

    LWC that runs LLMs locally on your computer and uses Salesforce records for context.

    Language:Apex135
  • arjunprabhulal/function-calling-gemma3

    Demo project showcasing Gemma3 function calling capabilities using Ollama. Enables automatic web searches via Serper.dev for up-to-date information and features an interactive Gradio chat interface.

    Language:Python12108
  • GridLLM/GridLLM

    Language:TypeScript9
  • mskry/dotfiles

    Alacritty + Fish + Zellij + Starship + Neovim + i3 + Supermaven + Ollama 🦙 = 🚀

    Language:Shell9201
  • arvindjuneja/OwnAI

    Local LLM (using Ollama) interface for MacOS

    Language:Swift8
  • statikfintechllc/AscendNet

    A P2P network where users: Share, trade, and sell AI prompts and prompt chains, Pool and rent distributed compute for inference/training, All transactions happen P2P, with a built-in 3% fee

    Language:TypeScript8