ggml
There are 88 repositories under ggml topic.
ggerganov/llama.cpp
LLM inference in C/C++
rustformers/llm
[Unmaintained, see README] An ecosystem of Rust libraries for working with large language models
xorbitsai/inference
Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.
LostRuins/koboldcpp
Run GGUF models easily with a KoboldAI UI. One File. Zero Install.
leejet/stable-diffusion.cpp
Stable Diffusion and Flux in pure C/C++
RWKV/rwkv.cpp
INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model
guinmoon/LLMFarm
llama and other large language models on iOS and MacOS offline using GGML library.
RahulSChand/gpu_poor
Calculate token/s & GPU memory requirement for any LLM. Supports llama.cpp/ggml/bnb/QLoRA quantization
PABannier/bark.cpp
Suno AI's Bark model in C/C++ for fast text-to-speech generation
abacaj/mpt-30B-inference
Run inference on MPT-30B using CPU
Maknee/minigpt4.cpp
Port of MiniGPT4 in C++ (4bit, 5bit, 6bit, 8bit, 16bit CPU inference with GGML)
azkadev/whisper
Whisper Dart is a cross platform library for dart and flutter that allows converting audio to text / speech to text / inference from Open AI models
the-crypt-keeper/can-ai-code
Self-evaluating interview for AI coders
monatis/clip.cpp
CLIP inference in plain C/C++ with no extra dependencies
shm007g/LLaMA-Cult-and-More
Large Language Models for All, 🦙 Cult and More, Stay in touch !
azkadev/bark
WIP Library Text To Speech From Suno AI's Bark in C/C++ for fast inference
azkadev/general_ai
GENERAL Ai Library For DART & Flutter
staghado/vit.cpp
Inference Vision Transformer (ViT) in plain C/C++ with ggml
mayooear/private-chatbot-mpt30b-langchain
Chat with your data privately using MPT-30b
balisujohn/tortoise.cpp
A ggml (C++) re-implementation of tortoise-tts
mgonzs13/llama_ros
llama.cpp (GGUF LLMs) and llava.cpp (GGUF VLMs) for ROS 2
abacaj/replit-3B-inference
Run inference on replit-3B code instruct model using CPU
chenhunghan/ialacol
🪶 Lightweight OpenAI drop-in replacement for Kubernetes
gotzmann/booster
Booster - open accelerator for LLM models. Better inference and debugging for AI hackers
zatevakhin/obsidian-local-llm
Obsidian Local LLM is a plugin for Obsidian that provides access to a powerful neural network, allowing users to generate text in a wide range of styles and formats using a local LLM.
zhouwg/kantv
workbench for learing&practising AI tech in real scenario on Android device, powered by GGML(Georgi Gerganov Machine Learning) and NCNN(Tencent NCNN) and FFmpeg
guoriyue/LangCommand
LangCommand is a local inference command-line tool that transforms natural language descriptions into shell commands.
sevagh/demucs.cpp
C++17 port of Demucs v3 (hybrid) and v4 (hybrid transformer) models with ggml and Eigen3
nrl-ai/CustomChar
Your customized AI assistant - Personal assistants on any hardware! With llama.cpp, whisper.cpp, ggml, LLaMA-v2.
rbourgeat/ImpAI
😈 ImpAI is an advanced role play app using large language and diffusion models.
mgonzs13/whisper_ros
Speech-to-Text based on SileroVAD + whisper.cpp (GGML Whisper) for ROS 2
Mobile-Artificial-Intelligence/maid_llm
maid_llm is a dart implementation of llama.cpp used by the mobile artificial intelligence distribution (maid)
ahoylabs/gguf.js
A Javascript library (with Typescript types) to parse metadata of GGML based GGUF files.
seasonjs/stable-diffusion
pure go for stable-diffusion and support cross-platform.
cztomsik/ggml-js
JavaScript bindings for the ggml-js library
latestissue/AltaeraAI
A set of bash scripts to automate deployment of GGML/GGUF models [default: RWKV] with the use of KoboldCpp on Android - Termux