cduk's Stars
stanfordnlp/dspy
DSPy: The framework for programming—not prompting—language models
facefusion/facefusion
Industry leading face manipulation platform
fishaudio/fish-speech
SOTA Open Source TTS
KwaiVGI/LivePortrait
Bring portraits to life!
voideditor/void
Ucas-HaoranWei/GOT-OCR2.0
Official code implementation of General OCR Theory: Towards OCR-2.0 via a Unified End-to-end Model
NVIDIA/cutlass
CUDA Templates for Linear Algebra Subroutines
leejet/stable-diffusion.cpp
Stable Diffusion and Flux in pure C/C++
prabirshrestha/vim-lsp
async language server protocol plugin for vim and neovim
mishushakov/llm-scraper
Turn any webpage into structured data using LLMs
codelion/optillm
Optimizing inference proxy for LLMs
pytorch/ao
PyTorch native quantization and sparsity for training and inference
pentacent/keila
Open Source Newsletter Tool.
mattn/vim-lsp-settings
Auto configurations for Language Server for vim-lsp
matatonic/openedai-speech
An OpenAI API compatible text to speech server using Coqui AI's xtts_v2 and/or piper tts as the backend.
BretFisher/docker-vackup
Script to easily backup and restore docker volumes
facefusion/facefusion-docker
Industry leading face manipulation platform
TheBlewish/Web-LLM-Assistant-Llamacpp-Ollama
A Python-based web-assisted large language model (LLM) search assistant using Llama.cpp
facefusion/facefusion-assets
Industry leading face manipulation platform
Agent-Tools/awesome-autonomous-web
CubicalBatch/deaddit
If Reddit's content was completely AI-generated.
ModelCloud/GPTQModel
Production ready LLM model compression/quantization toolkit with accelerated inference support for both cpu/gpu via HF, vLLM, and SGLang.
BorealisAI/flora-opt
This is the official repository for the paper "Flora: Low-Rank Adapters Are Secretly Gradient Compressors" in ICML 2024.
perk11/large-model-proxy
Large Model Proxy is designed to make it easy to run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources. It listens on a dedicated port for each proxied LM, making them always available to the clients connecting to these ports.
sasha0552/nvidia-pstated
A daemon that automatically manages the performance states of NVIDIA GPUs.
michaelfeil/embed
A stable, fast and easy-to-use inference library with a focus on a sync-to-async API
cduk/vllm-pascal
A fork of vLLM enabling Pascal architecture GPUs
diegovelilla/reddit-omni
AI Reddit bot that scrapes subreddits for questions, conducts research, and posts automated answers to help users with relevant information.
merefield/discourse-frotz
A plugin that uses Frotz to give you an interactive fiction experience on your Discourse forum
the-crypt-keeper/llama-srb-api
Single Request Batching API server backed by llama.cpp