ztsweet's Stars
Lightning-AI/LitServe
Lightning-fast serving engine for any AI model of any size. Flexible. Easy. Enterprise-scale.
zed-industries/zed
Code at the speed of thought – Zed is a high-performance, multiplayer code editor from the creators of Atom and Tree-sitter.
mitchellh/libxev
libxev is a cross-platform, high-performance event loop that provides abstractions for non-blocking IO, timers, events, and more and works on Linux (io_uring or epoll), macOS (kqueue), and Wasm + WASI. Available as both a Zig and C API.
cdgriffith/Box
Python dictionaries with advanced dot notation access
linkedin/Liger-Kernel
Efficient Triton Kernels for LLM Training
ziglang/zig-spec
AdjectiveAllison/zvdb
Zig Vector Database!
fabioarnold/nanovg-zig
A small anti-aliased hardware-accelerated vector graphics library
gusye1234/nano-vectordb
A simple, easy-to-hack Vector Database
skyzh/mini-lsm
A tutorial of building an LSM-Tree storage engine in a week.
karthik-codex/Autogen_GraphRAG_Ollama
Microsoft's GraphRAG + AutoGen + Ollama + Chainlit = Fully Local & Free Multi-Agent RAG Superbot
QwenLM/Qwen2.5
Qwen2.5 is the large language model series developed by Qwen team, Alibaba Cloud.
gusye1234/nano-graphrag
A simple, easy-to-hack GraphRAG implementation
Hejsil/zig-clap
Command line argument parsing library
pytorch/torchchat
Run PyTorch LLMs locally on servers, desktop and mobile
THUDM/LongWriter
LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs
huggingface/datasets
🤗 The largest hub of ready-to-use datasets for ML models with fast, easy-to-use and efficient data manipulation tools
comfyanonymous/ComfyUI
The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.
aria2/aria2
aria2 is a lightweight multi-protocol & multi-source, cross platform download utility operated in command-line. It supports HTTP/HTTPS, FTP, SFTP, BitTorrent and Metalink.
quickwit-oss/quickwit
Cloud-native search engine for observability. An open-source alternative to Datadog, Elasticsearch, Loki, and Tempo.
Dao-AILab/flash-attention
Fast and memory-efficient exact attention
EricLBuehler/candle-vllm
Efficent platform for inference and serving local LLMs including an OpenAI compatible API server.
InfiniTensor/InfiniLM
LlamaEdge/LlamaEdge
The easiest & fastest way to run customized and fine-tuned LLMs locally or on the edge
bbycroft/llm-viz
3D Visualization of an GPT-style LLM
CompendiumLabs/ziggy
Embedding, quantization, and vector indexing. Designed for speed.
tensorlakeai/indexify
A realtime serving engine for Data-Intensive Generative AI Applications
PaddlePaddle/PaddleOCR
Awesome multilingual OCR toolkits based on PaddlePaddle (practical ultra lightweight OCR system, support 80+ languages recognition, provide data annotation and synthesis tools, support training and deployment among server, mobile, embedded and IoT devices)
BerriAI/litellm
Python SDK, Proxy Server (LLM Gateway) to call 100+ LLM APIs in OpenAI format - [Bedrock, Azure, OpenAI, VertexAI, Cohere, Anthropic, Sagemaker, HuggingFace, Replicate, Groq]
MistApproach/callm
Run Generative AI models directly on your hardware