anttttti's Stars
All-Hands-AI/OpenHands
🙌 OpenHands: Code Less, Make More
ztjhz/BetterChatGPT
An amazing UI for OpenAI's ChatGPT (Website + Windows + MacOS + Linux)
peak/s5cmd
Parallel S3 and local filesystem execution tool.
neuralmagic/sparseml
Libraries for applying sparsification recipes to neural networks with a few lines of code, enabling faster and smaller models
refuel-ai/autolabel
Label, clean and enrich text datasets with LLMs.
princeton-nlp/SWE-bench
[ICLR 2024] SWE-Bench: Can Language Models Resolve Real-world Github Issues?
unslothai/hyperlearn
2-2000x faster ML algos, 50% less memory usage, works on all hardware - new and old.
anthropics/anthropic-sdk-python
pemistahl/lingua-py
The most accurate natural language detection library for Python, suitable for short text and mixed-language text
pemistahl/lingua-rs
The most accurate natural language detection library for Rust, suitable for short text and mixed-language text
prometheus-eval/prometheus-eval
Evaluate your LLM's response with Prometheus and GPT4 💯
IST-DASLab/marlin
FP16xINT4 LLM inference kernel that can achieve near-ideal ~4x speedups up to medium batchsizes of 16-32 tokens.
efeslab/Nanoflow
A throughput-oriented high-performance serving framework for LLMs
vllm-project/llm-compressor
Transformers-compatible library for applying various compression algorithms to LLMs for optimized deployment with vLLM
nlpxucan/evol-instruct
neuralmagic/nm-vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
unifyai/unify
LLMs Run Riot in Production. Get Back in The Driving Seat. Build Your Own Evals, Iterate Quickly, and Go from Prototype to Production in No Time ⚡
perpetual-ml/perpetual
A self-generalizing gradient boosting machine which doesn't need hyperparameter optimization
IBM/data-prep-kit
Open source project for data preparation of LLM application builders
neuralmagic/AutoFP8
premAI-io/benchmarks
🕹️ Performance Comparison of MLOps Engines, Frameworks, and Languages on Mainstream AI Models.
ModelCloud/GPTQModel
An easy-to-use LLM quantization and inference toolkit based on GPTQ algorithm (weight-only quantization).
cognitivecomputations/spectrum
HandH1998/QQQ
QQQ is an innovative and hardware-optimized W4A8 quantization solution.
IST-DASLab/Sparse-Marlin
Boosting 4-bit inference kernels with 2:4 Sparsity
anthropics/anthropic-bedrock-python
Qurrent-AI/RES-Q
RES-Q: Evaluating the Code-Editing Capability of Large Language Model Systems at the Repository Scale
woheller69/LLAMA_TK_CHAT
Simple chat interface for local AI using llama-cpp-python and llama-cpp-agent
TeHikuMedia/nga-tautohetohe-reo
Corpus of Te Reo derived from the New Zealand Hansard
sasha0552/pascal-pkgs-ci
The main repository for building Pascal-compatible versions of ML applications and libraries.