chenbong's Stars
ggerganov/llama.cpp
LLM inference in C/C++
xai-org/grok-1
Grok open release
psf/black
The uncompromising Python code formatter
hiyouga/LLaMA-Factory
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
flameshot-org/flameshot
Powerful yet simple to use screenshot software :desktop_computer: :camera_flash:
mlc-ai/mlc-llm
Universal LLM Deployment Engine with ML Compilation
richards199999/Thinking-Claude
Let your Claude able to think
d2phap/ImageGlass
🏞 A lightweight, versatile image viewer
xiaoyaDev/xiaoya-alist
小雅Alist的相关周边
InternLM/lmdeploy
LMDeploy is a toolkit for compressing, deploying, and serving LLMs.
Calcium-Ion/new-api
AI模型接口管理与分发系统,支持将多种大模型转为OpenAI格式调用、支持Midjourney Proxy、Suno、Rerank,兼容易支付协议,可供个人或者企业内部管理与分发渠道使用,本项目基于One API二次开发。🍥 The next-generation LLM gateway and AI asset management system supports multiple languages.
jurplel/qView
Practical and minimal image viewer
HuangJunJie2017/BEVDet
Code base of the BEVDet series .
ZhangGe6/onnx-modifier
A tool to modify ONNX models in a visualization fashion, based on Netron and Flask.
NexaAI/Awesome-LLMs-on-device
Awesome LLMs on Device: A Comprehensive Survey
OpenGVLab/OmniQuant
[ICLR2024 spotlight] OmniQuant is a simple and powerful quantization technique for LLMs.
onnx/optimizer
Actively maintained ONNX Optimizer
mit-han-lab/qserve
QServe: W4A8KV4 Quantization and System Co-design for Efficient LLM Serving
ModelTC/llmc
[EMNLP 2024 Industry Track] This is the official PyTorch implementation of "LLMC: Benchmarking Large Language Model Quantization with a Versatile Compression Toolkit".
HFAiLab/hai-platform
一种任务级GPU算力分时调度的高性能深度学习训练平台
facebookresearch/LLM-QAT
Code repo for the paper "LLM-QAT Data-Free Quantization Aware Training for Large Language Models"
jy-yuan/KIVI
[ICML 2024] KIVI: A Tuning-Free Asymmetric 2bit Quantization for KV Cache
OpenGVLab/EfficientQAT
EfficientQAT: Efficient Quantization-Aware Training for Large Language Models
nbasyl/LLM-FP4
The official implementation of the EMNLP 2023 paper LLM-FP4
tsingmicro-toolchain/OnnxSlim
A Toolkit to Help Optimize Large Onnx Model
gmalivenko/onnx-opcounter
Count number of parameters / MACs / FLOPS for ONNX models.
BestAnHongjun/LMDeploy-Jetson
Deploying LLMs offline on the NVIDIA Jetson platform marks the dawn of a new era in embodied intelligence, where devices can function independently without continuous internet access.
TCLResearchEurope/torch-dag
pruning vision models in torch
AbelLin1214/FastMarriageBooker
用于自动预约民政局婚姻登记处的号,限广东省民政局
richardliu11/marry_robber
初始提交