Pinned Repositories
awesome-model-quantization
A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (papers, repositories) that are missed by the repo.
Body-sense-simulation-against-game-system-based-on-Kinect
基于Kinect体感的仿真对抗游戏系统
cnn-quantization
Quantization of Convolutional Neural networks.
Design-and-Implementation-of-Face-Recognition-based-on-PYNQ
Face recognition, computer vision, deep learning, PYNQ, Movidius NCS
Electronics-Design-Competition
Remote amplitude frequency characteristic test device—Electronics Design Competition(Instrument class)
Lightweight-facial-expression-recognition
Face detection, expression recognition, model compression, terminal deployment
micronet
micronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、Low-Bit(≤2b)/Ternary and Binary(TWN/BNN/XNOR-Net); post-training-quantization(PTQ), 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization fuse for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape
Object-Tracking
Object Tracking
Object-tracking-system-based-on-deep-learning
Object tracking system based on deep learning. Master-Jetson TX2, ROS-image + control, image (DL) - detection + tracking, control - TurtleBot
Platform-Robot-Competition
Chinese Robot Competition—Platform Robot Competition
666DZY666's Repositories
666DZY666/micronet
micronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、Low-Bit(≤2b)/Ternary and Binary(TWN/BNN/XNOR-Net); post-training-quantization(PTQ), 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization fuse for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape
666DZY666/Awesome-LLM-Inference
📖A curated list of Awesome LLM Inference Paper with codes, TensorRT-LLM, vLLM, streaming-llm, AWQ, SmoothQuant, WINT8/4, Continuous Batching, FlashAttention, PagedAttention etc.
666DZY666/mmrazor
666DZY666/Torch-Pruning
[CVPR-2023] Towards Any Structural Pruning; LLMs / Diffusion / YOLOv8 / CNNs / Transformers
666DZY666/torchdynamo
A Python-level JIT compiler designed to make unmodified PyTorch programs faster.
666DZY666/aimet
AIMET is a library that provides advanced quantization and compression techniques for trained neural network models.
666DZY666/BRECQ
Pytorch implementation of BRECQ, ICLR 2021
666DZY666/CenterNet
Object detection, 3D detection, and pose estimation using center point detection:
666DZY666/examples
A set of examples around pytorch in Vision, Text, Reinforcement Learning, etc.
666DZY666/how-to-learn-deep-learning-framework
how to learn PyTorch and OneFlow
666DZY666/IntraQ
Pytorch implementation of our paper accepted by CVPR 2022 -- IntraQ: Learning Synthetic Images with Intra-Class Heterogeneity for Zero-Shot Network Quantization
666DZY666/KuiperInfer
带你从零实现一个高性能的深度学习推理库,支持Unet、Yolov5、Resnet等模型的推理。Implement a high-performance deep learning inference library step by step
666DZY666/mindspore
MindSpore is a new open source deep learning training/inference framework that could be used for mobile, edge and cloud scenarios.
666DZY666/mlc-llm
Enable everyone to develop, optimize and deploy AI models natively on everyone's devices.
666DZY666/MQBench
Model Quantization Benchmark
666DZY666/nndeploy
nndeploy is a cross-platform, high-performing, and straightforward AI model deployment framework. We strive to deliver a consistent and user-friendly experience across various inference framework in complex deployment environments and focus on performance. nndeploy一款跨平台、高性能、简单易用的模型端到端部署框架。我们致力于屏蔽不同推理框架的差异,提供一致且用户友好的编程体验,同时专注于部署全流程的性能。
666DZY666/onnx-modifier
A tool to modify onnx models in a visualization fashion, based on Netron and flask.
666DZY666/oscillations-qat
666DZY666/ppq
666DZY666/Python-100-Days
Python - 100天从新手到大师
666DZY666/python-patterns
A collection of design patterns/idioms in Python
666DZY666/pytorch-OpCounter
Count the MACs / FLOPs of your PyTorch model.
666DZY666/pytorch-tutorial
PyTorch Tutorial for Deep Learning Researchers
666DZY666/Quantformer
This is the official pytorch implementation for the paper: *Quantformer: Learning Extremely Low-precision Vision Transformers*.
666DZY666/TinyNeuralNetwork
TinyNeuralNetwork is an efficient and easy-to-use deep learning model compression framework.
666DZY666/torchdistill
A coding-free framework built on PyTorch for reproducible deep learning studies. 🏆20 knowledge distillation methods presented at CVPR, ICLR, ECCV, NeurIPS, ICCV, etc are implemented so far. 🎁 Trained models, training logs and configurations are available for ensuring the reproducibiliy and benchmark.
666DZY666/transformer-quantization
666DZY666/tuning_playbook
A playbook for systematically maximizing the performance of deep learning models.
666DZY666/tvm_learn
tvm learn
666DZY666/vision
Datasets, Transforms and Models specific to Computer Vision