pruning
There are 432 repositories under pruning topic.
datawhalechina/leedl-tutorial
《李宏毅深度学习教程》(李宏毅老师推荐👍),PDF下载地址:https://github.com/datawhalechina/leedl-tutorial/releases
IntelLabs/distiller
Neural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://intellabs.github.io/distiller
neuralmagic/deepsparse
Sparsity-aware deep learning inference runtime for CPUs
VainF/Torch-Pruning
[CVPR 2023] Towards Any Structural Pruning; LLMs / SAM / Diffusion / Transformers / YOLOv8 / CNNs
he-y/Awesome-Pruning
A curated list of neural network pruning resources.
666DZY666/micronet
micronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、Low-Bit(≤2b)/Ternary and Binary(TWN/BNN/XNOR-Net); post-training-quantization(PTQ), 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization fuse for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape
intel/neural-compressor
SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX Runtime
neuralmagic/sparseml
Libraries for applying sparsification recipes to neural networks with a few lines of code, enabling faster and smaller models
quic/aimet
AIMET is a library that provides advanced quantization and compression techniques for trained neural network models.
PaddlePaddle/PaddleSlim
PaddleSlim is an open-source library for deep model compression and architecture search.
tensorflow/model-optimization
A toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.
open-mmlab/mmrazor
OpenMMLab Model Compression Toolbox and Benchmark.
cupcakearmy/autorestic
Config driven, easy backup cli for restic.
huawei-noah/Efficient-Computing
Efficient computing methods developed by Huawei Noah's Ark Lab
jacobgil/pytorch-pruning
PyTorch Implementation of [1611.06440] Pruning Convolutional Neural Networks for Resource Efficient Inference
openvinotoolkit/nncf
Neural Network Compression Framework for enhanced OpenVINO™ inference
Syencil/mobile-yolov5-pruning-distillation
mobilev2-yolov5s剪枝、蒸馏,支持ncnn,tensorRT部署。ultra-light but better performence!
csarron/awesome-emdl
Embedded and mobile deep learning research resources
alibaba/TinyNeuralNetwork
TinyNeuralNetwork is an efficient and easy-to-use deep learning model compression framework.
horseee/LLM-Pruner
[NeurIPS 2023] LLM-Pruner: On the Structural Pruning of Large Language Models. Support LLaMA, Llama-2, BLOOM, Vicuna, Baichuan, etc.
SforAiDl/KD_Lib
A Pytorch Knowledge Distillation library for benchmarking and extending works in the domains of Knowledge Distillation, Pruning, and Quantization.
he-y/filter-pruning-geometric-median
Filter Pruning via Geometric Median for Deep Convolutional Neural Networks Acceleration (CVPR 2019 Oral)
princeton-nlp/LLM-Shearing
[ICLR 2024] Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruning
cedrickchee/awesome-ml-model-compression
Awesome machine learning model compression research papers, tools, and learning material.
SpursLipu/YOLOv3v4-ModelCompression-MultidatasetTraining-Multibackbone
YOLO ModelCompression MultidatasetTraining
BenWhetton/keras-surgeon
Pruning and other network surgery for trained Keras models.
he-y/soft-filter-pruning
Soft Filter Pruning for Accelerating Deep Convolutional Neural Networks
1duo/awesome-ai-infrastructures
Infrastructures™ for Machine Learning Training/Inference in Production.
neuralmagic/sparsezoo
Neural network model repository for highly sparse and sparse-quantized models with matching sparsification recipes
airaria/TextPruner
A PyTorch-based model pruning toolkit for pre-trained language models
talebolano/yolov3-network-slimming
yolov3 network slimming剪枝的一种实现
huggingface/optimum-intel
🤗 Optimum Intel: Accelerate inference with Intel optimization tools
mehtadushy/SelecSLS-Pytorch
Reference ImageNet implementation of SelecSLS CNN architecture proposed in the SIGGRAPH 2020 paper "XNect: Real-time Multi-Person 3D Motion Capture with a Single RGB Camera". The repository also includes code for pruning the model based on implicit sparsity emerging from adaptive gradient descent methods, as detailed in the CVPR 2019 paper "On implicit filter level sparsity in Convolutional Neural Networks".
megvii-research/Sparsebit
A model compression and acceleration toolbox based on pytorch.
rahulvigneswaran/Lottery-Ticket-Hypothesis-in-Pytorch
This repository contains a Pytorch implementation of the paper "The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks" by Jonathan Frankle and Michael Carbin that can be easily adapted to any model/dataset.
neuralmagic/sparsify
ML model optimization product to accelerate inference.