qat
There are 29 repositories under qat topic.
Xilinx/brevitas
Brevitas: neural network quantization in PyTorch
Bobo-y/flexible-yolov5
More readable and flexible yolov5 with more backbone(gcn, resnet, shufflenet, moblienet, efficientnet, hrnet, swin-transformer, etc) and (cbam,dcn and so on), and tensorrt
NVIDIA-AI-IOT/yolo_deepstream
yolo model qat and deploy with deepstream&tensorrt
sony/model_optimization
Model Compression Toolkit (MCT) is an open source project for neural network model optimization under efficient, constrained hardware. This project provides researchers, developers, and engineers advanced quantization and compression tools for deploying state-of-the-art neural networks.
THU-MIG/torch-model-compression
针对pytorch模型的自动化模型结构分析和修改工具集,包含自动分析模型结构的模型压缩算法库
NVIDIA-AI-IOT/clip-distillation
Zero-label image classification via OpenCLIP knowledge distillation
levipereira/yolov9-qat
Implementation of YOLOv9 QAT optimized for deployment on TensorRT platforms.
electrocucaracha/krd
Kubernetes Reference Deployment
DeadAt0m/LSQFakeQuantize-PyTorch
FakeQuantize with Learned Step Size(LSQ+) as Observer in PyTorch
ZLkanyo009/MQBench_Quantize
QAT(quantize aware training) for classification with MQBench
TsingmaoAI/MI-optimize
mi-optimize is a versatile tool designed for the quantization and evaluation of large language models (LLMs). The library's seamless integration of various quantization methods and evaluation techniques empowers users to customize their approaches according to specific requirements and constraints, providing a high level of flexibility.
intel/intel-technology-enabling-for-openshift
The project delivers a comprehensive full-stack solution for the Intel® Enterprise AI Foundation on the OpenShift platform to provision Intel AI and Xeon accelerators, integrate AI software, and enable key AI workloads, such as LLM inferencing and fine-tuning for enterprise AI. GPU network provisioning is currently in the planning stage.
lix19937/tensorrt-insight
Deep insight tensorrt, including but not limited to qat, ptq, plugin, triton_inference, cuda
AXERA-TECH/yolov5-qat
Try to export the ONNX QDQ model that conforms to the AXERA NPU quantization specification. Currently, only w8a8 is supported.
yester31/Quantization_EX
quantization example for pqt & qat
chris010970/qat
Training U-Net based Convolutional Neural Network model to automatically identify and delineate areas of qat agriculture in Sentinel-2 multispectral imagery.
yester31/TensorRT_Examples
TensorRT in Practice: Model Conversion, Extension, and Advanced Inference Optimization
BlindOver/blindover_AI
Build AI model to classify beverages for blind individuals
electrocucaracha/bootstrap-vagrant
Vagrant installation script
qatlang/qatlang.org
Official website of the qat programming language
Warrfie/combidata
Combidata is a flexible and powerful Python library designed for generating various combinations of test data based on defined cases and rules. It is especially useful for testing, debugging, and analyzing software applications and systems.
OmidGhadami95/EfficientNetV2_Quantization_CK
EfficientNetV2 (Efficientnetv2-b2) and quantization int8 and fp32 (QAT and PTQ) on CK+ dataset . fine-tuning, augmentation, solving imbalanced dataset, etc.
qatlang/qatls
Official LSP for qat
qatlang/tree-sitter-qat
Official implementation of the treesitter grammar for qat
dohdoh64/qat-website
A repo for our website that we are making
witmemtech/Witin-NN-Tool-
The "witin_nn" framework, based on PyTorch, maps neural networks to chip computations and supports operators including Linear, Conv2d, and GruCell. It enables 8-12 bit quantization for inputs/outputs and weights, implementing QAT.
aldrinmathew/AldrinMathew
This is a special repository for showcasing information about me and my projects...
ambideXtrous9/Quantization-of-Models-PTQ-and-QAT
Quantization of Models : Post-Training Quantization(PTQ) and Quantize Aware Training(QAT)
jahongir7174/CTDNet-qat
Quantization Aware Training Implementation of CTDNet