qat

There are 29 repositories under qat topic.

  • Xilinx/brevitas

    Brevitas: neural network quantization in PyTorch

    Language:Python1.4k30512227
  • Bobo-y/flexible-yolov5

    More readable and flexible yolov5 with more backbone(gcn, resnet, shufflenet, moblienet, efficientnet, hrnet, swin-transformer, etc) and (cbam,dcn and so on), and tensorrt

    Language:Python67910146119
  • NVIDIA-AI-IOT/yolo_deepstream

    yolo model qat and deploy with deepstream&tensorrt

    Language:Python5861565143
  • sony/model_optimization

    Model Compression Toolkit (MCT) is an open source project for neural network model optimization under efficient, constrained hardware. This project provides researchers, developers, and engineers advanced quantization and compression tools for deploying state-of-the-art neural networks.

    Language:Python3832214164
  • THU-MIG/torch-model-compression

    针对pytorch模型的自动化模型结构分析和修改工具集,包含自动分析模型结构的模型压缩算法库

    Language:Python252122041
  • NVIDIA-AI-IOT/clip-distillation

    Zero-label image classification via OpenCLIP knowledge distillation

    Language:Python1347314
  • levipereira/yolov9-qat

    Implementation of YOLOv9 QAT optimized for deployment on TensorRT platforms.

    Language:Python12341418
  • electrocucaracha/krd

    Kubernetes Reference Deployment

    Language:Shell416610
  • DeadAt0m/LSQFakeQuantize-PyTorch

    FakeQuantize with Learned Step Size(LSQ+) as Observer in PyTorch

    Language:C++34136
  • ZLkanyo009/MQBench_Quantize

    QAT(quantize aware training) for classification with MQBench

    Language:Python28023
  • TsingmaoAI/MI-optimize

    mi-optimize is a versatile tool designed for the quantization and evaluation of large language models (LLMs). The library's seamless integration of various quantization methods and evaluation techniques empowers users to customize their approaches according to specific requirements and constraints, providing a high level of flexibility.

    Language:Python23005
  • intel/intel-technology-enabling-for-openshift

    The project delivers a comprehensive full-stack solution for the Intel® Enterprise AI Foundation on the OpenShift platform to provision Intel AI and Xeon accelerators, integrate AI software, and enable key AI workloads, such as LLM inferencing and fine-tuning for enterprise AI. GPU network provisioning is currently in the planning stage.

    Language:Python1754213
  • lix19937/tensorrt-insight

    Deep insight tensorrt, including but not limited to qat, ptq, plugin, triton_inference, cuda

    Language:C++161710
  • AXERA-TECH/yolov5-qat

    Try to export the ONNX QDQ model that conforms to the AXERA NPU quantization specification. Currently, only w8a8 is supported.

    Language:Python9001
  • yester31/Quantization_EX

    quantization example for pqt & qat

    Language:Python7102
  • chris010970/qat

    Training U-Net based Convolutional Neural Network model to automatically identify and delineate areas of qat agriculture in Sentinel-2 multispectral imagery.

    Language:Jupyter Notebook3102
  • yester31/TensorRT_Examples

    TensorRT in Practice: Model Conversion, Extension, and Advanced Inference Optimization

    Language:Python3202
  • BlindOver/blindover_AI

    Build AI model to classify beverages for blind individuals

    Language:Python2041
  • electrocucaracha/bootstrap-vagrant

    Vagrant installation script

    Language:Shell2211
  • qatlang/qatlang.org

    Official website of the qat programming language

    Language:TypeScript2100
  • Warrfie/combidata

    Combidata is a flexible and powerful Python library designed for generating various combinations of test data based on defined cases and rules. It is especially useful for testing, debugging, and analyzing software applications and systems.

    Language:Python2100
  • OmidGhadami95/EfficientNetV2_Quantization_CK

    EfficientNetV2 (Efficientnetv2-b2) and quantization int8 and fp32 (QAT and PTQ) on CK+ dataset . fine-tuning, augmentation, solving imbalanced dataset, etc.

    Language:Jupyter Notebook1100
  • qatlang/qatls

    Official LSP for qat

    Language:Go1100
  • qatlang/tree-sitter-qat

    Official implementation of the treesitter grammar for qat

  • dohdoh64/qat-website

    A repo for our website that we are making

    Language:HTML0300
  • witmemtech/Witin-NN-Tool-

    The "witin_nn" framework, based on PyTorch, maps neural networks to chip computations and supports operators including Linear, Conv2d, and GruCell. It enables 8-12 bit quantization for inputs/outputs and weights, implementing QAT.

    Language:Python0010
  • aldrinmathew/AldrinMathew

    This is a special repository for showcasing information about me and my projects...

  • ambideXtrous9/Quantization-of-Models-PTQ-and-QAT

    Quantization of Models : Post-Training Quantization(PTQ) and Quantize Aware Training(QAT)

    Language:Jupyter Notebook10
  • jahongir7174/CTDNet-qat

    Quantization Aware Training Implementation of CTDNet

    Language:Python00