qat

There are 26 repositories under qat topic.

  • Xilinx/brevitas

    Brevitas: neural network quantization in PyTorch

    Language:Python1.2k34433192
  • Bobo-y/flexible-yolov5

    More readable and flexible yolov5 with more backbone(gcn, resnet, shufflenet, moblienet, efficientnet, hrnet, swin-transformer, etc) and (cbam,dcn and so on), and tensorrt

    Language:Python66510146120
  • NVIDIA-AI-IOT/yolo_deepstream

    yolo model qat and deploy with deepstream&tensorrt

    Language:Python5361661136
  • sony/model_optimization

    Model Compression Toolkit (MCT) is an open source project for neural network model optimization under efficient, constrained hardware. This project provides researchers, developers, and engineers advanced quantization and compression tools for deploying state-of-the-art neural networks.

    Language:Python3002311749
  • THU-MIG/torch-model-compression

    针对pytorch模型的自动化模型结构分析和修改工具集,包含自动分析模型结构的模型压缩算法库

    Language:Python236122040
  • NVIDIA-AI-IOT/clip-distillation

    Zero-label image classification via OpenCLIP knowledge distillation

    Language:Python1077315
  • levipereira/yolov9-qat

    Implementation of YOLOv9 QAT optimized for deployment on TensorRT platforms.

    Language:Python7431011
  • electrocucaracha/krd

    Kubernetes Reference Deployment

    Language:Shell396610
  • ZLkanyo009/MQBench_Quantize

    QAT(quantize aware training) for classification with MQBench

    Language:Python35023
  • DeadAt0m/LSQFakeQuantize-PyTorch

    FakeQuantize with Learned Step Size(LSQ+) as Observer in PyTorch

    Language:C++32136
  • TsingmaoAI/MI-optimize

    mi-optimize is a versatile tool designed for the quantization and evaluation of large language models (LLMs). The library's seamless integration of various quantization methods and evaluation techniques empowers users to customize their approaches according to specific requirements and constraints, providing a high level of flexibility.

    Language:Python18104
  • intel/intel-technology-enabling-for-openshift

    The project focuses on Intel’s enterprise AI and cloud native foundation for Red Hat OpenShift Container Platform (RHOCP) solution enablement and innovation including Intel data center hardware features, Intel technology enhanced AI platform and the referenced AI workloads provisioning for OpenShift.

    Language:Python176389
  • lix19937/tensorrt-insight

    Deep insight tensorrt, including but not limited to qat, ptq, plugin, triton_inference, cuda

    Language:C++61410
  • yester31/Quantization_EX

    quantization example for pqt & qat

    Language:Python4102
  • chris010970/qat

    Training U-Net based Convolutional Neural Network model to automatically identify and delineate areas of qat agriculture in Sentinel-2 multispectral imagery.

    Language:Jupyter Notebook3102
  • BlindOver/blindover_AI

    Build AI model to classify beverages for blind individuals

    Language:Python2141
  • electrocucaracha/bootstrap-vagrant

    Vagrant installation script

    Language:Shell2211
  • qatlang/QatDev

    Official website of qat

    Language:TypeScript2100
  • Warrfie/combidata

    Combidata is a flexible and powerful Python library designed for generating various combinations of test data based on defined cases and rules. It is especially useful for testing, debugging, and analyzing software applications and systems.

    Language:Python2200
  • OmidGhadami95/EfficientNetV2_Quantization_CK

    EfficientNetV2 (Efficientnetv2-b2) and quantization int8 and fp32 (QAT and PTQ) on CK+ dataset . fine-tuning, augmentation, solving imbalanced dataset, etc.

    Language:Jupyter Notebook1100
  • qatlang/tree-sitter-qat

    Official implementation of the treesitter grammar for qat

    10
  • dohdoh64/qat-website

    A repo for our website that we are making

    Language:HTML0300
  • qatlang/QatDevServer

    Server for https://qat.dev - official site of the Qat programming language...

    Language:Go0100
  • witmem/Witin-NN-Tool-

    The "witin_nn" framework, based on PyTorch, maps neural networks to chip computations and supports operators including Linear, Conv2d, and GruCell. It enables 8-12 bit quantization for inputs/outputs and weights, implementing QAT.

    Language:Python0100
  • AldrinMathew/AldrinMathew

    This is a special repository for showcasing information about me and my projects...

  • ambideXtrous9/Quantization-of-Models-PTQ-and-QAT

    Quantization of Models : Post-Training Quantization(PTQ) and Quantize Aware Training(QAT)

    Language:Jupyter Notebook10