int8-inference
There are 13 repositories under int8-inference topic.
anilsathyan7/Portrait-Segmentation
Real-time portrait segmentation for mobile devices
BUG1989/caffe-int8-convert-tools
Generate a quantization parameter file for ncnn framework int8 inference
DerryHub/BEVFormer_tensorrt
BEVFormer inference on TensorRT, including INT8 Quantization and Custom TensorRT Plugins (float/half/half2/int8).
jahongir7174/YOLOv8-qat
Quantization Aware Training
Howell-Yang/onnx2trt
将端上模型部署过程中,常见的问题以及解决办法记录并汇总,希望能给其他人带来一点帮助。
whitelok/tensorrt-int8-python-sample
TensorRT Int8 Python version sample. TensorRT Int8 Python 实现例子。TensorRT Int8 Pythonの例です
ENOT-AutoDL/gpt-j-6B-tensorrt-int8
GPT-J 6B inference on TensorRT with INT-8 precision
JohnClaw/chatllm.vb
VB.NET api wrapper for llm-inference chatllm.cpp
yester31/TensorRT_ONNX
Generating tensorrt model using onnx
JohnClaw/chatllm.cs
C# api wrapper for llm-inference chatllm.cpp
akashAD98/yolov7_vino_with_object_tracking
it has support for openvino converted model of yolov7-int.xml ,yolov7x,
daniel-rychlewski/cnn-planesnet
Compressed CNNs for airplane classification in satellite images (APoZ-based parameter pruning, INT8 weight quantization)