/tiny-tensorrt

a simple, efficient, easy-to-use nvidia TensorRT wrapper for cnn,sopport c++ and python

Primary LanguageC++

English | 中文简体

image

tiny-tensorrt

a simple, efficient, easy-to-use nvidia TensorRT wrapper for cnn with c++ and python api,support caffe, uff and onnx format models.

Features

  • Custom plugin tutorial and well_commented sample! ---2019-12-11 🔥🔥🔥
  • Custom onnx model output node ---2019.10.18
  • Upgrade with TensorRT 6.0.1.5 --- 2019.9.29
  • Support onnx,caffe and tensorflow model
  • Support more model and layer --working on
  • PReLU and up-sample plugin
  • Engine serialization and deserialization
  • INT8 support for caffe model
  • Python api support
  • Set device

System Requirements

cuda 10.0+

TensorRT 6.x

for python api, python 2.x/3.x and numpy in needed

this project is fully test in ubuntu 16.04. I tested it with 1060ti, 1050ti, 1080ti, 1660ti, 2080, 2080ti and p4.

Docs

User Guide

Custom Plugin Tutorial

Extra Support layer

  • upsample with custom scale, under test with yolov3.
  • yolo-det, last layer of yolov3 which sum three scales output and generate final result for nms. under test with yolov3.
  • PRELU, under test with openpose and mtcnn.

About License

For the 3rd-party module and TensorRT, maybe you need to follow their license

For the part I wrote, you can do anything you want