triton-server
There are 18 repositories under triton-server topic.
Bobo-y/flexible-yolov5
More readable and flexible yolov5 with more backbone(gcn, resnet, shufflenet, moblienet, efficientnet, hrnet, swin-transformer, etc) and (cbam,dcn and so on), and tensorrt
notAI-tech/fastDeploy
Deploy DL/ ML inference pipelines with minimal extra code.
levipereira/triton-server-yolo
This repository serves as an example of deploying the YOLO models on Triton Server for performance and testing purposes
chiehpower/Setup-deeplearning-tools
Set up CI in DL/ cuda/ cudnn/ TensorRT/ onnx2trt/ onnxruntime/ onnxsim/ Pytorch/ Triton-Inference-Server/ Bazel/ Tesseract/ PaddleOCR/ NVIDIA-docker/ minIO/ Supervisord on AGX or PC from scratch.
omarabid59/yolov8-triton
Provides an ensemble model to deploy a YoloV8 ONNX model to Triton
kozistr/triton-grpc-proxy-rs
Proxy server for triton gRPC server that inferences embedding model in Rust
levipereira/triton-client-yolo
This repository utilizes the Triton Inference Server Client, which streamlines the complexity of model deployment.
levipereira/deepstream-yolo-triton-server-rtsp-out
The Purpose of this repository is to create a DeepStream/Triton-Server sample application that utilizes yolov7, yolov7-qat, yolov9 models to perform inference on video files or RTSP streams.
Bobo-y/django_ai_server_template
A template for delpoy AI server use django with tf_serving or triton_inference_serving
levipereira/Docker-Yolov7-Nvidia-Kit
This repository provides an out-of-the-box deployment solution for creating an end-to-end procedure to train, deploy, and use Yolov7 models on Nvidia GPUs using Triton Server and Deepstream.
levipereira/nvdsinfer_yolov7_efficient_nms
NvDsInferYolov7EfficientNMS for Gst-nvinferserver
Bobo-y/triton-inference-server-backend-demo
This is a cutom triton-inference-server backend demo for process image (resize + norm)
tamanna18/Triton-Inference-Server-Deployment-with-ONNX-Models
Triton Inference Server Deployment with ONNX Models
Achiwilms/NVIDIA-Triton-Deployment-Quickstart
QuickStart for Deploying a Basic Model on the Triton Inference Server
msclock/transformersplus
Add Some plus extra features to transformers
RostigerDagmer/hermann-reesearch
Pipeline to insert text embeddings generated from self hosted embedding model into qdrant vector database using grpc in rust
cnwangjie/triton-client-js
A Node.js client for the Triton Inference Server.
xdilshod/retail-yolov8
Yolov8 onnx inference running on triton server