SadTalker Triton server demo

Env

  1. 服务镜像 >= tritonserver_22.11 差的包,参考*_reference.txt
  2. 客户端(或tritonclient 镜像) 差的包,参考*_reference.txt

Using

百度网盘链接 提取码: gogs

  1. download weights.zip and unzip weights.zip in onnx_weights and add generator.onnx and kp_detector.onnx
  2. cd SadTalkerTriton
  3. docker run -v $PWD:/models --gpus device=1 --shm-size 1g -it --name sadtalker_triton tritonserver_22.11_pypack(your image name)
  4. in server container: tritonserve --model-repo ./
  5. in client container: python client.py

You can do

  1. Dynamic batching
  2. More efficient organization compliant with triton
  3. int8 for generator network

Reminder

  1. generator.onnx and kp_detector.onnx you must do it by yourself,it's big and some licence.

galary

example