zhiheng-ldj's Stars
triton-inference-server/server
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
alibaba/TinyNeuralNetwork
TinyNeuralNetwork is an efficient and easy-to-use deep learning model compression framework.
NVIDIA/DALI
A GPU-accelerated library containing highly optimized building blocks and an execution engine for data processing to accelerate deep learning training and inference applications.