Pinned Repositories
onnx
Open Neural Network Exchange
onnx-feedstock
A conda-smithy repository for onnx.
onnx-tensorrt
ONNX-TensorRT: TensorRT backend for ONNX
onnxruntime
ONNX Runtime: cross-platform, high performance scoring engine for ML models
optimum
🚀 Accelerate training and inference of 🤗 Transformers and 🤗 Diffusers with easy to use hardware optimization tools
protobuf
Protocol Buffers - Google's data interchange format
pytorch
Tensors and Dynamic neural networks in Python with strong GPU acceleration
TensorRT
NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.
TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
wheel-builder
Utils for building and publishing ONNX wheels
kevinch-nv's Repositories
kevinch-nv/onnx-tensorrt
ONNX-TensorRT: TensorRT backend for ONNX
kevinch-nv/onnx
Open Neural Network Exchange
kevinch-nv/onnx-feedstock
A conda-smithy repository for onnx.
kevinch-nv/onnxruntime
ONNX Runtime: cross-platform, high performance scoring engine for ML models
kevinch-nv/optimum
🚀 Accelerate training and inference of 🤗 Transformers and 🤗 Diffusers with easy to use hardware optimization tools
kevinch-nv/protobuf
Protocol Buffers - Google's data interchange format
kevinch-nv/pytorch
Tensors and Dynamic neural networks in Python with strong GPU acceleration
kevinch-nv/TensorRT
NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.
kevinch-nv/wheel-builder
Utils for building and publishing ONNX wheels
kevinch-nv/working-groups
Repository for ONNX working group artifacts