Pinned Repositories
aistore
AIStore: scalable storage for AI applications
cuopt
GPU accelerated decision optimization
cuopt-examples
NVIDIA cuOpt examples for decision optimization
DeepLearningExamples
State-of-the-Art Deep Learning scripts organized by models - easy to train and deploy with reproducible accuracy and performance on enterprise-grade infrastructure.
GenerativeAIExamples
Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.
Megatron-LM
Ongoing research training transformer models at scale
nvidia-container-toolkit
Build and run containers leveraging NVIDIA GPUs
nvidia-docker
Build and run Docker containers leveraging NVIDIA GPUs
open-gpu-kernel-modules
NVIDIA Linux open GPU kernel module source
TensorRT
NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.
NVIDIA Corporation's Repositories
NVIDIA/Megatron-LM
Ongoing research training transformer models at scale
NVIDIA/TensorRT-LLM
TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in a performant way.
NVIDIA/cuda-python
CUDA Python: Performance meets Productivity
NVIDIA/TransformerEngine
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit and 4-bit floating point (FP8 and FP4) precision on Hopper, Ada and Blackwell GPUs, to provide better performance with lower memory utilization in both training and inference.
NVIDIA/nv-ingest
NeMo Retriever extraction is a scalable, performance-oriented document content and metadata extraction microservice. NeMo Retriever extraction uses specialized NVIDIA NIM microservices to find, contextualize, and extract text, tables, charts and images that you can use in downstream generative applications.
NVIDIA/gpu-operator
NVIDIA GPU Operator creates, configures, and manages GPUs in Kubernetes
NVIDIA/stdexec
`std::execution`, the proposed C++ framework for asynchronous and parallel programming.
NVIDIA/cccl
CUDA Core Compute Libraries
NVIDIA/aistore
AIStore: scalable storage for AI applications
NVIDIA/TensorRT-Model-Optimizer
A unified library of state-of-the-art model optimization techniques like quantization, pruning, distillation, speculative decoding, etc. It compresses deep learning models for downstream deployment frameworks like TensorRT-LLM or TensorRT to optimize inference speed.
NVIDIA/NeMo-Agent-Toolkit
The NVIDIA NeMo Agent toolkit is an open-source library for efficiently connecting and optimizing teams of AI agents.
NVIDIA/KAI-Scheduler
KAI Scheduler is an open source Kubernetes Native scheduler for AI workloads at large scale
NVIDIA/cuda-quantum
C++ and Python support for the CUDA Quantum programming model for heterogeneous quantum-classical workflows
NVIDIA/NeMo-Skills
A project to improve skills of large language models
NVIDIA/bionemo-framework
BioNeMo Framework: For building and adapting AI models in drug discovery at scale
NVIDIA/cuopt
GPU accelerated decision optimization
NVIDIA/tilus
Tilus is a tile-level kernel programming language with explicit control over shared memory and registers.
NVIDIA/Fuser
A Fusion Code Generator for NVIDIA GPUs (commonly known as "nvFuser")
NVIDIA/JAX-Toolbox
JAX-Toolbox
NVIDIA/mig-parted
MIG Partition Editor for NVIDIA GPUs
NVIDIA/nim-deploy
A collection of YAML files, Helm Charts, Operator code, and guides to act as an example reference implementation for NVIDIA NIM deployment.
NVIDIA/recsys-examples
Examples for Recommenders - easy to train and deploy on accelerated infrastructure.
NVIDIA/vgpu-device-manager
NVIDIA vGPU Device Manager manages NVIDIA vGPU devices on top of Kubernetes
NVIDIA/cudaqx
Accelerated libraries for quantum-classical computing built on CUDA-Q.
NVIDIA/NV-Kernels
Ubuntu kernels which are optimized for NVIDIA server systems
NVIDIA/gontainer
Simple but powerful dependency injection container for Go projects!
NVIDIA/doca-platform
DOCA Platform manages provisioning and service orchestration for Bluefield DPUs
NVIDIA/spark-rapids-jni
RAPIDS Accelerator JNI For Apache Spark
NVIDIA/cloud-native-docs
Documentation repository for NVIDIA Cloud Native Technologies
NVIDIA/doca-sosreport
A unified tool for collecting system logs and other debug information