Pinned Repositories
aistore
AIStore: scalable storage for AI applications
DeepLearningExamples
State-of-the-Art Deep Learning scripts organized by models - easy to train and deploy with reproducible accuracy and performance on enterprise-grade infrastructure.
FastPhotoStyle
Style transfer, deep learning, feature transform
Megatron-LM
Ongoing research training transformer models at scale
NeMo
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
nvidia-container-toolkit
Build and run containers leveraging NVIDIA GPUs
nvidia-docker
Build and run Docker containers leveraging NVIDIA GPUs
open-gpu-kernel-modules
NVIDIA Linux open GPU kernel module source
tensorflow
An Open Source Machine Learning Framework for Everyone
TensorRT
NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.
NVIDIA Corporation's Repositories
NVIDIA/NeMo
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
NVIDIA/apex
A PyTorch Extension: Tools for easy mixed precision and distributed training in Pytorch
NVIDIA/DALI
A GPU-accelerated library containing highly optimized building blocks and an execution engine for data processing to accelerate deep learning training and inference applications.
NVIDIA/cutlass
CUDA Templates for Linear Algebra Subroutines
NVIDIA/ChatRTX
A developer reference project for creating Retrieval Augmented Generation (RAG) chatbots on Windows using TensorRT-LLM
NVIDIA/warp
A Python framework for high performance GPU simulation and graphics
NVIDIA/TransformerEngine
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
NVIDIA/stdexec
`std::execution`, the proposed C++ framework for asynchronous and parallel programming.
NVIDIA/gpu-operator
NVIDIA GPU Operator creates/configures/manages GPUs atop Kubernetes
NVIDIA/MatX
An efficient C++17 GPU numerical computing library with Python-like syntax
NVIDIA/cccl
CUDA C++ Core Libraries
NVIDIA/spark-rapids
Spark RAPIDS plugin - accelerate Apache Spark with GPUs
NVIDIA/modulus
Open-source deep-learning framework for building, training, and fine-tuning deep learning models using state-of-the-art Physics-ML methods
NVIDIA/NVFlare
NVIDIA Federated Learning Application Runtime Environment
NVIDIA/cuCollections
NVIDIA/cuda-quantum
C++ and Python support for the CUDA Quantum programming model for heterogeneous quantum-classical workflows
NVIDIA/NeMo-Framework-Launcher
NeMo Megatron launcher and tools
NVIDIA/NeMo-Aligner
Scalable toolkit for efficient model alignment
NVIDIA/Fuser
A Fusion Code Generator for NVIDIA GPUs (commonly known as "nvFuser")
NVIDIA/nvbandwidth
A tool for bandwidth measurements on NVIDIA GPUs.
NVIDIA/edk2-nvidia
NVIDIA EDK2 platform support
NVIDIA/gpu-driver-container
The NVIDIA GPU driver container allows the provisioning of the NVIDIA driver through the use of containers.
NVIDIA/spark-rapids-tools
User tools for Spark RAPIDS
NVIDIA/earth2studio
Open-source deep-learning framework for exploring, building and deploying AI weather/climate workflows.
NVIDIA/knavigator
knavigator is a development, testing, and optimization toolkit for AI/ML scheduling systems at scale on Kubernetes.
NVIDIA/spark-rapids-jni
RAPIDS Accelerator JNI For Apache Spark
NVIDIA/edk2
NVIDIA fork of tianocore/edk2
NVIDIA/cloudai
CloudAI Benchmark Framework
NVIDIA/numbast
Numbast is a tool to build an automated pipeline that converts CUDA APIs into Numba bindings.
NVIDIA/tinylinux-scripts
Scripts for building minimal Linux distribution for diagnostics