Pinned Repositories
bert-finetune
finetune BERT with lightning and hugging face
cuda-lab
simple CUDA kernels and Python bindings
dspy
DSPy: The framework for programming—not prompting—language models
Hands-On-Large-Language-Models
Official code repo for the O'Reilly Book - "Hands-On Large Language Models"
HF-smol-course
A course on aligning smol models.
jxtngx
nemo-lab
Generative AI with NVIDIA NeMo
NVIDIA-GenerativeAIExamples
Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.
NVIDIA-NeMo
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
OpenHackathons-End-to-End-LLM
This repository is an AI Bootcamp material that consist of a workflow for LLM
jxtngx's Repositories
jxtngx/vision-lab
vision transformers with pytorch and pytorch lightning
jxtngx/bert-finetune
finetune BERT with lightning and hugging face
jxtngx/cuda-lab
simple CUDA kernels and Python bindings
jxtngx/Hands-On-Large-Language-Models
Official code repo for the O'Reilly Book - "Hands-On Large Language Models"
jxtngx/HF-smol-course
A course on aligning smol models.
jxtngx/jxtngx
jxtngx/lightning-lab
deep learning with pytorch lightning
jxtngx/LLMs-from-scratch
Implement a ChatGPT-like LLM in PyTorch from scratch, step by step
jxtngx/Meta-pytorch
Tensors and Dynamic neural networks in Python with strong GPU acceleration
jxtngx/nemo-lab
Generative AI with NVIDIA NeMo
jxtngx/NVIDIA-GenerativeAIExamples
Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.
jxtngx/NVIDIA-NeMo
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
jxtngx/NVIDIA-TensorRT
NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.
jxtngx/NVIDIA-TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
jxtngx/NVIDIA-triton-inference-server
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
jxtngx/OpenHackathons-End-to-End-LLM
This repository is an AI Bootcamp material that consist of a workflow for LLM
jxtngx/NVIDIA-DeepLearningExamples
State-of-the-Art Deep Learning scripts organized by models - easy to train and deploy with reproducible accuracy and performance on enterprise-grade infrastructure.
jxtngx/NVIDIA-digital-human
NIM Agent Blueprint for digital human for customer service.
jxtngx/NVIDIA-langchain-nvidia
jxtngx/NVIDIA-NeMo-Aligner
Scalable toolkit for efficient model alignment
jxtngx/NVIDIA-NeMo-Curator
Scalable data pre processing and curation toolkit for LLMs
jxtngx/NVIDIA-NeMo-Guardrails
NeMo Guardrails is an open-source toolkit for easily adding programmable guardrails to LLM-based conversational systems.
jxtngx/NVIDIA-NeMo-Run
A tool to configure, launch and manage your machine learning experiments.
jxtngx/NVIDIA-nim-anywhere
Accelerate your Gen AI with NVIDIA NIM and NVIDIA AI Workbench
jxtngx/NVIDIA-nim-deploy
A collection of YAML files, Helm Charts, Operator code, and guides to act as an example reference implementation for NVIDIA NIM deployment.
jxtngx/OpenHackathons-End-to-End-AI-for-Science
This repository containts materials for End-to-End AI for Science
jxtngx/resnet-finetune
finetune ResNet with lightning and hugging face
jxtngx/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs