Pinned Repositories
optimum
🚀 Accelerate inference and training of 🤗 Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimization tools
optimum-neuron
Easy, fast and very cheap training and inference on AWS Trainium and Inferentia chips.
transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
accelerate-bert-with-tvm
A benchmark on accelerating BERT based on passes in TVM.
onnxruntime-inference-benchmark
JingyaHuang's Repositories
JingyaHuang/onnxruntime-inference-benchmark
JingyaHuang/accelerate-bert-with-tvm
A benchmark on accelerating BERT based on passes in TVM.
JingyaHuang/aws-neuron-sdk
Powering AWS purpose-built machine learning chips. Blazing fast and cost effective, natively integrated into PyTorch and TensorFlow and integrated with your favorite AWS services
JingyaHuang/deep-learning-containers
AWS Deep Learning Containers (DLCs) are a set of Docker images for training and serving models in TensorFlow, TensorFlow 2, PyTorch, and MXNet.
JingyaHuang/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
JingyaHuang/ML-training-benchmark
Benchmark for training within different backend.
JingyaHuang/models
A collection of pre-trained, state-of-the-art models in the ONNX format
JingyaHuang/notebooks
Notebooks using the Hugging Face libraries 🤗
JingyaHuang/onnx
Open standard for machine learning interoperability
JingyaHuang/ort-playground
Demos testing caveat reported while using onnxruntime
JingyaHuang/diffusers
🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch and FLAX.
JingyaHuang/llm-hosting-container
Large Language Model Hosting Container
JingyaHuang/neuron-playground
Demos testing neuron (inf1 / inf2 / trn1)
JingyaHuang/onnx-tutorial
This repo gathers examples of ONNX graph surgery.
JingyaHuang/onnx.github.io
JingyaHuang/onnxruntime
ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
JingyaHuang/onnxruntime-inference-examples
Examples for using ONNX Runtime for machine learning inferencing.
JingyaHuang/optimum
🏎️ Accelerate training and inference of 🤗 Transformers with easy to use hardware optimization tools
JingyaHuang/optimum-neuron
Easy, fast and very cheap training and inference on AWS Trainium and Inferentia chips.
JingyaHuang/PaddleViT
:robot: PaddleViT: State-of-the-art Visual Transformer and MLP Models for PaddlePaddle 2.0+
JingyaHuang/recurrent-batch-normalization-pytorch
PyTorch implementation of recurrent batch normalization
JingyaHuang/sagemaker-python-sdk
A library for training and deploying machine learning models on Amazon SageMaker
JingyaHuang/spleeter
Deezer source separation library including pretrained models.
JingyaHuang/text-generation-inference
Large Language Model Text Generation Inference
JingyaHuang/transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
JingyaHuang/xla
Enabling PyTorch on XLA Devices (e.g. Google TPU)