yiliu30's Stars
microsoft/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
ray-project/ray
Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
huggingface/pytorch-image-models
The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (ViT), MobileNetV4, MobileNet-V3 & V2, RegNet, DPN, CSPNet, Swin Transformer, MaxViT, CoAtNet, ConvNeXt, and more
microsoft/onnxruntime
ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
apache/tvm
Open deep learning compiler stack for cpu, gpu and specialized accelerators
hyperopt/hyperopt
Distributed Asynchronous Hyperparameter Optimization in Python
MegEngine/MegEngine
MegEngine 是一个快速、可拓展、易于使用且支持自动求导的深度学习框架
facebookincubator/AITemplate
AITemplate is a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (NVIDIA GPU) and MatrixCore (AMD GPU) inference.
IntelLabs/distiller
Neural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://intellabs.github.io/distiller
mindspore-ai/mindspore
MindSpore is a new open source deep learning training/inference framework that could be used for mobile, edge and cloud scenarios.
Jittor/jittor
Jittor is a high-performance deep learning framework based on JIT compiling and meta-operators.
automl/Auto-PyTorch
Automatic architecture search and hyperparameter optimization for PyTorch
google/gemmlowp
Low-precision matrix multiplication
facebookresearch/TensorComprehensions
A domain specific language to express machine learning workloads.
flame/how-to-optimize-gemm
tensorflow/model-optimization
A toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.
baidu-research/DeepBench
Benchmarking Deep Learning operations on different hardware
dmlc/dlpack
common in-memory tensor structure
huggingface/optimum-intel
🤗 Optimum Intel: Accelerate inference with Intel optimization tools
jakc4103/DFQ
PyTorch implementation of Data Free Quantization Through Weight Equalization and Bias Correction.
camlsys/degree-quant
ICLR 2021
pengzhao-intel/oneAPI_course
oneAPI - Data Parallel C++ course for students