Pinned Repositories
adaptdl
Resource-adaptive cluster scheduler for deep learning training.
angular-6wh5ma
Created with StackBlitz ⚡️
annotated-transformer
An annotated implementation of the Transformer paper.
ant-design
An enterprise-class UI design language and React UI library
arena
A CLI for Kubeflow.
Baichuan-13B
A 13B large language model developed by Baichuan Intelligent Technology
Baichuan-7B
A large-scale 7B pretraining language model developed by BaiChuan-Inc.
BetterDL
DeepSpeedExamples
Example models using DeepSpeed
pytorch_examples
A set of examples around pytorch in Vision, Text, Reinforcement Learning, etc.
tingshua-yts's Repositories
tingshua-yts/BetterDL
tingshua-yts/DeepSpeedExamples
Example models using DeepSpeed
tingshua-yts/annotated-transformer
An annotated implementation of the Transformer paper.
tingshua-yts/arena
A CLI for Kubeflow.
tingshua-yts/Baichuan-13B
A 13B large language model developed by Baichuan Intelligent Technology
tingshua-yts/Baichuan-7B
A large-scale 7B pretraining language model developed by BaiChuan-Inc.
tingshua-yts/BetterServer
tingshua-yts/bigscience
Central place for the engineering/scaling WG: documentation, SLURM scripts and logs, compute environment and data.
tingshua-yts/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
tingshua-yts/djl-serving
A universal scalable machine learning model deployment solution
tingshua-yts/dlrover
DLRover: An Automatic Distributed Deep Learning System
tingshua-yts/et-operator
Kubernetes Operator for AI and Bigdata Elastic Training
tingshua-yts/FasterTransformer
Transformer related optimization, including BERT, GPT
tingshua-yts/fastertransformer_backend
tingshua-yts/gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
tingshua-yts/huggingface-transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
tingshua-yts/kubedl
Run your deep learning workloads on Kubernetes more easily and efficiently.
tingshua-yts/lightseq
LightSeq: A High Performance Library for Sequence Processing and Generation
tingshua-yts/Megatron-LM
Ongoing research training transformer models at scale
tingshua-yts/Pai-Megatron-Patch
tingshua-yts/pytorch
Tensors and Dynamic neural networks in Python with strong GPU acceleration
tingshua-yts/tensorflow
An Open Source Machine Learning Framework for Everyone
tingshua-yts/TensorRT
NVIDIA® TensorRT™, an SDK for high-performance deep learning inference, includes a deep learning inference optimizer and runtime that delivers low latency and high throughput for inference applications.
tingshua-yts/training-operator
Training operators on Kubernetes.
tingshua-yts/transformers-bloom-inference
Fast Inference Solutions for BLOOM
tingshua-yts/triton_core
The core library and APIs implementing the Triton Inference Server.
tingshua-yts/triton_pytorch_backend
The Triton backend for the PyTorch TorchScript models.
tingshua-yts/triton_server
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
tingshua-yts/tvm
Open deep learning compiler stack for cpu, gpu and specialized accelerators
tingshua-yts/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs