Chenyangzh's Stars
rackslab/Slurm-web
Open source web interface for Slurm HPC clusters
NISP-GmbH/SLURM
SLURM Installer
pyenv/pyenv
Simple Python version management
sgl-project/sglang
SGLang is a fast serving framework for large language models and vision language models.
meta-llama/llama3
The official Meta Llama 3 GitHub site
ml-explore/mlx
MLX: An array framework for Apple silicon
eip-work/kuboard-press
Kuboard 是基于 Kubernetes 的微服务管理界面。同时提供 Kubernetes 免费中文教程,入门教程,最新版本的 Kubernetes v1.23.4 安装手册,(k8s install) 在线答疑,持续更新。
tickstep/aliyunpan
阿里云盘命令行客户端,支持JavaScript插件,支持同步备份功能。
wenge-research/YAYI2
YAYI 2 是中科闻歌研发的新一代开源大语言模型,采用了超过 2 万亿 Tokens 的高质量、多语言语料进行预训练。(Repo for YaYi 2 Chinese LLMs)
NVIDIA/TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
ray-project/ray
Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
wenge-research/YAYI
雅意大模型:为客户打造安全可靠的专属大模型,基于大规模中英文多领域指令数据训练的 LlaMA 2 & BLOOM 系列模型,由中科闻歌算法团队研发。(Repo for YaYi Chinese LLMs based on LlaMA2 & BLOOM)
microsoft/pai
Resource scheduling and cluster management for AI
apache/tvm
Open deep learning compiler stack for cpu, gpu and specialized accelerators
deepinsight/insightface
State-of-the-art 2D and 3D Face Analysis Project