zhao9797's Stars
codecrafters-io/build-your-own-x
Master programming by recreating your favorite technologies from scratch.
langgenius/dify
Dify is an open-source LLM app development platform. Dify's intuitive interface combines AI workflow, RAG pipeline, agent capabilities, model management, observability features and more, letting you quickly go from prototype to production.
xai-org/grok-1
Grok open release
ray-project/ray
Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
chatchat-space/Langchain-Chatchat
Langchain-Chatchat(原Langchain-ChatGLM)基于 Langchain 与 ChatGLM, Qwen 与 Llama 等语言模型的 RAG 与 Agent 应用 | Langchain-Chatchat (formerly langchain-ChatGLM), local knowledge based LLM (like ChatGLM, Qwen and Llama) RAG and Agent app with langchain
myshell-ai/OpenVoice
Instant voice cloning by MIT and MyShell. Audio foundation model.
meta-llama/llama3
The official Meta Llama 3 GitHub site
opendatalab/MinerU
A high-quality tool for convert PDF to Markdown and JSON.一站式开源高质量数据提取工具,将PDF转换成Markdown和JSON格式。
hpcaitech/Open-Sora
Open-Sora: Democratizing Efficient Video Production for All
BerriAI/litellm
Python SDK, Proxy Server (LLM Gateway) to call 100+ LLM APIs in OpenAI format - [Bedrock, Azure, OpenAI, VertexAI, Cohere, Anthropic, Sagemaker, HuggingFace, Replicate, Groq]
NVIDIA/Megatron-LM
Ongoing research training transformer models at scale
lucidrains/imagen-pytorch
Implementation of Imagen, Google's Text-to-Image Neural Network, in Pytorch
apple/corenet
CoreNet: A library for training deep neural networks
pytorch/torchchat
Run PyTorch LLMs locally on servers, desktop and mobile
pytorch/torchtitan
A PyTorch native library for large model training
avinashkranjan/Amazing-Python-Scripts
🚀 Curated collection of Amazing Python scripts from Basics to Advance with automation task scripts.
NVIDIA/TransformerEngine
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
baaivision/Emu
Emu Series: Generative Multimodal Models from BAAI
facebookresearch/multimodal
TorchMultimodal is a PyTorch library for training state-of-the-art multimodal multi-task models at scale.
intelligent-machine-learning/dlrover
DLRover: An Automatic Distributed Deep Learning System
BAAI-DCAI/Bunny
A family of lightweight multimodal models.
HITsz-TMG/UMOE-Scaling-Unified-Multimodal-LLMs
The codes about "Uni-MoE: Scaling Unified Multimodal Models with Mixture of Experts"
zhuzilin/ring-flash-attention
Ring attention implementation with flash attention
UpstageAI/dataverse
The Universe of Data. All about data, data science, and data engineering
FlagOpen/FlagGems
FlagGems is an operator library for large language models implemented in Triton Language.
OpenGVLab/OmniCorpus
OmniCorpus: A Unified Multimodal Corpus of 10 Billion-Level Images Interleaved with Text
THUDM/LongAlign
[EMNLP 2024] LongAlign: A Recipe for Long Context Alignment of LLMs
YangLing0818/VideoTetris
[NeurIPS 2024] VideoTetris: Towards Compositional Text-To-Video Generation
XiaoMi/subllm
This repository is the official implementation of the ECAI 2024 conference paper SUBLLM: A Novel Efficient Architecture with Token Sequence Subsampling for LLM