wswcfan's Stars
Dao-AILab/flash-attention
Fast and memory-efficient exact attention
togethercomputer/RedPajama-Data
The RedPajama-Data repository contains code for preparing large datasets for training large language models.
deepspeedai/DeepSpeedExamples
Example models using DeepSpeed
AntonOsika/gpt-engineer
CLI platform to experiment with codegen. Precursor to: https://lovable.dev
yoheinakajima/babyagi
tensorchord/envd
🏕️ Reproducible development environment
RUCAIBox/LLMSurvey
The official GitHub page for the survey paper "A Survey of Large Language Models".
Mooler0410/LLMsPracticalGuide
A curated list of practical guide resources of LLMs (LLMs Tree, Examples, Papers)
kubernetes-sigs/node-feature-discovery
Node feature discovery for Kubernetes
OpenNLPLab/TransnormerLLM
Official implementation of TransNormerLLM: A Faster and Better LLM
bigcode-project/starcoder
Home of StarCoder: fine-tuning & inference!
InternLM/InternLM
Official release of InternLM series (InternLM, InternLM2, InternLM2.5, InternLM3).
Facico/Chinese-Vicuna
Chinese-Vicuna: A Chinese Instruction-following LLaMA-based Model —— 一个中文低资源的llama+lora方案,结构参考alpaca
openlm-research/open_llama
OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset
roclark/NeMo-Megatron-Launcher
NeMo Megatron launcher and tools
NVIDIA/nccl
Optimized primitives for collective multi-GPU communication
kubeflow/trainer
Distributed ML Training and Fine-Tuning on Kubernetes
ggml-org/llama.cpp
LLM inference in C/C++
lm-sys/FastChat
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
sylabs/wlm-operator
Singularity implementation of k8s operator for interacting with SLURM.
IBM/Bridge-Operator
Bridge operator repo
SchedMD/slurm
Slurm: A Highly Scalable Workload Manager
Vision-CAIR/MiniGPT-4
Open-sourced codes for MiniGPT-4 and MiniGPT-v2 (https://minigpt-4.github.io, https://minigpt-v2.github.io/)
deepspeedai/Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
alpa-projects/alpa
Training and serving large-scale neural networks with auto parallelization.
flyteorg/flyte
Scalable and flexible workflow orchestration platform that seamlessly unifies data, ML and analytics stacks.
huggingface/transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
NVIDIA/NeMo
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
NVIDIA/DeepLearningExamples
State-of-the-Art Deep Learning scripts organized by models - easy to train and deploy with reproducible accuracy and performance on enterprise-grade infrastructure.
onnx/onnx
Open standard for machine learning interoperability