lambda7xx's Stars
comfyanonymous/ComfyUI
The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.
bytedance/monolith
A Lightweight Recommendation System
NVIDIA/Cosmos
Cosmos is a world model development platform that consists of world foundation models, tokenizers and video processing pipeline to accelerate the development of Physical AI at Robotics & AV labs. Cosmos is purpose built for physical AI. The Cosmos repository will enable end users to run the Cosmos models, run inference scripts and generate videos.
MingchaoZhu/DeepLearning
Python for《Deep Learning》,该书为《深度学习》(花书) 数学推导、原理剖析与源码级别代码实现
CarperAI/trlx
A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)
dreamgaussian/dreamgaussian
[ICLR 2024 Oral] Generative Gaussian Splatting for Efficient 3D Content Creation
opendilab/LightZero
[NeurIPS 2023 Spotlight] LightZero: A Unified Benchmark for Monte Carlo Tree Search in General Sequential Decision Scenarios (awesome MCTS)
PRIME-RL/PRIME
Scalable RL solution for advanced reasoning of language models
huggingface/search-and-learn
Recipes to scale inference-time compute of open models
FoundationVision/Infinity
Infinity ∞ : Scaling Bitwise AutoRegressive Modeling for High-Resolution Image Synthesis
huggingface/picotron
Minimalistic 4D-parallelism distributed training framework for education purpose
stanford-crfm/levanter
Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax
MarioSieg/magnetron
(WIP) A small but powerful, homemade PyTorch from scratch.
WooooDyy/AgentGym
Code and implementations for the paper "AgentGym: Evolving Large Language Model-based Agents across Diverse Environments" by Zhiheng Xi et al.
alibaba/ChatLearn
A flexible and efficient training framework for large-scale alignment tasks
PKU-DAIR/Hetu
A high-performance distributed deep learning system targeting large-scale and automated distributed training.
rkinas/triton-resources
A curated list of resources for learning and exploring Triton, OpenAI's programming language for writing efficient GPU code.
hustvl/LightningDiT
[arXiv'25] Reconstruction vs. Generation: Taming Optimization Dilemma in Latent Diffusion Models
pytorch/torchft
PyTorch per step fault tolerance (actively under development)
yuandong-tian/arXiv_recbot
A Telegram bot to recommend arXiv papers
ServiceNow/AgentLab
AgentLab: An open-source framework for developing, testing, and benchmarking web agents on diverse tasks, designed for scalability and reproducibility.
modelscope/dash-infer
DashInfer is a native LLM inference engine aiming to deliver industry-leading performance atop various hardware architectures, including CUDA, x86 and ARMv9.
ezelikman/STaR
Code for STaR: Bootstrapping Reasoning With Reasoning (NeurIPS 2022)
MLSys-Learner-Resources/Awesome-MLSys-Blogger
The repository has collected a batch of noteworthy MLSys bloggers (Algorithms/Systems)
DefTruth/cuffpa-py
📚[WIP] FFPA: Yet antother Faster Flash Prefill Attention with O(1)🎉GPU SRAM complexity for headdim > 256, 1.5x~2x🎉faster vs SDPA EA.
LemonTwoL/ReNeg
ReNeg: Learning Negative Embedding with Reward Guidance
thuwzy/ZhuSuan-PyTorch
An Elegant Library for Bayesian Deep Learning in PyTorch
feifeibear/DPSKV3MFU
Estimate MFU for DeepSeekV3
thomaschlt/mla.c
Implementation from scratch in C of the Multi-head latent attention used in the Deepseek-v3 technical paper.
cylinbao/cylinbao.github.io