video-generation
There are 520 repositories under video-generation topic.
KwaiVGI/LivePortrait
Bring portraits to life!
zai-org/CogVideo
text and image to video generation: CogVideoX (2024) and CogVideo (ICLR 2023)
Wan-Video/Wan2.2
Wan: Open and Advanced Large-Scale Video Generative Models
Tencent-Hunyuan/HunyuanVideo
HunyuanVideo: A Systematic Framework For Large Video Generation Model
showlab/Awesome-Video-Diffusion
A curated list of recent diffusion models for video generation, editing, and various other applications.
AILab-CVC/VideoCrafter
VideoCrafter2: Overcoming Data Limitations for High-Quality Video Diffusion Models
fudan-generative-vision/champ
[ECCV 2024] Champ: Controllable and Consistent Human Image Animation with 3D Parametric Guidance
Picsart-AI-Research/Text2Video-Zero
[ICCV 2023 Oral] Text-to-Image Diffusion Models are Zero-Shot Video Generators
SandAI-org/MAGI-1
MAGI-1: Autoregressive Video Generation at Scale
ali-vilab/VACE
[ICCV 2025] Official implementations for paper: VACE: All-in-One Video Creation and Editing
OpenGVLab/InternGPT
InternGPT (iGPT) is an open source demo platform where you can easily showcase your AI models. Now it supports DragGAN, ChatGPT, ImageBind, multimodal chat like GPT-4, SAM, interactive image editing, etc. Try it at igpt.opengvlab.com (支持DragGAN、ChatGPT、ImageBind、SAM的在线Demo系统)
jy0205/Pyramid-Flow
[ICLR 2025] Pyramidal Flow Matching for Efficient Video Generative Modeling
Doubiiu/DynamiCrafter
[ECCV 2024, Oral] DynamiCrafter: Animating Open-domain Images with Video Diffusion Priors
TMElyralab/MuseV
MuseV: Infinite-length and High Fidelity Virtual Human Video Generation with Visual Conditioned Parallel Denoising
SamurAIGPT/AI-Youtube-Shorts-Generator
A python tool that uses GPT-4, FFmpeg, and OpenCV to automatically analyze videos, extract the most interesting sections, and crop them for an improved viewing experience.
thu-ml/SageAttention
[ICLR2025, ICML2025, NeurIPS2025 Spotlight] Quantized Attention achieves speedup of 2-5x compared to FlashAttention, without losing end-to-end metrics across language, image, and video models.
hao-ai-lab/FastVideo
A unified inference and post-training framework for accelerated video generation.
Tencent/MimicMotion
High-Quality Human Motion Video Generation with Confidence-aware Pose Guidance
ali-vilab/dreamtalk
Official implementations for paper: DreamTalk: When Expressive Talking Head Generation Meets Diffusion Probabilistic Models
SkyworkAI/Matrix-Game
Matrix-Game 2.0: An Open-Source, Real-Time, and Streaming Interactive World Model
KwaiVGI/ReCamMaster
[ICCV'25 Best Paper Finalist] ReCamMaster: Camera-Controlled Generative Rendering from A Single Video
menyifang/MIMO
Official implementation of "MIMO: Controllable Character Video Synthesis with Spatial Decomposed Modeling"
Phantom-video/Phantom
Phantom: Subject-Consistent Video Generation via Cross-Modal Alignment
lucidrains/video-diffusion-pytorch
Implementation of Video Diffusion Models, Jonathan Ho's new paper extending DDPMs to Video Generation - in Pytorch
mayuelala/FollowYourPose
[AAAI 2024] Follow-Your-Pose: This repo is the official implementation of "Follow-Your-Pose : Pose-Guided Text-to-Video Generation using Pose-Free Videos"
PKU-YuanGroup/MagicTime
[TPAMI 2025🔥] MagicTime: Time-lapse Video Generation Models as Metamorphic Simulators
Vchitect/VBench
[CVPR2024 Highlight] VBench - We Evaluate Video Generation
snap-research/articulated-animation
Code for Motion Representations for Articulated Animation paper
mini-sora/minisora
MiniSora: A community aims to explore the implementation path and future development direction of Sora.
Tencent-Hunyuan/HunyuanCustom
HunyuanCustom: A Multimodal-Driven Architecture for Customized Video Generation
ali-vilab/UniAnimate
Code for SCIS-2025 Paper "UniAnimate: Taming Unified Video Diffusion Models for Consistent Human Image Animation".
ali-vilab/TeaCache
Timestep Embedding Tells: It's Time to Cache for Video Diffusion Model
cure-lab/MagicDrive
[ICLR24] Official implementation of the paper “MagicDrive: Street View Generation with Diverse 3D Geometry Control”
Francis-Rings/StableAvatar
We present StableAvatar, the first end-to-end video diffusion transformer, which synthesizes infinite-length high-quality audio-driven avatar videos without any post-processing, conditioned on a reference image and audio.
wladradchenko/wunjo.wladradchenko.ru
Wunjo CE: Face Swap, Lip Sync, Control Remove Objects & Text & Background, Restyling, Audio Separator, Clone Voice, Video Generation. Open Source, Local & Free.
showlab/MotionDirector
[ECCV 2024 Oral] MotionDirector: Motion Customization of Text-to-Video Diffusion Models.