/awesome-4d-generation

List of papers on 4D Generation.

MIT LicenseMIT

Awesome 4D Generation

This repo collects papers for 4D generation.

Table of Contents

Camera Control for Video Diffusion

VD3D: Taming Large Video Diffusion Transformers for 3D Camera Control

📄 Paper | 🌐 Project Page

Controlling Space and Time with Diffusion Models

📄 Paper | 🌐 Project Page

CamCo: Camera-Controllable 3D-Consistent Image-to-Video Generation

📄 Paper | 🌐 Project Page

Collaborative Video Diffusion: Consistent Multi-video Generation with Camera Control

📄 Paper | 🌐 Project Page

Multi-view for Video Diffusion

SV4D: Dynamic 3D Content Generation with Multi-Frame and Multi-View Consistency, Xie et al., Arxiv 2024

📄 Paper | 🌐 Project Page | 💻 Code

L4GM: Large 4D Gaussian Reconstruction Model, Ren et al., Arxiv 2024

📄 Paper | 🌐 Project Page

4Diffusion: Multi-view Video Diffusion Model for 4D Generation, Zhang et al., Arxiv 2024

📄 Paper | 🌐 Project Page | 💻 Code

Diffusion4D: Fast Spatial-temporal Consistent 4D Generation via Video Diffusion Models, Liang et al., Arxiv 2024

📄 Paper | 🌐 Project Page | 💻 Code | 🎥 Video

Distillation from Video Diffusion

Streetscapes: Large-scale Consistent Street View Generation Using Autoregressive Video Diffusion, Deng et al., SIGGRAPH 2024

📄 Paper | 🌐 Project Page | 🎥 Video

4Dynamic: Text-to-4D Generation with Hybrid Priors, Yuan et al., Arxiv 2024

📄 Paper

STAR: Skeleton-aware Text-based 4D Avatar Generation with In-Network Motion Retargeting, Chai et al., Arxiv 2024

📄 Paper | 🌐 Project Page | 💻 Code

MotionDreamer: Zero-Shot 3D Mesh Animation from Video Diffusion Models, Uzolas et al., Arxiv 2024

📄 Paper | 💻 Code

PLA4D: Pixel-Level Alignments for Text-to-4D Gaussian Splatting, Miao et al., Arxiv 2024

📄 Paper | 🌐 Project Page

MagicPose4D: Crafting Articulated Models with Appearance and Motion Control, Zhang et al., Arxiv 2024

📄 Paper | 🌐 Project Page | 💻 Code

SC4D: Sparse-Controlled Video-to-4D Generation and Motion Transfer, Wu et al., Arxiv 2024

📄 Paper | 🌐 Project Page | 💻 Code | 🎥 Video

TC4D: Trajectory-Conditioned Text-to-4D Generation, Bahmani et al., Arxiv 2024

📄 Paper | 🌐 Project Page | 💻 Code

Comp4D: LLM-Guided Compositional 4D Scene Generation, Xu et al., Arxiv 2024

📄 Paper | 🌐 Project Page | 💻 Code | 🎥 Video

STAG4D: Spatial-Temporal Anchored Generative 4D Gaussians, Zetn et al., Arxiv 2024

📄 Paper | 🌐 Project Page | 💻 Code

GaussianFlow: Splatting Gaussian Dynamics for 4D Content Creation, Gao et al., Arxiv 2024

📄 Paper | 🌐 Project Page | 💻 Code

4DGen: Grounded 4D Content Generation with Spatial-temporal Consistency, Yin et al., Arxiv 2023

📄 Paper | 🌐 Project Page | 💻 Code | 🎥 Video

DreamGaussian4D: Generative 4D Gaussian Splatting, Ren et al., CVPR 2024

📄 Paper | 🌐 Project Page

Align Your Gaussians: Text-to-4D with Dynamic 3D Gaussians and Composed Diffusion Models, Ling et al., Arxiv 2023

📄 Paper | 🌐 Project Page | 💻 Code

AnimatableDreamer: Text-Guided Non-rigid 3D Model Generation and Reconstruction with Canonical Score Distillation, Wang et al., Arxiv 2023

📄 Paper | 🌐 Project Page | 💻 Code

4D-fy: Text-to-4D Generation Using Hybrid Score Distillation Sampling, Bahmani et al., CVPR 2024

📄 Paper | 🌐 Project Page | 💻 Code

A Unified Approach for Text- and Image-guided 4D Scene Generation, Zheng et al., CVPR 2024

📄 Paper | 🌐 Project Page | 💻 Code

Animate124: Animating One Image to 4D Dynamic Scene, Zhao et al., Arxiv 2023

📄 Paper | 🌐 Project Page | 💻 Code

Consistent4D: Consistent 360° Dynamic Object Generation from Monocular Video, Jiang et al., Arxiv 2024

📄 Paper | 🌐 Project Page | 💻 Code

Text-To-4D Dynamic Scene Generation, Singer et al., Arxiv 2023

📄 Paper | 🌐 Project Page

Generation by Reconstruction

4Real: Towards Photorealistic 4D Scene Generation via Video Diffusion Models, Yu et al., Arxiv 2024

📄 Paper | 🌐 Project Page

Vidu4D: Single Generated Video to High-Fidelity 4D Reconstruction with Dynamic Gaussian Surfels, Wang et al., Arxiv 2024

📄 Paper | 🌐 Project Page

4D Editing

Instruct 4D-to-4D: Editing 4D Scenes as Pseudo-3D Scenes Using 2D Diffusion, Mou et al., CVPR 2024

📄 Paper | 🌐 Project Page

Physics

Sync4D: Video Guided Controllable Dynamics for Physics-Based 4D Generation, Fu et al., Arxiv 2024

📄 Paper | 🌐 Project Page