controllable-generation
There are 59 repositories under controllable-generation topic.
yang-song/score_sde_pytorch
PyTorch implementation for Score-Based Generative Modeling through Stochastic Differential Equations (ICLR 2021, Oral)
ttengwang/Caption-Anything
Caption-Anything is a versatile tool combining image segmentation, visual captioning, and ChatGPT, generating tailored captions with diverse controls for user preferences. https://huggingface.co/spaces/TencentARC/Caption-Anything https://huggingface.co/spaces/VIPLab/Caption-Anything
yang-song/score_sde
Official code for Score-Based Generative Modeling through Stochastic Differential Equations (ICLR 2021, Oral)
ermongroup/SDEdit
PyTorch implementation for SDEdit: Image Synthesis and Editing with Stochastic Differential Equations
PRIV-Creation/Awesome-Controllable-T2I-Diffusion-Models
A collection of resources on controllable generation with text-to-image diffusion models.
MyNiuuu/MOFA-Video
[ECCV 2024] MOFA-Video: Controllable Image Animation via Generative Motion Field Adaptions in Frozen Image-to-Video Diffusion Model.
SkyworkAI/SkyReels-A2
SkyReels-A2: Compose anything in video diffusion transformers
mayuelala/Awesome-Controllable-Video-Generation
[ArXiv 2025] A survey about controllable video generation: This repo is the official awesome of "Controllable video generation: A survey"
open-mmlab/StyleShot
StyleShot: A SnapShot on Any Style. 一款可以迁移任意风格到任意内容的模型,无需针对图片微调,即能生成高质量的个性风格化图片!
kpandey008/DiffuseVAE
Official implementation of "DiffuseVAE: Efficient, Controllable and High-Fidelity Generation from Low-Dimensional Latents"
haofanwang/awesome-conditional-content-generation
Update-to-data resources for conditional content generation, including human motion generation, image or video generation and editing.
hancyran/LiDAR-Diffusion
[CVPR 2024] Official implementation of "Towards Realistic Scene Generation with LiDAR Diffusion Models"
xyfJASON/ctrlora
[ICLR 2025] Codebase for "CtrLoRA: An Extensible and Efficient Framework for Controllable Image Generation"
WU-CVGL/MVControl
[3DV-2025] Official implementation of "Controllable Text-to-3D Generation via Surface-Aligned Gaussian Splatting"
KangweiiLiu/Awesome_Audio-driven_Talking-Face-Generation
A curated list of resources of audio-driven talking face generation
samxuxiang/hnc-cad
[ICML 2023] Official PyTorch Implementation of "Hierarchical Neural Coding for Controllable CAD Model Generation".
open-mmlab/AnyControl
[ECCV 2024] AnyControl, a multi-control image synthesis model that supports any combination of user provided control signals. 一个支持用户自由输入控制信号的图像生成模型,能够根据多种控制生成自然和谐的结果!
GGGHSL/InfoSwap-master
Official PyTorch Implementation for InfoSwap
wangqiang9/Awesome-Controllable-Video-Diffusion
Awesome Controllable Video Generation with Diffusion Models
inFaaa/Awesome-Personalized-Video-Creation
📖 This is a repository for organizing papers, codes, and other resources related to personalized video generation and editing.
ImKeTT/CTG-latentAEs
[Paperlist] Awesome paper list of controllable text generation via latent auto-encoders. Contributions of any kind are welcome.
LabShuHangGU/PerLDiff
ICCV 2025-PerLDiff: Controllable Street View Synthesis Using Perspective-Layout Diffusion Models
jxzhangjhu/awesome-LLM-controlled-decoding-generation
awesome-LLM-controlled-constrained-generation
ChocoWu/Any2Caption
This is the project for 'Any2Caption', Interpreting Any Condition to Caption for Controllable Video Generation
VSAnimator/collage-diffusion
Implementation of Collage Diffusion (https://arxiv.org/abs/2303.00262)
ImKeTT/AdaVAE
[Preprint] AdaVAE: Exploring Adaptive GPT-2s in VAEs for Language Modeling PyTorch Implementation
Tele-AI/OmniVDiff
Omni Controllable Video Diffusion
ImKeTT/PCAE
[KBS] PCAE: A Framework of Plug-in Conditional Auto-Encoder for Controllable Text Generation PyTorch Implementation
Chernobyllight/SaMST
ACCV 2024
bladewaltz1/ModeCap
Controllable mage captioning model with unsupervised modes
yuanze-lin/IllumiCraft
The official code for "IllumiCraft: Unified Geometry and Illumination Diffusion for Controllable Video Generation"
Justin900429/autonomous_driving_with_diffusion_model
Using diffusion model to reach controllable end-to-end driving with Carla simulation environment.
Robin-WZQ/CCLAP
[ICME'23, oral] CCLAP: Controllable Chinese Landscape Painting Generation
ha-lins/DTG-SI
Code for "Data-to-text Generation with Style Imitation." [Findings of EMNLP 2020]
faverogian/controlNet
An implementation of ControlNet as described in "Adding Conditional Control to Text-to-Image Diffusion Models" published by Zhang et al.