WuTao-CS's Stars
Stability-AI/stablediffusion
High-Resolution Image Synthesis with Latent Diffusion Models
IDEA-Research/Grounded-Segment-Anything
Grounded SAM: Marrying Grounding DINO with Segment Anything & Stable Diffusion & Recognize Anything - Automatically Detect , Segment and Generate Anything
HumanAIGC/AnimateAnyone
Animate Anyone: Consistent and Controllable Image-to-Video Synthesis for Character Animation
PKU-YuanGroup/Open-Sora-Plan
This project aim to reproduce Sora (Open AI T2V model), we wish the open source community contribute to this project.
cloneofsimo/lora
Using Low-rank adaptation to quickly fine-tune diffusion models.
pengsida/learning_research
本人的科研经验
tencent-ailab/IP-Adapter
The image prompt adapter is designed to enable a pretrained text-to-image diffusion model to generate images with image prompt.
ahmetbersoz/chatgpt-prompts-for-academic-writing
This list of writing prompts covers a range of topics and tasks, including brainstorming research ideas, improving language and style, conducting literature reviews, and developing research plans.
ali-vilab/VGen
Official repo for VGen: a holistic video generation ecosystem for video generation building on diffusion models
Doubiiu/DynamiCrafter
[ECCV 2024, Oral] DynamiCrafter: Animating Open-domain Images with Video Diffusion Priors
InternLM/InternLM-XComposer
InternLM-XComposer2.5-OmniLive: A Comprehensive Multimodal System for Long-term Streaming Video and Audio Interactions
ChenHsing/Awesome-Video-Diffusion-Models
[CSUR] A Survey on Video Diffusion Models
baaivision/Emu
Emu Series: Generative Multimodal Models from BAAI
TencentARC/MotionCtrl
Official Code for MotionCtrl [SIGGRAPH 2024]
ali-vilab/videocomposer
Official repo for VideoComposer: Compositional Video Synthesis with Motion Controllability
Vchitect/LaVie
[IJCV 2024] LaVie: High-Quality Video Generation with Cascaded Latent Diffusion Models
ExponentialML/Text-To-Video-Finetuning
Finetune ModelScope's Text To Video model using Diffusers 🧨
m-bain/webvid
Large-scale text-video dataset. 10 million captioned short videos.
Delppine1024/TGreen
Some files work well on T v1.1 (The latest support v1.8.10/1.8.3-dev), Powered by TC
ali-vilab/Cones-V2
[NeurIPS 2023] Official implementations for paper: Customizable Image Synthesis with Multiple Subjects
YingqingHe/Awesome-LLMs-meet-Multimodal-Generation
🔥🔥🔥 A curated list of papers on LLMs-based multimodal generation (image, video, 3D and audio).
ID-Animator/ID-Animator
atfortes/Awesome-Controllable-Generation
Papers and resources on Controllable Generation using Diffusion Models, including ControlNet, DreamBooth, T2I-Adapter, IP-Adapter.
OPPO-Mente-Lab/Subject-Diffusion
Subject-Diffusion:Open Domain Personalized Text-to-Image Generation without Test-time Fine-tuning
Akaneqwq/360DVD
[CVPR2024] 360DVD: Controllable Panorama Video Generation with 360-Degree Video Diffusion Model
NPU-YanChi/diff-gaussian-rasterization-for-gsslam
Zehong-Ma/OVMR
OVMR: Open-Vocabulary Recognition with Multi-Modal References (CVPR24)
Whalesong-zrs/Towards-Fine-grained-HBOE
The code for Fine-grained HBOE | AAAI 2024 (official version and optimized version).
neurowelt/AnimateDiff
Official implementation of AnimateDiff.
ryanpo/custom-diffusion-lora
Custom Diffusion: Multi-Concept Customization of Text-to-Image Diffusion (CVPR 2023)