18445864529's Stars
huggingface/diffusers
🤗 Diffusers: State-of-the-art diffusion models for image, video, and audio generation in PyTorch and FLAX.
Stability-AI/generative-models
Generative Models by Stability AI
guoyww/AnimateDiff
Official implementation of AnimateDiff.
salesforce/LAVIS
LAVIS - A One-stop Library for Language-Vision Intelligence
showlab/Tune-A-Video
[ICCV 2023] Tune-A-Video: One-Shot Tuning of Image Diffusion Models for Text-to-Video Generation
Picsart-AI-Research/Text2Video-Zero
[ICCV 2023 Oral] Text-to-Image Diffusion Models are Zero-Shot Video Generators
princeton-vl/RAFT
ali-vilab/VGen
Official repo for VGen: a holistic video generation ecosystem for video generation building on diffusion models
Doubiiu/DynamiCrafter
[ECCV 2024, Oral] DynamiCrafter: Animating Open-domain Images with Video Diffusion Priors
InternLM/InternLM-XComposer
InternLM-XComposer2.5-OmniLive: A Comprehensive Multimodal System for Long-term Streaming Video and Audio Interactions
ChenHsing/Awesome-Video-Diffusion-Models
[CSUR] A Survey on Video Diffusion Models
Picsart-AI-Research/StreamingT2V
StreamingT2V: Consistent, Dynamic, and Extendable Long Video Generation from Text
Yujun-Shi/DragDiffusion
[CVPR2024, Highlight] Official code for DragDiffusion
VainF/pytorch-msssim
Fast and differentiable MS-SSIM and SSIM for pytorch.
Vchitect/SEINE
[ICLR 2024] SEINE: Short-to-Long Video Diffusion Model for Generative Transition and Prediction
ali-vilab/videocomposer
Official repo for VideoComposer: Compositional Video Synthesis with Motion Controllability
Vchitect/LaVie
[IJCV 2024] LaVie: High-Quality Video Generation with Cascaded Latent Diffusion Models
mayuelala/FollowYourClick
[AAAI 2025] Follow-Your-Click: This repo is the official implementation of "Follow-Your-Click: Open-domain Regional Image Animation via Short Prompts"
DirtyHarryLYL/LLM-in-Vision
Recent LLM-based CV and related works. Welcome to comment/contribute!
alibaba/animate-anything
Fine-Grained Open Domain Image Animation with Motion Guidance
willisma/SiT
Official PyTorch Implementation of "SiT: Exploring Flow and Diffusion-based Generative Models with Scalable Interpolant Transformers"
ExponentialML/Text-To-Video-Finetuning
Finetune ModelScope's Text To Video model using Diffusers 🧨
TonyLianLong/LLM-groundedDiffusion
LLM-grounded Diffusion: Enhancing Prompt Understanding of Text-to-Image Diffusion Models with Large Language Models (LLM-grounded Diffusion: LMD, TMLR 2024)
RQ-Wu/LAMP
[CVPR 2024] | LAMP: Learn a Motion Pattern for Few-Shot Based Video Generation
TIGER-AI-Lab/ConsistI2V
ConsistI2V: Enhancing Visual Consistency for Image-to-Video Generation (TMLR 2024)
sutdcv/Animal-Kingdom
[CVPR2022] Animal Kingdom: A Large and Diverse Dataset for Animal Behavior Understanding
microsoft/ReCo
ReCo: Region-Controlled Text-to-Image Generation, CVPR 2023
JianhongBai/UniEdit
UniEdit: A Unified Tuning-Free Framework for Video Motion and Appearance Editing
SooLab/Free-Bloom
[NeurIPS 2023] Free-Bloom: Zero-Shot Text-to-Video Generator with LLM Director and LDM Animator
kylehkhsu/latent_quantization