zhw-zhang's Stars
hpcaitech/Open-Sora
Open-Sora: Democratizing Efficient Video Production for All
KwaiVGI/LivePortrait
Bring portraits to life!
OpenTalker/SadTalker
[CVPR 2023] SadTalker:Learning Realistic 3D Motion Coefficients for Stylized Audio-Driven Single Image Talking Face Animation
voxel51/fiftyone
Refine high-quality datasets and visual AI models
THUDM/CogVideo
text and image to video generation: CogVideoX (2024) and CogVideo (ICLR 2023)
baowenbo/DAIN
Depth-Aware Video Frame Interpolation (CVPR 2019)
LiheYoung/Depth-Anything
[CVPR 2024] Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data. Foundation Model for Monocular Depth Estimation
IDEA-Research/GroundingDINO
[ECCV 2024] Official implementation of the paper "Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection"
lxfater/inpaint-web
A free and open-source inpainting & image-upscaling tool powered by webgpu and wasm on the browser。| 基于 Webgpu 技术和 wasm 技术的免费开源 inpainting & image-upscaling 工具, 纯浏览器端实现。
sail-sg/EditAnything
Edit anything in images powered by segment-anything, ControlNet, StableDiffusion, etc. (ACM MM)
ali-vilab/VGen
Official repo for VGen: a holistic video generation ecosystem for video generation building on diffusion models
yerfor/GeneFace
GeneFace: Generalized and High-Fidelity 3D Talking Face Synthesis; ICLR 2023; Official code
gligen/GLIGEN
Open-Set Grounded Text-to-Image Generation
baaivision/Emu
Emu Series: Generative Multimodal Models from BAAI
facebookresearch/consistent_depth
We estimate dense, flicker-free, geometrically consistent depth from monocular video, for example hand-held cell phone video.
ali-vilab/dreamtalk
Official implementations for paper: DreamTalk: When Expressive Talking Head Generation Meets Diffusion Probabilistic Models
Junyi42/monst3r
Official Implementation of paper "MonST3R: A Simple Approach for Estimating Geometry in the Presence of Motion"
MyNiuuu/MOFA-Video
[ECCV 2024] MOFA-Video: Controllable Image Animation via Generative Motion Field Adaptions in Frozen Image-to-Video Diffusion Model.
sniklaus/softmax-splatting
an implementation of softmax splatting for differentiable forward warping using PyTorch
Zhen-Dong/Magic-Me
Codes for ID-Specific Video Customized Diffusion
BrokenSource/DepthFlow
🌊 Images to → 2.5D Parallax Effect Video. A Free and Open Source ImmersityAI alternative
yzhang2016/video-generation-survey
A reading list of video generation
google/dynamic-video-depth
Code for the SIGGRAPH 2021 paper "Consistent Depth of Moving Objects in Video".
aim-uofa/MovieDreamer
16lemoing/dot
Dense Optical Tracking: Connecting the Dots
yanghb22-fdu/Hi3D-Official
[MM24] Official codes and datasets for ACM MM24 paper "Hi3D: Pursuing High-Resolution Image-to-3D Generation with Video Diffusion Models".
jhaoshao/ChronoDepth
ChronoDepth: Learning Temporally Consistent Video Depth from Video Diffusion Priors
neufieldrobotics/NeuFlow_v2
hi-zhengcheng/vividzoo
mickaelseznec/flowpy
Utilities for working with optical flows in python