di-mi-ta's Stars
soumik-kanad/diff2lip
joonson/syncnet_python
Out of time: automated lip sync in the wild
willisma/SiT
Official PyTorch Implementation of "SiT: Exploring Flow and Diffusion-based Generative Models with Scalable Interpolant Transformers"
HumanAIGC/EMO
Emote Portrait Alive: Generating Expressive Portrait Videos with Audio2Video Diffusion Model under Weak Conditions
CelebV-HQ/CelebV-HQ
[ECCV 2022] CelebV-HQ: A Large-Scale Video Facial Attributes Dataset
gpu-mode/resource-stream
GPU programming related news and material links
ali-vilab/dreamtalk
Official implementations for paper: DreamTalk: When Expressive Talking Head Generation Meets Diffusion Probabilistic Models
beartype/beartype
Unbearably fast near-real-time hybrid runtime-static type-checking in pure Python.
sczhou/Upscale-A-Video
[CVPR 2024] Upscale-A-Video: Temporal-Consistent Diffusion Model for Real-World Video Super-Resolution
vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
Rudrabha/Wav2Lip
This repository contains the codes of "A Lip Sync Expert Is All You Need for Speech to Lip Generation In the Wild", published at ACM Multimedia 2020. For HD commercial model, please try out Sync Labs
guanjz20/StyleSync_PyTorch
PyTorch implementation of "StyleSync: High-Fidelity Generalized and Personalized Lip Sync in Style-based Generator"
bala1144/Imitator
TimoBolkart/BFM_to_FLAME
Convert from Basel Face Model (BFM) to the FLAME head model
Stability-AI/generative-models
Generative Models by Stability AI
VinAIResearch/PhoGPT
PhoGPT: Generative Pre-training for Vietnamese (2023)
weihaox/awesome-digital-human
A collection of resources on digital human including clothed people digitalization, virtual try-on, and other related directions.
ocornut/imgui
Dear ImGui: Bloat-free Graphical User interface for C++ with minimal dependencies
google-research/big_vision
Official codebase used to develop Vision Transformer, SigLIP, MLP-Mixer, LiT and more.
z-x-yang/Segment-and-Track-Anything
An open-source project dedicated to tracking and segmenting any objects in videos, either automatically or interactively. The primary algorithms utilized include the Segment Anything Model (SAM) for key-frame segmentation and Associating Objects with Transformers (AOT) for efficient tracking and propagation purposes.
DaddyJin/awesome-faceReenactment
papers about Face Reenactment/Talking Face Generation
Hannibal046/Awesome-LLM
Awesome-LLM: a curated list of Large Language Model
ChenyangSi/FreeU
FreeU: Free Lunch in Diffusion U-Net (CVPR2024 Oral)
facefusion/facefusion
Industry leading face manipulation platform
michaildoukas/head2head
PyTorch implementation for Head2Head and Head2Head++. It can be used to fully transfer the head pose, facial expression and eye movements from a source video to a target identity.
haofeixu/gmflow
[CVPR'22 Oral] GMFlow: Learning Optical Flow via Global Matching
guoyww/AnimateDiff
Official implementation of AnimateDiff.
williamyang1991/Rerender_A_Video
[SIGGRAPH Asia 2023] Rerender A Video: Zero-Shot Text-Guided Video-to-Video Translation
openai/tiktoken
tiktoken is a fast BPE tokeniser for use with OpenAI's models.
iejMac/video2dataset
Easily create large video dataset from video urls