Human Video Generation
Paper List
2018
- Face2Face: "Real-time Face Capture and Reenactment of RGB Videos" "CVPR" (2016) [paper][project]
- PSGAN: "Pose Guided Human Video Generation" "ECCV" (2018) [paper]
- DVP: "Deep Video Portraits" "Siggraph"(2018) [paper][project]
- Recycle-GAN: "Recycle-GAN: Unsupervised Video Retargeting" "ECCV"(2018) [paper][project][code]
- X2Face: "X2Face: A network for controlling face generation by using images, audio, and pose codes" "ECCV"(2018) [paper][project][code]
- EBDN: "Everybody Dance Now" "arXiv"(2018) [paper][project]
- Vid2Vid: "Video-to-Video Synthesis" "NIPS"(2018) [paper][project][code]
2019
- NAR: "Neural Animation and Reenactment of Human Actor Videos" "Siggraph"(2019) [paper][project]
- TETH: "Text-based Editing of Talking-head Video" "Siggraph"(2019) [paper][project]
- VPC: "Deep Video-Based Performance Cloning" "Eurographics"(2019) [paper]
- FSTH: "Few-Shot Adversarial Learning of Realistic Neural Talking Head Models" "CVPR"(2019) [paper][code unofficial]
- TNA: "Textured Neural Avatars" "CVPR"(2019) [paper][project]
- VOCA: "Voice Operated Character Animation" "CVPR"(2019) [paper][project][code]
- Audio2Face: "Audio2Face: Generating Speech/Face Animation from Single Audio with Attention-Based Bidirectional LSTM Networks" "arXiv"(2019) [paper
- RSDA: "Realistic Speech-Driven Animation with GANs" "arXiv"(2019) [paper][project][code]
- LISCG: "Learning Individual Styles of Conversational Gesture" "arXiv"(2019) [paper] [project][code]
- AUDIO2FACE: "EAUDIO2FACE: GENERATING SPEECH/FACE ANIMATION FROM SINGLE AUDIO WITH ATTENTION-BASED BIDIRECTIONAL LSTM NETWORKS" "ICMI"(2019)
- AvatarSim: "A High-Fidelity Open Embodied Avatar with Lip Syncing and Expression Capabilities" "ICMI"(2019) code
- NVP: "Neural Voice Puppetry: Audio-driven Facial Reenactment" "arXiv"(2019) [paper]
- CSGN: "Convolutional Sequence Generation for Skeleton-Based Action Synthesis" "ICCV"(2019) [paper]
- Few shot VID2VID: "Few-shot Video-to-Video Synthesis" [paper] [project] [code]
- FOM: "First Order Motion Model for Image Animation" "NIPS"(2019) [paper] [project] [code]
2020
- TransMoMo: "TransMoMo: Invariance-Driven Unsupervised Video Motion Retargeting" "CVPR"(2020) [paper] [project] [code]
- poseflow: "Deep Image Spatial Transformation for Person Image Generation" "CVPR"(2020) [paper] [project] [code]
- PIFuHD: "PIFuHD: Multi-Level Pixel-Aligned Implicit Function for High-Resolution 3D Human Digitization" "CVPR(Oral)"(2020) [paper] [project] [code]
- Hifi3dface: "High-Fidelity 3D Digital Human Creation from RGB-D Selfies" "arXiv"(2020.10) [paper][project] [code]
- face-vid2vid: "One-Shot Free-View Neural Talking-Head Synthesis for Video Conferencing" "arXiv"(2020.11) [paper] [project] [code]
- HeadGan: "HeadGAN: Video-and-Audio-Driven Talking Head Synthesis" "arXiv"(2020.12) [paper]
- "Audio-driven Talking Face Video Generation with Learning-based Personalized Head Pose" "arXiv"(2020) [paper][code]
2021
- Talking-Face_PC-AVS: "Pose-Controllable Talking Face Generation by Implicitly Modularized Audio-Visual Representation" "CVPR"(2021) [code][project][demo]
- Pixel Codec Avatar "Pixel Codec Avatars" "arXiv"(2021.04) [paper]
- MRAA "Motion Representations for Articulated Animation" "CVPR"(2021) [project]
- NWT "Towards natural audio-to-video generation with representation learning" "arXiv"(2021)[paper][project]
- LipSync3D Data-Efficient Learning of Personalized 3D Talking Faces from Video using Pose and Lighting Normalization "arXiv"(2021) [paper][demo]
Applications
Face Swap
- ZAO: a hot app.