Human Video Generation

Paper List

2018

  • Face2Face: "Real-time Face Capture and Reenactment of RGB Videos" "CVPR" (2016) [paper][project]
  • PSGAN: "Pose Guided Human Video Generation" "ECCV" (2018) [paper]
  • DVP: "Deep Video Portraits" "Siggraph"(2018) [paper][project]
  • Recycle-GAN: "Recycle-GAN: Unsupervised Video Retargeting" "ECCV"(2018) [paper][project][code]
  • X2Face: "X2Face: A network for controlling face generation by using images, audio, and pose codes" "ECCV"(2018) [paper][project][code]
  • EBDN: "Everybody Dance Now" "arXiv"(2018) [paper][project]
  • Vid2Vid: "Video-to-Video Synthesis" "NIPS"(2018) [paper][project][code]

2019

  • NAR: "Neural Animation and Reenactment of Human Actor Videos" "Siggraph"(2019) [paper][project]
  • TETH: "Text-based Editing of Talking-head Video" "Siggraph"(2019) [paper][project]
  • VPC: "Deep Video-Based Performance Cloning" "Eurographics"(2019) [paper]
  • FSTH: "Few-Shot Adversarial Learning of Realistic Neural Talking Head Models" "CVPR"(2019) [paper][code unofficial]
  • TNA: "Textured Neural Avatars" "CVPR"(2019) [paper][project]
  • VOCA: "Voice Operated Character Animation" "CVPR"(2019) [paper][project][code]
  • Audio2Face: "Audio2Face: Generating Speech/Face Animation from Single Audio with Attention-Based Bidirectional LSTM Networks" "arXiv"(2019) [paper
  • RSDA: "Realistic Speech-Driven Animation with GANs" "arXiv"(2019) [paper][project][code]
  • LISCG: "Learning Individual Styles of Conversational Gesture" "arXiv"(2019) [paper] [project][code]
  • AUDIO2FACE: "EAUDIO2FACE: GENERATING SPEECH/FACE ANIMATION FROM SINGLE AUDIO WITH ATTENTION-BASED BIDIRECTIONAL LSTM NETWORKS" "ICMI"(2019)
  • AvatarSim: "A High-Fidelity Open Embodied Avatar with Lip Syncing and Expression Capabilities" "ICMI"(2019) code
  • NVP: "Neural Voice Puppetry: Audio-driven Facial Reenactment" "arXiv"(2019) [paper]
  • CSGN: "Convolutional Sequence Generation for Skeleton-Based Action Synthesis" "ICCV"(2019) [paper]
  • Few shot VID2VID: "Few-shot Video-to-Video Synthesis" [paper] [project] [code]
  • FOM: "First Order Motion Model for Image Animation" "NIPS"(2019) [paper] [project] [code]

2020

  • TransMoMo: "TransMoMo: Invariance-Driven Unsupervised Video Motion Retargeting" "CVPR"(2020) [paper] [project] [code]
  • poseflow: "Deep Image Spatial Transformation for Person Image Generation" "CVPR"(2020) [paper] [project] [code]
  • PIFuHD: "PIFuHD: Multi-Level Pixel-Aligned Implicit Function for High-Resolution 3D Human Digitization" "CVPR(Oral)"(2020) [paper] [project] [code]
  • Hifi3dface: "High-Fidelity 3D Digital Human Creation from RGB-D Selfies" "arXiv"(2020.10) [paper][project] [code]
  • face-vid2vid: "One-Shot Free-View Neural Talking-Head Synthesis for Video Conferencing" "arXiv"(2020.11) [paper] [project] [code]
  • HeadGan: "HeadGAN: Video-and-Audio-Driven Talking Head Synthesis" "arXiv"(2020.12) [paper]

Applications

Face Swap

  • ZAO: a hot app.

Video generated based on ZAO

AI Host:

Video generated by SouGou

Dataset

Researchers & Teams

  1. Graphics, Vision & Video at MPII
  2. REAL VIRTUAL HUMANS at MPII
  3. Visual Computing Group at TUM
  4. Perceiving Systems Department at MPII