Stars
[ECCV 2024] MOFA-Video: Controllable Image Animation via Generative Motion Field Adaptions in Frozen Image-to-Video Diffusion Model.
PixArt-α: Fast Training of Diffusion Transformer for Photorealistic Text-to-Image Synthesis
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
[CVPR2024 Highlight] VBench - We Evaluate Video Generation
Latte: Latent Diffusion Transformer for Video Generation.
[CVPR 2024] This is the official source for our paper "SyncTalk: The Devil is in the Synchronization for Talking Head Synthesis"
MusePose: a Pose-Driven Image-to-Video Framework for Virtual Human Generation
Code for Paper "UniAnimate: Taming Unified Video Diffusion Models for Consistent Human Image Animation".
⏰ Collaboratively track deadlines of conferences recommended by CCF (Website, Python Cli, Wechat Applet) / If you find it useful, please star this project, thanks~
V-Express aims to generate a talking head video under the control of a reference image, an audio, and a sequence of V-Kps images.
Character Animation (AnimateAnyone, Face Reenactment)
[CVPR 2024] Make-Your-Anchor: A Diffusion-based 2D Avatar Generation Framework.
Real3D-Portrait: One-shot Realistic 3D Talking Portrait Synthesis; ICLR 2024 Spotlight; Official code
Pytorch implementation of Learning Disentangled Representations via Mutual Information Estimation (ECCV 2020)
AniPortrait: Audio-Driven Synthesis of Photorealistic Portrait Animation
Latex-format paper templates, including Elsevier, arXiv and IEEE Access.
📖 A curated list of resources dedicated to talking face.
[GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simple, user-friendly …
Champ: Controllable and Consistent Human Image Animation with 3D Parametric Guidance
A PyTorch implementation of the paper "ZigMa: A DiT-Style Mamba-based Diffusion Model" (ECCV 2024)