Abstract
This work focuses on open-domain 4D avatarization, with the purpose of creating a 4D avatar from a portrait image in an arbitrary style. We select parametric triplanes as the intermediate 4D representation and propose a practical training paradigm that takes advantage of both generative adversarial networks (GANs) and diffusion models. Our design stems from the observation that 4D GANs excel at bridging images and triplanes without supervision yet usually face challenges in handling diverse data distributions. A robust 2D diffusion prior emerges as the solution, assisting the GAN in transferring its expertise across various domains. The synergy between these experts permits the construction of a multi-domain image-triplane dataset, which drives the development of a general 4D avatar creator. Extensive experiments suggest that our model, AvatarArtist, is capable of producing high-quality 4D avatars with strong robustness to various source image domains. The code, the data, and the models will be made publicly available to facilitate future studies..
Community
π’ AvatarArtist: Open-Domain 4D Avatarization ππ
Weβre excited to introduce AvatarArtist, a new method for open-domain 4D avatar generation β accepted to CVPR 2025! π
π Highlights:
β
Single image β 4D avatar animation
β
Works across styles: photorealistic, anime, sculpture, game characters
β
Combines 4D GANs + diffusion models to overcome data bottlenecks
β
Novel Motion-Aware Cross-Domain Renderer for precise expression & motion control
π‘ Powered by parametric triplanes and a new data construction pipeline, our method bridges 2D diffusion & 4D GANs to generate high-quality, cross-style avatars. Experiments show strong generalization in open-domain 4D avatar animation! π₯
π¨ Project: AvatarArtist Homepage
π» Code: GitHub Repository
π Paper: Arxiv
π Online Demo: Try on Hugging Face
π§ Model: Hugging Face Model
Try it out & let us know what you think! π¨β¨
#CVPR2025 #Avatar #AI #4DAvatar #GenerativeAI
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Zero-1-to-A: Zero-Shot One Image to Animatable Head Avatars Using Video Diffusion (2025)
- Controlling Avatar Diffusion with Learnable Gaussian Embedding (2025)
- HunyuanPortrait: Implicit Condition Control for Enhanced Portrait Animation (2025)
- Avat3r: Large Animatable Gaussian Reconstruction Model for High-fidelity 3D Head Avatars (2025)
- Unlock Pose Diversity: Accurate and Efficient Implicit Keypoint-based Spatiotemporal Diffusion for Audio-driven Talking Portrait (2025)
- InstaFace: Identity-Preserving Facial Editing with Single Image Inference (2025)
- High-Fidelity Relightable Monocular Portrait Animation with Lighting-Controllable Video Diffusion Model (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 1
Datasets citing this paper 0
No dataset linking this paper