SViM3D: Stable Video Material Diffusion for Single Image 3D Generation
Abstract
A latent video diffusion model predicts multi-view consistent PBR materials from a single image, enabling relighting and novel view synthesis with high quality.
We present Stable Video Materials 3D (SViM3D), a framework to predict multi-view consistent physically based rendering (PBR) materials, given a single image. Recently, video diffusion models have been successfully used to reconstruct 3D objects from a single image efficiently. However, reflectance is still represented by simple material models or needs to be estimated in additional steps to enable relighting and controlled appearance edits. We extend a latent video diffusion model to output spatially varying PBR parameters and surface normals jointly with each generated view based on explicit camera control. This unique setup allows for relighting and generating a 3D asset using our model as neural prior. We introduce various mechanisms to this pipeline that improve quality in this ill-posed setting. We show state-of-the-art relighting and novel view synthesis performance on multiple object-centric datasets. Our method generalizes to diverse inputs, enabling the generation of relightable 3D assets useful in AR/VR, movies, games and other visual media.
Community
Stable Video Materials 3D (SviM3D) generates camera conditioned material videos from a single image.
Based on a video diffusion model the framework outputs spatially-varying physically based rendering (PBR) parameters and surface normals, jointly with each generated view. This unique setup allows for direct relighting and generating 3D assets using the material model as neural prior.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- DreamLifting: A Plug-in Module Lifting MV Diffusion Models for 3D Asset Generation (2025)
- Large Material Gaussian Model for Relightable 3D Generation (2025)
- A Scalable Attention-Based Approach for Image-to-3D Texture Mapping (2025)
- GOGS: High-Fidelity Geometry and Relighting for Glossy Objects via Gaussian Surfels (2025)
- ROGR: Relightable 3D Objects using Generative Relighting (2025)
- Chord: Chain of Rendering Decomposition for PBR Material Estimation from Generated Texture Images (2025)
- GSFix3D: Diffusion-Guided Repair of Novel Views in Gaussian Splatting (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper