ReflectanceFusion: Diffusion-based text to SVBRDF Generation
dc.contributor.author | Xue, Bowen | en_US |
dc.contributor.author | Guarnera, Giuseppe Claudio | en_US |
dc.contributor.author | Zhao, Shuang | en_US |
dc.contributor.author | Montazeri, Zahra | en_US |
dc.contributor.editor | Haines, Eric | en_US |
dc.contributor.editor | Garces, Elena | en_US |
dc.date.accessioned | 2024-06-25T11:05:47Z | |
dc.date.available | 2024-06-25T11:05:47Z | |
dc.date.issued | 2024 | |
dc.description.abstract | We introduce ReflectanceFusion (Reflectance Diffusion), a new neural text-to-texture model capable of generating high-fidelity SVBRDF maps from textual descriptions. Our method leverages a tandem neural approach, consisting of two modules, to accurately model the distribution of spatially varying reflectance as described by text prompts. Initially, we employ a pre-trained stable diffusion 2 model to generate a latent representation that informs the overall shape of the material and serves as our backbone model. Then, our ReflectanceUNet enables fine-tuning control over the material's physical appearance and generates SVBRDF maps. ReflectanceUNet module is trained on an extensive dataset comprising approximately 200,000 synthetic spatially varying materials. Our generative SVBRDF diffusion model allows for the synthesis of multiple SVBRDF estimates from a single textual input, offering users the possibility to choose the output that best aligns with their requirements. We illustrate our method's versatility by generating SVBRDF maps from a range of textual descriptions, both specific and broad. Our ReflectanceUNet model can integrate optional physical parameters, such as roughness and specularity, enhancing customization. When the backbone module is fixed, the ReflectanceUNet module refines the material, allowing direct edits to its physical attributes. Comparative evaluations demonstrate that ReflectanceFusion achieves better accuracy than existing text-to-material models, such as Text2Mat, while also providing the benefits of editable and relightable SVBRDF maps. | en_US |
dc.description.sectionheaders | Light and Textures | |
dc.description.seriesinformation | Eurographics Symposium on Rendering | |
dc.identifier.doi | 10.2312/sr.20241152 | |
dc.identifier.isbn | 978-3-03868-262-2 | |
dc.identifier.issn | 1727-3463 | |
dc.identifier.pages | 10 pages | |
dc.identifier.uri | https://doi.org/10.2312/sr.20241152 | |
dc.identifier.uri | https://diglib.eg.org/handle/10.2312/sr20241152 | |
dc.publisher | The Eurographics Association | en_US |
dc.rights | Attribution 4.0 International License | |
dc.rights.uri | https://creativecommons.org/licenses/by/4.0/ | |
dc.subject | CCS Concepts: Computing methodologies -> Reflectance modeling | |
dc.subject | CCS Concepts | |
dc.subject | Computing methodologies | |
dc.subject | > Reflectance modeling | |
dc.title | ReflectanceFusion: Diffusion-based text to SVBRDF Generation | en_US |