Yiwei Hu, Paul Guerrero, Miloš Hašan, H. Rushmeier, V. Deschaintre
{"title":"Generating Procedural Materials from Text or Image Prompts","authors":"Yiwei Hu, Paul Guerrero, Miloš Hašan, H. Rushmeier, V. Deschaintre","doi":"10.1145/3588432.3591520","DOIUrl":null,"url":null,"abstract":"Node graph systems are used ubiquitously for material design in computer graphics. They allow the use of visual programming to achieve desired effects without writing code. As high-level design tools they provide convenience and flexibility, but mastering the creation of node graphs usually requires professional training. We propose an algorithm capable of generating multiple node graphs from different types of prompts, significantly lowering the bar for users to explore a specific design space. Previous work [Guerrero et al. 2022] was limited to unconditional generation of random node graphs, making the generation of an envisioned material challenging. We propose a multi-modal node graph generation neural architecture for high-quality procedural material synthesis which can be conditioned on different inputs (text or image prompts), using a CLIP-based encoder. We also create a substantially augmented material graph dataset, key to improving the generation quality. Finally, we generate high-quality graph samples using a regularized sampling process and improve the matching quality by differentiable optimization for top-ranked samples. We compare our methods to CLIP-based database search baselines (which are themselves novel) and achieve superior or similar performance without requiring massive data storage. We further show that our model can produce a set of material graphs unconditionally, conditioned on images, text prompts or partial graphs, serving as a tool for automatic visual programming completion.","PeriodicalId":280036,"journal":{"name":"ACM SIGGRAPH 2023 Conference Proceedings","volume":"23 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-04-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ACM SIGGRAPH 2023 Conference Proceedings","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3588432.3591520","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3
Abstract
Node graph systems are used ubiquitously for material design in computer graphics. They allow the use of visual programming to achieve desired effects without writing code. As high-level design tools they provide convenience and flexibility, but mastering the creation of node graphs usually requires professional training. We propose an algorithm capable of generating multiple node graphs from different types of prompts, significantly lowering the bar for users to explore a specific design space. Previous work [Guerrero et al. 2022] was limited to unconditional generation of random node graphs, making the generation of an envisioned material challenging. We propose a multi-modal node graph generation neural architecture for high-quality procedural material synthesis which can be conditioned on different inputs (text or image prompts), using a CLIP-based encoder. We also create a substantially augmented material graph dataset, key to improving the generation quality. Finally, we generate high-quality graph samples using a regularized sampling process and improve the matching quality by differentiable optimization for top-ranked samples. We compare our methods to CLIP-based database search baselines (which are themselves novel) and achieve superior or similar performance without requiring massive data storage. We further show that our model can produce a set of material graphs unconditionally, conditioned on images, text prompts or partial graphs, serving as a tool for automatic visual programming completion.
节点图系统在计算机图形学的材料设计中被广泛使用。它们允许使用可视化编程来实现所需的效果,而无需编写代码。作为高级设计工具,它们提供了便利性和灵活性,但掌握节点图的创建通常需要专业培训。我们提出了一种能够从不同类型的提示生成多个节点图的算法,大大降低了用户探索特定设计空间的门槛。先前的工作[Guerrero et al. 2022]仅限于无条件生成随机节点图,这使得生成设想的材料具有挑战性。我们提出了一个多模态节点图生成神经结构,用于高质量的程序材料合成,可以根据不同的输入(文本或图像提示),使用基于clip的编码器。我们还创建了一个实质性增强的材料图数据集,这是提高生成质量的关键。最后,我们使用正则化采样过程生成高质量的图样本,并通过对排名靠前的样本进行可微优化来提高匹配质量。我们将我们的方法与基于clip的数据库搜索基线(它们本身是新颖的)进行了比较,并在不需要大量数据存储的情况下获得了更好或类似的性能。我们进一步表明,我们的模型可以无条件地生成一组材料图,以图像、文本提示或部分图为条件,作为自动可视化编程完成的工具。