Furkan Mert Algan, Umut Yazgan, Driton Salihu, Cem Eteke, Eckehard Steinbach
{"title":"LEMON:利用网格优化和神经着色器进行局部编辑","authors":"Furkan Mert Algan, Umut Yazgan, Driton Salihu, Cem Eteke, Eckehard Steinbach","doi":"arxiv-2409.12024","DOIUrl":null,"url":null,"abstract":"In practical use cases, polygonal mesh editing can be faster than generating\nnew ones, but it can still be challenging and time-consuming for users.\nExisting solutions for this problem tend to focus on a single task, either\ngeometry or novel view synthesis, which often leads to disjointed results\nbetween the mesh and view. In this work, we propose LEMON, a mesh editing\npipeline that combines neural deferred shading with localized mesh\noptimization. Our approach begins by identifying the most important vertices in\nthe mesh for editing, utilizing a segmentation model to focus on these key\nregions. Given multi-view images of an object, we optimize a neural shader and\na polygonal mesh while extracting the normal map and the rendered image from\neach view. By using these outputs as conditioning data, we edit the input\nimages with a text-to-image diffusion model and iteratively update our dataset\nwhile deforming the mesh. This process results in a polygonal mesh that is\nedited according to the given text instruction, preserving the geometric\ncharacteristics of the initial mesh while focusing on the most significant\nareas. We evaluate our pipeline using the DTU dataset, demonstrating that it\ngenerates finely-edited meshes more rapidly than the current state-of-the-art\nmethods. We include our code and additional results in the supplementary\nmaterial.","PeriodicalId":501130,"journal":{"name":"arXiv - CS - Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"LEMON: Localized Editing with Mesh Optimization and Neural Shaders\",\"authors\":\"Furkan Mert Algan, Umut Yazgan, Driton Salihu, Cem Eteke, Eckehard Steinbach\",\"doi\":\"arxiv-2409.12024\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In practical use cases, polygonal mesh editing can be faster than generating\\nnew ones, but it can still be challenging and time-consuming for users.\\nExisting solutions for this problem tend to focus on a single task, either\\ngeometry or novel view synthesis, which often leads to disjointed results\\nbetween the mesh and view. In this work, we propose LEMON, a mesh editing\\npipeline that combines neural deferred shading with localized mesh\\noptimization. Our approach begins by identifying the most important vertices in\\nthe mesh for editing, utilizing a segmentation model to focus on these key\\nregions. Given multi-view images of an object, we optimize a neural shader and\\na polygonal mesh while extracting the normal map and the rendered image from\\neach view. By using these outputs as conditioning data, we edit the input\\nimages with a text-to-image diffusion model and iteratively update our dataset\\nwhile deforming the mesh. This process results in a polygonal mesh that is\\nedited according to the given text instruction, preserving the geometric\\ncharacteristics of the initial mesh while focusing on the most significant\\nareas. We evaluate our pipeline using the DTU dataset, demonstrating that it\\ngenerates finely-edited meshes more rapidly than the current state-of-the-art\\nmethods. We include our code and additional results in the supplementary\\nmaterial.\",\"PeriodicalId\":501130,\"journal\":{\"name\":\"arXiv - CS - Computer Vision and Pattern Recognition\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-18\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Computer Vision and Pattern Recognition\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.12024\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Computer Vision and Pattern Recognition","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.12024","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
LEMON: Localized Editing with Mesh Optimization and Neural Shaders
In practical use cases, polygonal mesh editing can be faster than generating
new ones, but it can still be challenging and time-consuming for users.
Existing solutions for this problem tend to focus on a single task, either
geometry or novel view synthesis, which often leads to disjointed results
between the mesh and view. In this work, we propose LEMON, a mesh editing
pipeline that combines neural deferred shading with localized mesh
optimization. Our approach begins by identifying the most important vertices in
the mesh for editing, utilizing a segmentation model to focus on these key
regions. Given multi-view images of an object, we optimize a neural shader and
a polygonal mesh while extracting the normal map and the rendered image from
each view. By using these outputs as conditioning data, we edit the input
images with a text-to-image diffusion model and iteratively update our dataset
while deforming the mesh. This process results in a polygonal mesh that is
edited according to the given text instruction, preserving the geometric
characteristics of the initial mesh while focusing on the most significant
areas. We evaluate our pipeline using the DTU dataset, demonstrating that it
generates finely-edited meshes more rapidly than the current state-of-the-art
methods. We include our code and additional results in the supplementary
material.