Weiyue Wang, Qiangui Huang, Suya You, Chao Yang, U. Neumann
{"title":"Shape Inpainting Using 3D Generative Adversarial Network and Recurrent Convolutional Networks","authors":"Weiyue Wang, Qiangui Huang, Suya You, Chao Yang, U. Neumann","doi":"10.1109/ICCV.2017.252","DOIUrl":null,"url":null,"abstract":"Recent advances in convolutional neural networks have shown promising results in 3D shape completion. But due to GPU memory limitations, these methods can only produce low-resolution outputs. To inpaint 3D models with semantic plausibility and contextual details, we introduce a hybrid framework that combines a 3D Encoder-Decoder Generative Adversarial Network (3D-ED-GAN) and a Longterm Recurrent Convolutional Network (LRCN). The 3DED- GAN is a 3D convolutional neural network trained with a generative adversarial paradigm to fill missing 3D data in low-resolution. LRCN adopts a recurrent neural network architecture to minimize GPU memory usage and incorporates an Encoder-Decoder pair into a Long Shortterm Memory Network. By handling the 3D model as a sequence of 2D slices, LRCN transforms a coarse 3D shape into a more complete and higher resolution volume. While 3D-ED-GAN captures global contextual structure of the 3D shape, LRCN localizes the fine-grained details. Experimental results on both real-world and synthetic data show reconstructions from corrupted models result in complete and high-resolution 3D objects.","PeriodicalId":6559,"journal":{"name":"2017 IEEE International Conference on Computer Vision (ICCV)","volume":"46 1","pages":"2317-2325"},"PeriodicalIF":0.0000,"publicationDate":"2017-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"150","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2017 IEEE International Conference on Computer Vision (ICCV)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCV.2017.252","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 150
Abstract
Recent advances in convolutional neural networks have shown promising results in 3D shape completion. But due to GPU memory limitations, these methods can only produce low-resolution outputs. To inpaint 3D models with semantic plausibility and contextual details, we introduce a hybrid framework that combines a 3D Encoder-Decoder Generative Adversarial Network (3D-ED-GAN) and a Longterm Recurrent Convolutional Network (LRCN). The 3DED- GAN is a 3D convolutional neural network trained with a generative adversarial paradigm to fill missing 3D data in low-resolution. LRCN adopts a recurrent neural network architecture to minimize GPU memory usage and incorporates an Encoder-Decoder pair into a Long Shortterm Memory Network. By handling the 3D model as a sequence of 2D slices, LRCN transforms a coarse 3D shape into a more complete and higher resolution volume. While 3D-ED-GAN captures global contextual structure of the 3D shape, LRCN localizes the fine-grained details. Experimental results on both real-world and synthetic data show reconstructions from corrupted models result in complete and high-resolution 3D objects.
卷积神经网络的最新进展在三维形状补全方面显示出有希望的结果。但是由于GPU内存的限制,这些方法只能产生低分辨率输出。为了绘制具有语义合理性和上下文细节的3D模型,我们引入了一个混合框架,该框架结合了3D编码器-解码器生成对抗网络(3D- ed - gan)和长期循环卷积网络(LRCN)。3DED- GAN是一种用生成对抗范式训练的3D卷积神经网络,用于填补低分辨率缺失的3D数据。LRCN采用循环神经网络架构,最大限度地减少GPU内存使用,并将编码器-解码器对集成到长短期记忆网络中。LRCN通过将3D模型处理为一系列2D切片,将粗糙的3D形状转换为更完整、更高分辨率的体积。3D- ed - gan捕获3D形状的全局上下文结构,而LRCN则定位细粒度细节。真实世界和合成数据的实验结果表明,损坏模型的重建结果是完整和高分辨率的3D物体。