Corner selection and dual network blender for efficient view synthesis in outdoor scenes

IF 7.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Mohannad Al-Jaafari , Firas Abedi , You Yang , Qiong Liu
{"title":"Corner selection and dual network blender for efficient view synthesis in outdoor scenes","authors":"Mohannad Al-Jaafari ,&nbsp;Firas Abedi ,&nbsp;You Yang ,&nbsp;Qiong Liu","doi":"10.1016/j.patcog.2025.111668","DOIUrl":null,"url":null,"abstract":"<div><div>Novel view synthesis (NVS) from freely distributed viewpoints from large-scale outdoor scenes offers a compelling user experience in several applications, including first-person hyper-lapse videos and virtual reality. However, high-quality NVS requires dense input images, which can be affected by color discrepancies caused by extreme brightness changes, varying viewing angles, incorrect estimation of camera parameters, and mobile objects. This paper introduces Competent View Synthesis (CVS), a cost-effective approach to generating high-quality NVS from large-scale outdoor scenes to address these challenges. CVS employs a three-stage pipeline, including a Corners Selection Algorithm (CSA) to reduce the number of required input images, a Tinkering mechanism to fill in missing pixel data, and a dual-network blending (DNB) model to fuse colors and calculate attention coefficients for feature refinement. The experimental results demonstrate the effectiveness of the proposed CVS in generating realistic viewpoints from a limited number of input viewpoints. Furthermore, the comparative evaluations against two baselines using metrics such as PSNR, SSIM, and LPIPS reveal significant performance improvements of 4.6%, 0.18%, and 30.13%, respectively, over the first baseline, while the proposed method significantly outperforms the second baseline in term of perceptual metrics. By optimizing the synthesis process for complex outdoor scenes, CVS enhances the quality of generated images and improves computational efficiency.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"166 ","pages":"Article 111668"},"PeriodicalIF":7.5000,"publicationDate":"2025-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Pattern Recognition","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0031320325003280","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Novel view synthesis (NVS) from freely distributed viewpoints from large-scale outdoor scenes offers a compelling user experience in several applications, including first-person hyper-lapse videos and virtual reality. However, high-quality NVS requires dense input images, which can be affected by color discrepancies caused by extreme brightness changes, varying viewing angles, incorrect estimation of camera parameters, and mobile objects. This paper introduces Competent View Synthesis (CVS), a cost-effective approach to generating high-quality NVS from large-scale outdoor scenes to address these challenges. CVS employs a three-stage pipeline, including a Corners Selection Algorithm (CSA) to reduce the number of required input images, a Tinkering mechanism to fill in missing pixel data, and a dual-network blending (DNB) model to fuse colors and calculate attention coefficients for feature refinement. The experimental results demonstrate the effectiveness of the proposed CVS in generating realistic viewpoints from a limited number of input viewpoints. Furthermore, the comparative evaluations against two baselines using metrics such as PSNR, SSIM, and LPIPS reveal significant performance improvements of 4.6%, 0.18%, and 30.13%, respectively, over the first baseline, while the proposed method significantly outperforms the second baseline in term of perceptual metrics. By optimizing the synthesis process for complex outdoor scenes, CVS enhances the quality of generated images and improves computational efficiency.

Abstract Image

角落选择和双网络搅拌器在室外场景中有效的视图合成
从大规模户外场景中自由分布的视点进行新颖视角合成(NVS),在包括第一人称超延时视频和虚拟现实在内的多个应用中提供了引人注目的用户体验。然而,高质量的NVS需要密集的输入图像,这可能受到极端亮度变化,视角变化,相机参数估计错误以及移动物体引起的色差的影响。本文介绍了胜任视图合成(CVS),这是一种从大规模户外场景生成高质量NVS的经济有效方法,以解决这些挑战。CVS采用了一个三阶段的管道,其中包括角点选择算法(CSA)来减少所需输入图像的数量,修补机制来填充缺失的像素数据,双网络混合(DNB)模型来融合颜色并计算关注系数以进行特征细化。实验结果表明,该算法能够从有限数量的输入视点生成真实视点。此外,使用PSNR、SSIM和LPIPS等指标对两个基线进行比较评估,结果显示,与第一个基线相比,该方法的性能分别提高了4.6%、0.18%和30.13%,而在感知指标方面,该方法的性能明显优于第二个基线。通过优化复杂户外场景的合成过程,CVS增强了生成图像的质量,提高了计算效率。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Pattern Recognition
Pattern Recognition 工程技术-工程:电子与电气
CiteScore
14.40
自引率
16.20%
发文量
683
审稿时长
5.6 months
期刊介绍: The field of Pattern Recognition is both mature and rapidly evolving, playing a crucial role in various related fields such as computer vision, image processing, text analysis, and neural networks. It closely intersects with machine learning and is being applied in emerging areas like biometrics, bioinformatics, multimedia data analysis, and data science. The journal Pattern Recognition, established half a century ago during the early days of computer science, has since grown significantly in scope and influence.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信