Demo: Real-time depth extraction and viewpoint interpolation on FPGA

Guanyu Yi, H. Yeh, G. Vanmeerbeeck, Ke Zhang, G. Lafruit
{"title":"Demo: Real-time depth extraction and viewpoint interpolation on FPGA","authors":"Guanyu Yi, H. Yeh, G. Vanmeerbeeck, Ke Zhang, G. Lafruit","doi":"10.1109/ICDSC.2011.6042943","DOIUrl":null,"url":null,"abstract":"In this demo, we demonstrate a real-time viewpoint interpolation application on FPGA. Viewpoint interpolation is the process of synthesizing plausible in-between views — so-called virtual camera views — from a couple of surrounding fixed camera views. Stereo matching is used to extract depth information, by computing a disparity map from a pair of input images. With the depth information, virtual views at any points between the two cameras are computed through view interpolation. To make viewpoint interpolation possible for low/moderate-power consumer applications, a further quality/complexity tradeoff study is required to conciliate algorithmic quality to architectural performance. In essence, the inter-dependencies between the different algorithmic steps in the processing chain are thoroughly analyzed, aiming at an overall quality-performance model that pinpoints which algorithmic functionalities can be simplified with minor (preferably no) global input-output quality degradation, while maximally reducing their implementation complexity w.r.t. arithmetic and line buffer requirements. Compared to state-of-the-art CPU and GPU platforms running at several GHz clock speed, our low-power 65 MHz FPGA implementation achieves speedups with one order of magnitude over state-of-the-art, without impeding on the visual quality, reaching over 60 frames per second high-definition (1024×768) high-quality, 64-disparity search range stereo matching and enabling viewpoint interpolation in low-power, embedded applications.","PeriodicalId":385052,"journal":{"name":"2011 Fifth ACM/IEEE International Conference on Distributed Smart Cameras","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2011-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2011 Fifth ACM/IEEE International Conference on Distributed Smart Cameras","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICDSC.2011.6042943","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

Abstract

In this demo, we demonstrate a real-time viewpoint interpolation application on FPGA. Viewpoint interpolation is the process of synthesizing plausible in-between views — so-called virtual camera views — from a couple of surrounding fixed camera views. Stereo matching is used to extract depth information, by computing a disparity map from a pair of input images. With the depth information, virtual views at any points between the two cameras are computed through view interpolation. To make viewpoint interpolation possible for low/moderate-power consumer applications, a further quality/complexity tradeoff study is required to conciliate algorithmic quality to architectural performance. In essence, the inter-dependencies between the different algorithmic steps in the processing chain are thoroughly analyzed, aiming at an overall quality-performance model that pinpoints which algorithmic functionalities can be simplified with minor (preferably no) global input-output quality degradation, while maximally reducing their implementation complexity w.r.t. arithmetic and line buffer requirements. Compared to state-of-the-art CPU and GPU platforms running at several GHz clock speed, our low-power 65 MHz FPGA implementation achieves speedups with one order of magnitude over state-of-the-art, without impeding on the visual quality, reaching over 60 frames per second high-definition (1024×768) high-quality, 64-disparity search range stereo matching and enabling viewpoint interpolation in low-power, embedded applications.
演示:基于FPGA的实时深度提取和视点插值
在这个演示中,我们演示了一个在FPGA上的实时视点插值应用。视点插值是一种从周围几个固定的摄像机视图合成可信的中间视图的过程,即所谓的虚拟摄像机视图。立体匹配通过计算一对输入图像的视差图来提取深度信息。利用深度信息,通过视图插值计算两个摄像机之间任意点的虚拟视图。为了使低/中等功率消费应用程序的视点插值成为可能,需要进行进一步的质量/复杂性权衡研究,以协调算法质量和架构性能。本质上,对处理链中不同算法步骤之间的相互依赖关系进行了彻底的分析,旨在建立一个整体质量-性能模型,该模型指出哪些算法功能可以通过轻微(最好没有)全局输入-输出质量退化来简化,同时最大限度地降低其实现复杂性。与运行在几GHz时钟速度下的最先进的CPU和GPU平台相比,我们的低功耗65 MHz FPGA实现的速度比最先进的速度快一个数量级,而不会影响视觉质量,达到每秒60帧以上的高清(1024×768)高质量,64视差搜索范围立体匹配,并支持低功耗嵌入式应用中的视点插值。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信