3-D motion and structure estimation for arbitrary scenes from 2-D optical and sonar video

S. Negahdaripour, A. Taatian
{"title":"3-D motion and structure estimation for arbitrary scenes from 2-D optical and sonar video","authors":"S. Negahdaripour, A. Taatian","doi":"10.1109/OCEANS.2008.5151985","DOIUrl":null,"url":null,"abstract":"Imaging system are routinely deployed for underwater search, inspection and scientific surveys of manmade and natural structures, etc. Optical cameras, while providing high resolution and target details, have range limitation according to water visibility and become ineffective in turbid environments. In comparison, high-frequency (MHz) 2-D imaging sonar video systems, introduced to the commercial market in recent years, image targets at distances of 10's of meters in highly turbid waters. Visibility permitting, the integration of visual cues in 2-D optical and sonar data would enable better performance compared to deploying either imaging system alone. We address the problem of motion estimation- e.g., for vision-based navigation and target-based positioning of a mobile submersible platform- from 2-D optical and sonar images. The application of structure from motion paradigm in this multimodal imaging scenario also enables the 3-D reconstruction of scene features. We rely on the tracking of features in the sonar and optical motion sequences independently, without the need to establish multi-modal association between corresponding optical and sonar features. In addition to improving the motion estimation accuracy, advantages of the proposed method comprise overcoming certain inherent ambiguities of monocular vision, e.g., the scale-factor ambiguity, and although rare, up to three interpretations for certain scene structures and camera motion. Experiment with synthetic and real data are presented in support of our technical contribution.","PeriodicalId":113677,"journal":{"name":"OCEANS 2008","volume":"79 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2008-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"9","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"OCEANS 2008","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/OCEANS.2008.5151985","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 9

Abstract

Imaging system are routinely deployed for underwater search, inspection and scientific surveys of manmade and natural structures, etc. Optical cameras, while providing high resolution and target details, have range limitation according to water visibility and become ineffective in turbid environments. In comparison, high-frequency (MHz) 2-D imaging sonar video systems, introduced to the commercial market in recent years, image targets at distances of 10's of meters in highly turbid waters. Visibility permitting, the integration of visual cues in 2-D optical and sonar data would enable better performance compared to deploying either imaging system alone. We address the problem of motion estimation- e.g., for vision-based navigation and target-based positioning of a mobile submersible platform- from 2-D optical and sonar images. The application of structure from motion paradigm in this multimodal imaging scenario also enables the 3-D reconstruction of scene features. We rely on the tracking of features in the sonar and optical motion sequences independently, without the need to establish multi-modal association between corresponding optical and sonar features. In addition to improving the motion estimation accuracy, advantages of the proposed method comprise overcoming certain inherent ambiguities of monocular vision, e.g., the scale-factor ambiguity, and although rare, up to three interpretations for certain scene structures and camera motion. Experiment with synthetic and real data are presented in support of our technical contribution.
二维光学和声纳视频中任意场景的三维运动和结构估计
成像系统通常用于水下搜索、人工和自然结构物的检查和科学调查等。光学相机虽然能提供高分辨率和目标细节,但受水面能见度的限制,在浑浊环境中效果不佳。相比之下,近年来引入商业市场的高频(MHz)二维成像声纳视频系统,在高度浑浊的水域中,图像目标距离为10米。在能见度允许的情况下,与单独部署任何成像系统相比,将视觉线索集成到2d光学和声纳数据中可以实现更好的性能。我们解决了运动估计的问题-例如,基于视觉的导航和基于目标的移动潜水平台定位-从二维光学和声纳图像。在这种多模态成像场景中,运动结构范式的应用也使场景特征的三维重建成为可能。我们独立依赖声纳和光学运动序列中的特征跟踪,而不需要在相应的光学和声纳特征之间建立多模态关联。除了提高运动估计精度外,该方法的优点还包括克服了单目视觉的某些固有歧义,例如比例因子歧义,以及对某些场景结构和摄像机运动的多达三种解释(尽管罕见)。用合成数据和实际数据进行了实验,以支持我们的技术贡献。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信