Seeing Through Pixel Motion: Learning Obstacle Avoidance From Optical Flow With One Camera

IF 4.6 2区 计算机科学 Q2 ROBOTICS
Yu Hu;Yuang Zhang;Yunlong Song;Yang Deng;Feng Yu;Linzuo Zhang;Weiyao Lin;Danping Zou;Wenxian Yu
{"title":"Seeing Through Pixel Motion: Learning Obstacle Avoidance From Optical Flow With One Camera","authors":"Yu Hu;Yuang Zhang;Yunlong Song;Yang Deng;Feng Yu;Linzuo Zhang;Weiyao Lin;Danping Zou;Wenxian Yu","doi":"10.1109/LRA.2025.3560842","DOIUrl":null,"url":null,"abstract":"Optical flow captures the motion of pixels in an image sequence over time, providing information about movement, depth, and environmental structure. Flying insects utilize this information to navigate and avoid obstacles, allowing them to execute highly agile maneuvers even in complex environments. Despite its potential, autonomous flying robots have yet to fully leverage this motion information to achieve comparable levels of agility and robustness. The main challenges are two-fold: 1) extracting accurate optical flow from visual data during high-speed flight and 2) designing a robust controller that can handle noisy optical flow estimations while ensuring robust performance in complex environments. To address these challenges, we propose a novel end-to-end system for quadrotor obstacle avoidance using monocular optical flow. We develop an efficient differentiable simulator coupled with a simplified quadrotor model, allowing our policy to be trained directly through first-order gradient optimization. Additionally, we introduce a central flow attention mechanism and an action-guided active sensing strategy that enhances the policy's focus on task-relevant optical flow observations to enable more responsive decision-making during flight. Our system is validated both in simulation and the real world using an FPV racing drone. Despite being trained in a simple environment in simulation, our system demonstrates agile and robust flight in various unknown, cluttered environments in the real world at speeds of up to 6 m/s.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"10 6","pages":"5871-5878"},"PeriodicalIF":4.6000,"publicationDate":"2025-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Robotics and Automation Letters","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10964683/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ROBOTICS","Score":null,"Total":0}
引用次数: 0

Abstract

Optical flow captures the motion of pixels in an image sequence over time, providing information about movement, depth, and environmental structure. Flying insects utilize this information to navigate and avoid obstacles, allowing them to execute highly agile maneuvers even in complex environments. Despite its potential, autonomous flying robots have yet to fully leverage this motion information to achieve comparable levels of agility and robustness. The main challenges are two-fold: 1) extracting accurate optical flow from visual data during high-speed flight and 2) designing a robust controller that can handle noisy optical flow estimations while ensuring robust performance in complex environments. To address these challenges, we propose a novel end-to-end system for quadrotor obstacle avoidance using monocular optical flow. We develop an efficient differentiable simulator coupled with a simplified quadrotor model, allowing our policy to be trained directly through first-order gradient optimization. Additionally, we introduce a central flow attention mechanism and an action-guided active sensing strategy that enhances the policy's focus on task-relevant optical flow observations to enable more responsive decision-making during flight. Our system is validated both in simulation and the real world using an FPV racing drone. Despite being trained in a simple environment in simulation, our system demonstrates agile and robust flight in various unknown, cluttered environments in the real world at speeds of up to 6 m/s.
透过像素运动:学习避障从光流与一个相机
光流捕获图像序列中像素随时间的运动,提供有关运动、深度和环境结构的信息。飞虫利用这些信息来导航和避开障碍物,使它们即使在复杂的环境中也能执行高度敏捷的机动。尽管有潜力,自主飞行机器人还没有充分利用这些运动信息来达到相当水平的敏捷性和鲁棒性。主要的挑战有两个方面:1)在高速飞行期间从视觉数据中提取准确的光流;2)设计一个鲁棒控制器,可以处理有噪声的光流估计,同时确保在复杂环境下的鲁棒性能。为了解决这些挑战,我们提出了一种新颖的端到端四旋翼避障系统,使用单目光流。我们开发了一个高效的可微模拟器与简化的四旋翼模型相结合,允许我们的策略直接通过一阶梯度优化进行训练。此外,我们引入了一个中心流注意机制和一个行动导向的主动感知策略,以增强政策对任务相关光流观测的关注,从而在飞行过程中做出更灵敏的决策。我们的系统在模拟和现实世界中使用FPV赛车无人机进行了验证。尽管在简单的模拟环境中进行了训练,但我们的系统在现实世界中各种未知、混乱的环境中以高达6米/秒的速度展示了敏捷和稳健的飞行。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
IEEE Robotics and Automation Letters
IEEE Robotics and Automation Letters Computer Science-Computer Science Applications
CiteScore
9.60
自引率
15.40%
发文量
1428
期刊介绍: The scope of this journal is to publish peer-reviewed articles that provide a timely and concise account of innovative research ideas and application results, reporting significant theoretical findings and application case studies in areas of robotics and automation.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信