Qi Li;Jiexin Luo;Chunxiao Chen;Jiaxin Cai;Wenjie Yang;Yuanlong Yu;Shengfeng He;Wenxi Liu
{"title":"Playing to the Strengths of High- and Low-Resolution Cues for Ultra-High Resolution Image Segmentation","authors":"Qi Li;Jiexin Luo;Chunxiao Chen;Jiaxin Cai;Wenjie Yang;Yuanlong Yu;Shengfeng He;Wenxi Liu","doi":"10.1109/LRA.2025.3579605","DOIUrl":null,"url":null,"abstract":"In ultra-high resolution image segmentation task for robotic platforms like AAVs and autonomous vehicles, existing paradigms process a downsampled input image through a deep network and the original high-resolution image through a shallow network, then fusing their features for final segmentation. Although these features are designed to be complementary, they often contain redundant or even conflicting semantic information, which leads to blurred edge contours, particularly for small objects. This is especially detrimental to robotics applications requiring precise spatial awareness. To address this challenge, we propose a novel paradigm that disentangles the task into two independent subtasks concerning high- and low-resolution inputs, leveraging high-resolution features exclusively to capture low-level structured details and low-resolution features for extracting semantics. Specifically, for the high-resolution input, we propose a region-pixel association experts scheme that partitions the image into multiple regions. For the low-resolution input, we assign compact semantic tokens to the partitioned regions. Additionally, we incorporate a high-resolution local perception scheme with an efficient field-enriched local context module to enhance small object recognition in case of incorrect semantic assignment. Extensive experiments demonstrate the state-of-the-art performance of our method and validate the effectiveness of each designed component.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"10 8","pages":"7787-7794"},"PeriodicalIF":5.3000,"publicationDate":"2025-06-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Robotics and Automation Letters","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/11034711/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ROBOTICS","Score":null,"Total":0}
引用次数: 0
Abstract
In ultra-high resolution image segmentation task for robotic platforms like AAVs and autonomous vehicles, existing paradigms process a downsampled input image through a deep network and the original high-resolution image through a shallow network, then fusing their features for final segmentation. Although these features are designed to be complementary, they often contain redundant or even conflicting semantic information, which leads to blurred edge contours, particularly for small objects. This is especially detrimental to robotics applications requiring precise spatial awareness. To address this challenge, we propose a novel paradigm that disentangles the task into two independent subtasks concerning high- and low-resolution inputs, leveraging high-resolution features exclusively to capture low-level structured details and low-resolution features for extracting semantics. Specifically, for the high-resolution input, we propose a region-pixel association experts scheme that partitions the image into multiple regions. For the low-resolution input, we assign compact semantic tokens to the partitioned regions. Additionally, we incorporate a high-resolution local perception scheme with an efficient field-enriched local context module to enhance small object recognition in case of incorrect semantic assignment. Extensive experiments demonstrate the state-of-the-art performance of our method and validate the effectiveness of each designed component.
期刊介绍:
The scope of this journal is to publish peer-reviewed articles that provide a timely and concise account of innovative research ideas and application results, reporting significant theoretical findings and application case studies in areas of robotics and automation.