{"title":"Cross-Scale Guidance Network for Few-Shot Moving Foreground Object Segmentation","authors":"Yi-Sheng Liao;Yen-Wei Lin;Ya-Han Chang;Chun-Rong Huang","doi":"10.1109/TITS.2025.3559144","DOIUrl":null,"url":null,"abstract":"Foreground object segmentation is one of the most important pre-processing steps in intelligent transportation and video surveillance systems. Although background modeling methods are efficient to segment foreground objects, their results are easily affected by dynamic backgrounds and updating strategies. Recently, deep learning-based methods have achieved more effective foreground object segmentation results compared with background modeling methods. However, a large number of labeled training frames are usually required. To reduce the number of training frames, we propose a novel cross-scale guidance network (CSGNet) for few-shot moving foreground object segmentation in surveillance videos. The proposed CSGNet contains the cross-scale feature expansion encoder and cross-scale feature guidance decoder. The encoder aims to represent the scenes by extracting cross-scale expansion features based on cross-scale and multiple field-of-view information learned from a limited number of training frames. The decoder aims to obtain accurate foreground object segmentation results under the guidance of the encoder features and the foreground loss. The proposed method outperforms the state-of-the-art background modeling methods and the deep learning-based methods around 2.6% and 3.1%, and the average computation time is 0.073 and 0.046 seconds for each frame in the CDNet2014 dataset and the UCSD dataset under a single GTX 1080 GPU computer. The source code will be available at <uri>https://github.com/nchucvml/CSGNet</uri>.","PeriodicalId":13416,"journal":{"name":"IEEE Transactions on Intelligent Transportation Systems","volume":"26 6","pages":"7726-7739"},"PeriodicalIF":8.4000,"publicationDate":"2025-04-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Intelligent Transportation Systems","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10972131/","RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, CIVIL","Score":null,"Total":0}
引用次数: 0
Abstract
Foreground object segmentation is one of the most important pre-processing steps in intelligent transportation and video surveillance systems. Although background modeling methods are efficient to segment foreground objects, their results are easily affected by dynamic backgrounds and updating strategies. Recently, deep learning-based methods have achieved more effective foreground object segmentation results compared with background modeling methods. However, a large number of labeled training frames are usually required. To reduce the number of training frames, we propose a novel cross-scale guidance network (CSGNet) for few-shot moving foreground object segmentation in surveillance videos. The proposed CSGNet contains the cross-scale feature expansion encoder and cross-scale feature guidance decoder. The encoder aims to represent the scenes by extracting cross-scale expansion features based on cross-scale and multiple field-of-view information learned from a limited number of training frames. The decoder aims to obtain accurate foreground object segmentation results under the guidance of the encoder features and the foreground loss. The proposed method outperforms the state-of-the-art background modeling methods and the deep learning-based methods around 2.6% and 3.1%, and the average computation time is 0.073 and 0.046 seconds for each frame in the CDNet2014 dataset and the UCSD dataset under a single GTX 1080 GPU computer. The source code will be available at https://github.com/nchucvml/CSGNet.
期刊介绍:
The theoretical, experimental and operational aspects of electrical and electronics engineering and information technologies as applied to Intelligent Transportation Systems (ITS). Intelligent Transportation Systems are defined as those systems utilizing synergistic technologies and systems engineering concepts to develop and improve transportation systems of all kinds. The scope of this interdisciplinary activity includes the promotion, consolidation and coordination of ITS technical activities among IEEE entities, and providing a focus for cooperative activities, both internally and externally.