{"title":"基于倍频变换器的伪装目标交叉感知网络","authors":"Feng Dong , Jinchao Zhu , Hongpeng Wang","doi":"10.1016/j.knosys.2025.114638","DOIUrl":null,"url":null,"abstract":"<div><div>Camouflaged object detection (COD) aims to identify objects that are fully blended into their surrounding environments. Current mainstream COD methods primarily focus on pixel-level optimization using convolutional neural networks (CNNs), without sufficiently addressing the significance of frequency interactions between candidate targets and noisy backgrounds, which are crucial for obtaining accurate edge and localization information. This paper explores the integration of multi-frequency features, and constructs a cross-frequency aware network (CFANet). The proposed network utilizes precisely learned deep-layer low-frequency features to guide other layers, achieving coarse localization. To further refine segmentation, the network employs both Transformer and CNN structures to facilitate the interaction and optimization of high- and low-frequency features at local and global levels. The model adopts a localization-guided decoder structure (LGS) that allows deep-layer low-frequency features to play a key role in guiding localization. The discussion module (DM) comprises three feature extraction experts, who engage in a teacher-student learning framework to derive more accurate deep-layer low-frequency features. In the Octave-Transformer module (OTM), the high- and low-frequency fused features based on octave convolution (OctConv) and Transformer deeply mine semantic features and detailed information. Compared to 33 existing state-of-the-art COD methods, the proposed network achieves overall superior performance across four benchmark datasets. Additionally, the network demonstrates excellent performance in other downstream tasks, such as polyp segmentation, surface defect detection. Our code is available at <span><span>https://github.com/wkkwll-df/CFANet</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49939,"journal":{"name":"Knowledge-Based Systems","volume":"330 ","pages":"Article 114638"},"PeriodicalIF":7.6000,"publicationDate":"2025-10-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Cross-frequency aware network for camouflaged object detection with octave-transformer\",\"authors\":\"Feng Dong , Jinchao Zhu , Hongpeng Wang\",\"doi\":\"10.1016/j.knosys.2025.114638\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Camouflaged object detection (COD) aims to identify objects that are fully blended into their surrounding environments. Current mainstream COD methods primarily focus on pixel-level optimization using convolutional neural networks (CNNs), without sufficiently addressing the significance of frequency interactions between candidate targets and noisy backgrounds, which are crucial for obtaining accurate edge and localization information. This paper explores the integration of multi-frequency features, and constructs a cross-frequency aware network (CFANet). The proposed network utilizes precisely learned deep-layer low-frequency features to guide other layers, achieving coarse localization. To further refine segmentation, the network employs both Transformer and CNN structures to facilitate the interaction and optimization of high- and low-frequency features at local and global levels. The model adopts a localization-guided decoder structure (LGS) that allows deep-layer low-frequency features to play a key role in guiding localization. The discussion module (DM) comprises three feature extraction experts, who engage in a teacher-student learning framework to derive more accurate deep-layer low-frequency features. In the Octave-Transformer module (OTM), the high- and low-frequency fused features based on octave convolution (OctConv) and Transformer deeply mine semantic features and detailed information. Compared to 33 existing state-of-the-art COD methods, the proposed network achieves overall superior performance across four benchmark datasets. Additionally, the network demonstrates excellent performance in other downstream tasks, such as polyp segmentation, surface defect detection. Our code is available at <span><span>https://github.com/wkkwll-df/CFANet</span><svg><path></path></svg></span>.</div></div>\",\"PeriodicalId\":49939,\"journal\":{\"name\":\"Knowledge-Based Systems\",\"volume\":\"330 \",\"pages\":\"Article 114638\"},\"PeriodicalIF\":7.6000,\"publicationDate\":\"2025-10-14\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Knowledge-Based Systems\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0950705125016776\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Knowledge-Based Systems","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0950705125016776","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Cross-frequency aware network for camouflaged object detection with octave-transformer
Camouflaged object detection (COD) aims to identify objects that are fully blended into their surrounding environments. Current mainstream COD methods primarily focus on pixel-level optimization using convolutional neural networks (CNNs), without sufficiently addressing the significance of frequency interactions between candidate targets and noisy backgrounds, which are crucial for obtaining accurate edge and localization information. This paper explores the integration of multi-frequency features, and constructs a cross-frequency aware network (CFANet). The proposed network utilizes precisely learned deep-layer low-frequency features to guide other layers, achieving coarse localization. To further refine segmentation, the network employs both Transformer and CNN structures to facilitate the interaction and optimization of high- and low-frequency features at local and global levels. The model adopts a localization-guided decoder structure (LGS) that allows deep-layer low-frequency features to play a key role in guiding localization. The discussion module (DM) comprises three feature extraction experts, who engage in a teacher-student learning framework to derive more accurate deep-layer low-frequency features. In the Octave-Transformer module (OTM), the high- and low-frequency fused features based on octave convolution (OctConv) and Transformer deeply mine semantic features and detailed information. Compared to 33 existing state-of-the-art COD methods, the proposed network achieves overall superior performance across four benchmark datasets. Additionally, the network demonstrates excellent performance in other downstream tasks, such as polyp segmentation, surface defect detection. Our code is available at https://github.com/wkkwll-df/CFANet.
期刊介绍:
Knowledge-Based Systems, an international and interdisciplinary journal in artificial intelligence, publishes original, innovative, and creative research results in the field. It focuses on knowledge-based and other artificial intelligence techniques-based systems. The journal aims to support human prediction and decision-making through data science and computation techniques, provide a balanced coverage of theory and practical study, and encourage the development and implementation of knowledge-based intelligence models, methods, systems, and software tools. Applications in business, government, education, engineering, and healthcare are emphasized.