Neural Networks最新文献

筛选
英文 中文
Intra-class progressive and adaptive self-distillation 阶级内部渐进和适应性的自我升华
IF 6 1区 计算机科学
Neural Networks Pub Date : 2025-03-21 DOI: 10.1016/j.neunet.2025.107404
Jianping Gou , Jiaye Lin , Lin Li , Weihua Ou , Baosheng Yu , Zhang Yi
{"title":"Intra-class progressive and adaptive self-distillation","authors":"Jianping Gou ,&nbsp;Jiaye Lin ,&nbsp;Lin Li ,&nbsp;Weihua Ou ,&nbsp;Baosheng Yu ,&nbsp;Zhang Yi","doi":"10.1016/j.neunet.2025.107404","DOIUrl":"10.1016/j.neunet.2025.107404","url":null,"abstract":"<div><div>In recent years, knowledge distillation (KD) has become widely used in compressing models, training compact and efficient students to reduce computational load and training time due to the increasing parameters in deep neural networks. To minimize training costs, self-distillation has been proposed, with methods like offline-KD and online-KD requiring pre-trained teachers and multiple networks. However, these self-distillation methods often overlook feature knowledge and category information. In this paper, we introduce Intra-class Progressive and Adaptive Self-Distillation (IPASD), which transfers knowledge from the front to the back in adjacent epochs. This method extracts class-typical features and promotes compactness within classes. By integrating feature-level and logits-level knowledge into strong teacher knowledge and using ground-truth labels as supervision signals, we adaptively optimize the model. We evaluated IPASD on CIFAR-10, CIFAR-100, Tiny ImageNet, Plant Village datasets, and ImageNet showing its superiority over state-of-the-art self-distillation methods in knowledge transfer and model compression. Our codes are available at: <span><span>https://github.com/JLinye/IPASD</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"188 ","pages":"Article 107404"},"PeriodicalIF":6.0,"publicationDate":"2025-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143704893","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Structure information preserving domain adaptation network for fault diagnosis of Sucker Rod Pumping systems 有杆抽油系统故障诊断的结构信息保持域自适应网络
IF 6 1区 计算机科学
Neural Networks Pub Date : 2025-03-21 DOI: 10.1016/j.neunet.2025.107392
Xiaohua Gu , Fei Lu , Liping Yang , Kan Wang , Lusi Li , Guang Yang , Yiling Sun
{"title":"Structure information preserving domain adaptation network for fault diagnosis of Sucker Rod Pumping systems","authors":"Xiaohua Gu ,&nbsp;Fei Lu ,&nbsp;Liping Yang ,&nbsp;Kan Wang ,&nbsp;Lusi Li ,&nbsp;Guang Yang ,&nbsp;Yiling Sun","doi":"10.1016/j.neunet.2025.107392","DOIUrl":"10.1016/j.neunet.2025.107392","url":null,"abstract":"<div><div>Fault diagnosis is of great importance to the reliability and security of Sucker Rod Pumping (SRP) oil production system. With the development of digital oilfield, data-driven deep learning SRP fault diagnosis has become the development trend of oilfield system. However, due to the different working conditions, time periods, and areas, the fault diagnosis models trained from certain SRP data do not consider the statistical discrepancy of different SRP systems, resulting in insufficient generalization. To consider the fault diagnosis and generalization performances of deep models at the same time, this paper proposes a Structure Information Preserving Domain Adaptation Network (SIP-DAN) for SRP fault diagnosis. Different from the usual domain adaptation methods, SIP-DAN divides the source domain data into different subdomains according to the fault categories of the source domain, and then realizes structure information preserving domain adaptation through subdomains alignment of the source domain and the target domain. Due to the lack of fault category information in the target domain, we designed a Classifier Voting Assisted Alignment (CVAA) mechanism. The target domain data are divided into clusters using fuzzy clustering algorithm. Then, fault diagnosis classifier trained in source domain is employed to classify the samples in each cluster, and the majority voting principle is used to assign pseudo-labels to each cluster in the target domain. With these pseudo-labels, source and target subdomains alignment is carried out by optimizing the Local Maximum Mean Discrepancy (LMMD) loss to achieve fine-grained domain adaptation. Experimental results illustrate that the proposed method is better than the existing methods in fault diagnosis of SRP systems.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"188 ","pages":"Article 107392"},"PeriodicalIF":6.0,"publicationDate":"2025-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143714782","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Central loss guides coordinated Transformer for reliable anatomical landmark detection 中心损失引导协调变压器可靠的解剖地标检测
IF 6 1区 计算机科学
Neural Networks Pub Date : 2025-03-21 DOI: 10.1016/j.neunet.2025.107391
Qikui Zhu , Yihui Bi , Jie Chen , Xiangpeng Chu , Danxin Wang , Yanqing Wang
{"title":"Central loss guides coordinated Transformer for reliable anatomical landmark detection","authors":"Qikui Zhu ,&nbsp;Yihui Bi ,&nbsp;Jie Chen ,&nbsp;Xiangpeng Chu ,&nbsp;Danxin Wang ,&nbsp;Yanqing Wang","doi":"10.1016/j.neunet.2025.107391","DOIUrl":"10.1016/j.neunet.2025.107391","url":null,"abstract":"<div><div>Heatmap-based anatomical landmark detection is still facing two unresolved challenges: (1) inability to accurately evaluate the distribution of heatmap; (2) inability to effectively exploit global spatial structure information. To address the computational inability challenge, we propose a novel position-aware and sample-aware central loss. Specifically, our central loss can absorb position information, enabling accurate evaluation of the heatmap distribution. More advanced is that our central loss is sample-aware, which can adaptively distinguish easy and hard samples and make the model more focused on hard samples while solving the challenge of extreme imbalance between landmarks and non-landmarks. To address the challenge of ignoring structure information, a Coordinated Transformer, called CoorTransformer, is proposed, which establishes long-range dependencies under the guidance of landmark coordinate information, making the attention more focused on the sparse landmarks while taking advantage of global spatial structure. Furthermore, CoorTransformer can speed up convergence, effectively avoiding the defect that Transformers have difficulty converging in sparse representation learning. Using the advanced CoorTransformer and central loss, we propose a generalized detection model that can handle various scenarios, inherently exploiting the underlying relationship between landmarks and incorporating rich structural knowledge around the target landmarks. We analyzed and evaluated CoorTransformer and central loss on three challenging landmark detection tasks. The experimental results show that our CoorTransformer outperforms state-of-the-art methods, and the central loss significantly improves the model’s performance with <span><math><mi>p</mi></math></span>-values <span><math><mrow><mo>&lt;</mo><mn>0</mn><mo>.</mo><mn>05</mn></mrow></math></span>. The source code of this work is available at the <span><span>GitHub repository</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"187 ","pages":"Article 107391"},"PeriodicalIF":6.0,"publicationDate":"2025-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143696808","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mathematical expression exploration with graph representation and generative graph neural network 用图表示和生成图神经网络探索数学表达式
IF 6 1区 计算机科学
Neural Networks Pub Date : 2025-03-21 DOI: 10.1016/j.neunet.2025.107405
Jingyi Liu , Weijun Li , Lina Yu , Min Wu , Wenqiang Li , Yanjie Li , Meilan Hao
{"title":"Mathematical expression exploration with graph representation and generative graph neural network","authors":"Jingyi Liu ,&nbsp;Weijun Li ,&nbsp;Lina Yu ,&nbsp;Min Wu ,&nbsp;Wenqiang Li ,&nbsp;Yanjie Li ,&nbsp;Meilan Hao","doi":"10.1016/j.neunet.2025.107405","DOIUrl":"10.1016/j.neunet.2025.107405","url":null,"abstract":"<div><div>Symbolic Regression (SR) methods in tree representations have exhibited commendable outcomes across Genetic Programming (GP) and deep learning search paradigms. Nonetheless, the tree representation of mathematical expressions occasionally embodies redundant substructures. Representing expressions as computation graphs is more succinct and intuitive through graph representation. Despite its adoption in evolutionary strategies within SR, deep learning paradigms remain under-explored. Acknowledging the profound advancements of deep learning in tree-centric SR approaches, we advocate for addressing SR tasks using the Directed Acyclic Graph (DAG) representation of mathematical expressions, complemented by a generative graph neural network. We name the proposed method as <em><strong>Graph</strong>-based <strong>D</strong>eep <strong>S</strong>ymbolic <strong>R</strong>egression (GraphDSR)</em>. We vectorize node types and employ an adjacent matrix to delineate connections. The graph neural networks craft the DAG incrementally, sampling node types and graph connections conditioned on previous DAG at every step. During each sample step, the valid check is implemented to avoid meaningless sampling, and four domain-agnostic constraints are adopted to further streamline the search. This process culminates once a coherent expression emerges. Constants undergo optimization by SGD and BFGS algorithms, and rewards refine the graph neural network through reinforcement learning. A comprehensive evaluation across 110 benchmarks underscores the potency of our approach.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"187 ","pages":"Article 107405"},"PeriodicalIF":6.0,"publicationDate":"2025-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143726271","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enabling scale and rotation invariance in convolutional neural networks with retina like transformation 在具有视网膜样变换的卷积神经网络中实现缩放和旋转不变性
IF 6 1区 计算机科学
Neural Networks Pub Date : 2025-03-20 DOI: 10.1016/j.neunet.2025.107395
Jiahong Zhang , Guoqi Li , Qiaoyi Su , Lihong Cao , Yonghong Tian , Bo Xu
{"title":"Enabling scale and rotation invariance in convolutional neural networks with retina like transformation","authors":"Jiahong Zhang ,&nbsp;Guoqi Li ,&nbsp;Qiaoyi Su ,&nbsp;Lihong Cao ,&nbsp;Yonghong Tian ,&nbsp;Bo Xu","doi":"10.1016/j.neunet.2025.107395","DOIUrl":"10.1016/j.neunet.2025.107395","url":null,"abstract":"<div><div>Traditional convolutional neural networks (CNNs) struggle with scale and rotation transformations, resulting in reduced performance on transformed images. Previous research focused on designing specific CNN modules to extract transformation-invariant features. However, these methods lack versatility and are not adaptable to a wide range of scenarios. Drawing inspiration from human visual invariance, we propose a novel brain-inspired approach to tackle the invariance problem in CNNs. If we consider a CNN as the visual cortex, we have the potential to design an “eye” that exhibits transformation invariance, allowing CNNs to perceive the world consistently. Therefore, we propose a retina module and then integrate it into CNNs to create transformation-invariant CNNs (TICNN), achieving scale and rotation invariance. The retina module comprises a retina-like transformation and a transformation-aware neural network (TANN). The retina-like transformation supports flexible image transformations, while the TANN regulates these transformations for scaling and rotation. Specifically, we propose a reference-based training method (RBTM) where the retina module learns to align input images with a reference scale and rotation, thereby achieving invariance. Furthermore, we provide mathematical substantiation for the retina module to confirm its feasibility. Experimental results also demonstrate that our method outperforms existing methods in recognizing images with scale and rotation variations. The code will be released at <span><span>https://github.com/JiaHongZ/TICNN</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"187 ","pages":"Article 107395"},"PeriodicalIF":6.0,"publicationDate":"2025-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143679904","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Feature-Tuning Hierarchical Transformer via token communication and sample aggregation constraint for object re-identification 基于令牌通信和样本聚合约束的特征调优分层变压器
IF 6 1区 计算机科学
Neural Networks Pub Date : 2025-03-20 DOI: 10.1016/j.neunet.2025.107394
Zhi Yu , Zhiyong Huang , Mingyang Hou , Jiaming Pei , Yan Yan , Yushi Liu , Daming Sun
{"title":"Feature-Tuning Hierarchical Transformer via token communication and sample aggregation constraint for object re-identification","authors":"Zhi Yu ,&nbsp;Zhiyong Huang ,&nbsp;Mingyang Hou ,&nbsp;Jiaming Pei ,&nbsp;Yan Yan ,&nbsp;Yushi Liu ,&nbsp;Daming Sun","doi":"10.1016/j.neunet.2025.107394","DOIUrl":"10.1016/j.neunet.2025.107394","url":null,"abstract":"<div><div>Recently, transformer-based methods have shown remarkable success in object re-identification. However, most works directly embed off-the-shelf transformer backbones for feature extraction. These methods treat all patch tokens equally, ignoring the difference of distinct patch tokens for feature representation. To solve this issue, this paper designs a feature-tuning mechanism for transformer backbones to emphasize important patches and attenuate unimportant patches. Specifically, a Feature-tuning Hierarchical Transformer (FHTrans) for object re-identification is proposed. First, we propose a plug-and-play Feature-tuning module via Token Communication (TCF) deployed within transformer encoder blocks. This module regards the class token as a pivot to achieve communication between patch tokens. Important patch tokens are emphasized, while unimportant patch tokens are attenuated, focusing more precisely on the discriminative features related to object distinction. Then, we construct a FHTrans based on the designed feature-tuning module. The encoder blocks are divided into three hierarchies considering the correlation between feature representativeness and transformer depth. As the hierarchy deepens, the communication between tokens becomes tighter. This enables the model to capture more crucial feature information. Finally, we propose a Sample Aggregation (SA) loss to impose more effective constraints on statistical characteristics among samples, thereby enhancing intra-class aggregation and guiding FHTrans to learn more discriminative features. Experiments on object re-identification benchmarks demonstrate that our method can achieve state-of-the-art performance.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"187 ","pages":"Article 107394"},"PeriodicalIF":6.0,"publicationDate":"2025-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143679899","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Estimating uncertainty from feed-forward network based sensing using quasi-linear approximation 基于准线性逼近的前馈网络传感不确定性估计
IF 6 1区 计算机科学
Neural Networks Pub Date : 2025-03-20 DOI: 10.1016/j.neunet.2025.107376
Songhan Zhang , Matthew Singh , Delsin Menolascino , ShiNung Ching
{"title":"Estimating uncertainty from feed-forward network based sensing using quasi-linear approximation","authors":"Songhan Zhang ,&nbsp;Matthew Singh ,&nbsp;Delsin Menolascino ,&nbsp;ShiNung Ching","doi":"10.1016/j.neunet.2025.107376","DOIUrl":"10.1016/j.neunet.2025.107376","url":null,"abstract":"<div><div>A fundamental problem in neural network theory is the quantification of uncertainty as it propagates through these constructs. Such quantification is crucial as neural networks become integrated into broader engineered systems that render decisions based on their outputs. In this paper, we engage the problem of estimating uncertainty in feedforward neural network constructs. Mathematically, the problem, in essence, amounts to understanding how the moments of an input distribution become modifies as they move through network layers. Despite its straightforward formulation, the nonlinear nature of modern feedforward architectures makes this is a mathematically challenging problem. Most contemporary approaches rely on some form of Monte Carlo sampling to construct inter-laminar distributions. Here, we borrow an approach from the control systems community known as quasilinear approximation, to enable a more analytical approach to the uncertainty quantification problem in this setting. Specifically, by using quasilinear approximation, nonlinearities are linearized in terms of the expectation of their gain in an input–output sense. We derive these expectations for several commonly used nonlinearities, under the assumption of Gaussian inputs. We then establish that the ensuing approximation is accurate relative to traditional linearization. Furthermore, we provide a rigorous example how this method can enable formal estimation of uncertainty in latent variables upstream of the network, within a target-tracking case study.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"188 ","pages":"Article 107376"},"PeriodicalIF":6.0,"publicationDate":"2025-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143704897","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Decomposition-based multi-scale transformer framework for time series anomaly detection 基于分解的多尺度变压器时间序列异常检测框架
IF 6 1区 计算机科学
Neural Networks Pub Date : 2025-03-20 DOI: 10.1016/j.neunet.2025.107399
Wenxin Zhang , Cuicui Luo
{"title":"Decomposition-based multi-scale transformer framework for time series anomaly detection","authors":"Wenxin Zhang ,&nbsp;Cuicui Luo","doi":"10.1016/j.neunet.2025.107399","DOIUrl":"10.1016/j.neunet.2025.107399","url":null,"abstract":"<div><div>Time series anomaly detection is crucial for maintaining stable systems. Existing methods face two main challenges. First, it is difficult to directly model the dependencies of diverse and complex patterns within the sequences. Second, many methods that optimize parameters using mean squared error struggle with noise in the time series, leading to performance deterioration. To address these challenges, we propose a transformer-based framework built on decomposition (TransDe) for multivariate time series anomaly detection. The key idea is to combine the strengths of time series decomposition and transformers to effectively learn the complex patterns in normal time series data. A multi-scale patch-based transformer architecture is proposed to exploit the representative dependencies of each decomposed component of the time series. Furthermore, a contrastive learn paradigm based on patch operation is proposed, which leverages KL divergence to align the positive pairs, namely the pure representations of normal patterns between different patch-level views. A novel asynchronous loss function with a stop-gradient strategy is further introduced to enhance the performance of TransDe effectively. It can avoid time-consuming and labor-intensive computation costs in the optimization process. Extensive experiments on five public datasets are conducted and TransDe shows superiority compared with twelve baselines in terms of F1 score. Our code is available at <span><span>https://github.com/shaieesss/TransDe</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"187 ","pages":"Article 107399"},"PeriodicalIF":6.0,"publicationDate":"2025-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143679900","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A shape composition method for named entity recognition 一种用于命名实体识别的形状组合方法。
IF 6 1区 计算机科学
Neural Networks Pub Date : 2025-03-19 DOI: 10.1016/j.neunet.2025.107389
Ying Hu , Yanping Chen , Yong Xu
{"title":"A shape composition method for named entity recognition","authors":"Ying Hu ,&nbsp;Yanping Chen ,&nbsp;Yong Xu","doi":"10.1016/j.neunet.2025.107389","DOIUrl":"10.1016/j.neunet.2025.107389","url":null,"abstract":"<div><div>Large language models (LLMs) roughly encode a sentence into a dense representation (a vector), which mixes up the semantic expression of all named entities within a sentence. So the decoding process is easily overwhelmed by sentence-specific information learned during the pre-training process. It results in seriously performance degeneration in recognizing named entities, especially annotated with nested structures. In contrast to LLMs condensing a sentence into a single vector, our model adopts a discriminative language model to map each sentence into a high-order semantic space. In this space, named entities are decomposed into entity body and entity edge. The decomposition is effective to decode complex semantic structures of named entities. In this paper, a shape composition method is proposed for recognizing named entities. This approach leverages a multi-objective learning neural architecture to simultaneously detect entity bodies and classify entity edges. During training, the dual objectives for body and edge learning guide the deep network to encode more task-relevant semantic information. Our method is evaluated on eight widely used public datasets and demonstrated competitive performance. Analytical experiments show that the strategy of let semantic expressions take its course aligns with the entity recognition task. This approach yields finer-grained semantic representations, which enhance not only NER but also other NLP tasks.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"187 ","pages":"Article 107389"},"PeriodicalIF":6.0,"publicationDate":"2025-03-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143674855","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Expert guidance and partially-labeled data collaboration for multi-organ segmentation 多器官分割的专家指导和部分标记数据协作
IF 6 1区 计算机科学
Neural Networks Pub Date : 2025-03-19 DOI: 10.1016/j.neunet.2025.107396
Li Li , Jianyi Liu , Hanguang Xiao , Guanqun Zhou , Qiyuan Liu , Zhicheng Zhang
{"title":"Expert guidance and partially-labeled data collaboration for multi-organ segmentation","authors":"Li Li ,&nbsp;Jianyi Liu ,&nbsp;Hanguang Xiao ,&nbsp;Guanqun Zhou ,&nbsp;Qiyuan Liu ,&nbsp;Zhicheng Zhang","doi":"10.1016/j.neunet.2025.107396","DOIUrl":"10.1016/j.neunet.2025.107396","url":null,"abstract":"<div><div>Abdominal multi-organ segmentation in computed tomography (CT) scans has exhibited successful applications in numerous real clinical scenarios. Nevertheless, prevailing methods for multi-organ segmentation often necessitate either a substantial volume of datasets derived from a single healthcare institution or the centralized storage of patient data obtained from diverse healthcare institutions. This prevailing approach significantly burdens data labeling and collection, thereby exacerbating the associated challenges. Compared to multi organ annotation labels, single organ annotation labels are extremely easy to obtain and have low costs. Therefor, this work establishes an effective collaborative mechanism between multi organ labels and single organ labels, and proposes an expert guided and partially-labeled data collaboration framework for multi organ segmentation, named EGPD-Seg. Firstly, a reward penalty loss function is proposed under the setting of partial labels to make the model more focused on the targets in single organ labels, while suppressing the influence of unlabeled organs on segmentation results. Then, an expert guided module is proposed to enable the model to learn prior knowledge, thereby enabling the model to obtain the ability to segment unlabeled organs on a single organ labeled dataset. The two modules interact with each other and jointly promote the multi organ segmentation performance of the model under label partial settings. This work has been effectively validated on five publicly available abdominal multi organ segmentation datasets, including internal datasets and invisible external datasets. Code: <span><span>https://github.com/LiLiXJTU/EGPDC-Seg</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"187 ","pages":"Article 107396"},"PeriodicalIF":6.0,"publicationDate":"2025-03-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143679994","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信