{"title":"MIU-Net: Advanced multi-scale feature extraction and imbalance mitigation for optic disc segmentation","authors":"Yichen Xiao , Yi Shao , Zhi Chen , Ruyi Zhang , Xuan Ding , Jing Zhao , Shengtao Liu , Teruko Fukuyama , Yu Zhao , Xiaoliao Peng , Guangyang Tian , Shiping Wen , Xingtao Zhou","doi":"10.1016/j.neunet.2024.106895","DOIUrl":"10.1016/j.neunet.2024.106895","url":null,"abstract":"<div><div>Pathological myopia is a severe eye condition that can cause serious complications like retinal detachment and macular degeneration, posing a threat to vision. Optic disc segmentation helps measure changes in the optic disc and observe the surrounding retina, aiding early detection of pathological myopia. However, these changes make segmentation difficult, resulting in accuracy levels that are not suitable for clinical use. To address this, we propose a new model called MIU-Net, which improves segmentation performance through several innovations. First, we introduce a multi-scale feature extraction (MFE) module to capture features at different scales, helping the model better identify optic disc boundaries in complex images. Second, we design a dual attention module that combines channel and spatial attention to focus on important features and improve feature use. To tackle the imbalance between optic disc and background pixels, we use focal loss to enhance the model’s ability to detect minority optic disc pixels. We also apply data augmentation techniques to increase data diversity and address the lack of training data. Our model was tested on the iChallenge-PM and iChallenge-AMD datasets, showing clear improvements in accuracy and robustness compared to existing methods. The experimental results demonstrate the effectiveness and potential of our model in diagnosing pathological myopia and other medical image processing tasks.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"182 ","pages":"Article 106895"},"PeriodicalIF":6.0,"publicationDate":"2024-11-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142645082","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Neural NetworksPub Date : 2024-11-13DOI: 10.1016/j.neunet.2024.106795
Yi Jiang , Mingxiao Feng , Wengang Zhou , Houqiang Li
{"title":"Recovering Permuted Sequential Features for effective Reinforcement Learning","authors":"Yi Jiang , Mingxiao Feng , Wengang Zhou , Houqiang Li","doi":"10.1016/j.neunet.2024.106795","DOIUrl":"10.1016/j.neunet.2024.106795","url":null,"abstract":"<div><div>When applying Reinforcement Learning (RL) to the real-world visual tasks, two major challenges necessitate consideration: sample inefficiency and limited generalization. To address the above two challenges, previous works focus primarily on learning semantic information from the visual state for improving sample efficiency, but they do not explicitly learn other valuable aspects, such as spatial information. Moreover, they improve generalization by learning representations that are invariant to alterations of task-irrelevant variables, without considering task-relevant variables. To enhance sample efficiency and generalization of the base RL algorithm in visual tasks, we propose an auxiliary task called Recovering Permuted Sequential Features (RPSF). Our method enhances generalization by learning the spatial structure information of the agent, which can mitigate the effects of changes in both task-relevant and task-irrelevant variables. Moreover, it explicitly learns both semantic and spatial information from the visual state by disordering and subsequently recovering a sequence of features to generate more holistic representations, thereby improving sample efficiency. Extensive experiments demonstrate that our method significantly improves the sample efficiency and generalization of the base RL algorithm and outperforms various state-of-the-art baselines across diverse tasks in unseen environments. Furthermore, our method exhibits compatibility with both CNN and Transformer architectures.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"182 ","pages":"Article 106795"},"PeriodicalIF":6.0,"publicationDate":"2024-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142645084","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Neural NetworksPub Date : 2024-11-12DOI: 10.1016/j.neunet.2024.106890
Yuze Li , Yan Zhang , Chunling Yang , Yu Chen
{"title":"UMS2-ODNet: Unified-scale domain adaptation mechanism driven object detection network with multi-scale attention","authors":"Yuze Li , Yan Zhang , Chunling Yang , Yu Chen","doi":"10.1016/j.neunet.2024.106890","DOIUrl":"10.1016/j.neunet.2024.106890","url":null,"abstract":"<div><div>Unsupervised domain adaptation techniques improve the generalization capability and performance of detectors, especially when the source and target domains have different distributions. Compared with two-stage detectors, one-stage detectors (especially YOLO series) provide better real-time capabilities and become primary choices in industrial fields. In this paper, to improve cross-domain object detection performance, we propose a Unified-Scale Domain Adaptation Mechanism Driven Object Detection Network with Multi-Scale Attention (UMS<span><math><msup><mrow></mrow><mrow><mn>2</mn></mrow></msup></math></span>-ODNet). UMS<span><math><msup><mrow></mrow><mrow><mn>2</mn></mrow></msup></math></span>-ODNet chooses YOLOv6 as the basic framework in terms of its balance between efficiency and accuracy. UMS<span><math><msup><mrow></mrow><mrow><mn>2</mn></mrow></msup></math></span>-ODNet considers the adaptation consistency across different scale feature maps, which tends to be ignored by existing methods. A unified-scale domain adaptation mechanism is designed to fully utilize and unify the discriminative information from different scales. A multi-scale attention module is constructed to further improve the multi-scale representation ability of features. A novel loss function is created to maintain the consistency of multi-scale information by considering the homology of the descriptions from the same latent feature. Multiply experiments are conducted on four widely used datasets. Our proposed method outperforms other state-of-the-art techniques, illustrating the feasibility and effectiveness of the proposed UMS<span><math><msup><mrow></mrow><mrow><mn>2</mn></mrow></msup></math></span>-ODNet.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"181 ","pages":"Article 106890"},"PeriodicalIF":6.0,"publicationDate":"2024-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142639982","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Neural NetworksPub Date : 2024-11-12DOI: 10.1016/j.neunet.2024.106884
Wenbiao Yan , Jihua Zhu , Jinqian Chen , Haozhe Cheng , Shunshun Bai , Liang Duan , Qinghai Zheng
{"title":"Partially multi-view clustering via re-alignment","authors":"Wenbiao Yan , Jihua Zhu , Jinqian Chen , Haozhe Cheng , Shunshun Bai , Liang Duan , Qinghai Zheng","doi":"10.1016/j.neunet.2024.106884","DOIUrl":"10.1016/j.neunet.2024.106884","url":null,"abstract":"<div><div>Multi-view clustering learns consistent information from multi-view data, aiming to achieve more significant clustering characteristics. However, data in real-world scenarios often exhibit temporal or spatial asynchrony, leading to views with unaligned instances. Existing methods primarily address this issue by learning transformation matrices to align unaligned instances, but this process of learning differentiable transformation matrices is cumbersome. To address the challenge of partially unaligned instances, we propose <strong>P</strong>artially <strong>M</strong>ulti-<strong>v</strong>iew <strong>C</strong>lustering via <strong>R</strong>e-alignment (PMVCR). Our approach integrates representation learning and data alignment through a two-stage training and a re-alignment process. Specifically, our training process consists of three stages: (i) In the coarse-grained alignment stage, we construct negative instance pairs for unaligned instances and utilize contrastive learning to preliminarily learn the view representations of the instances. (ii) In the re-alignment stage, we match unaligned instances based on the similarity of their view representations, aligning them with the primary view. (iii) In the fine-grained alignment stage, we further enhance the discriminative power of the view representations and the model’s ability to differentiate between clusters. Compared to existing models, our method effectively leverages information between unaligned samples and enhances model generalization by constructing negative instance pairs. Clustering experiments on several popular multi-view datasets demonstrate the effectiveness and superiority of our method. Our code is publicly available at <span><span>https://github.com/WenB777/PMVCR.git</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"182 ","pages":"Article 106884"},"PeriodicalIF":6.0,"publicationDate":"2024-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142645083","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Neural NetworksPub Date : 2024-11-12DOI: 10.1016/j.neunet.2024.106858
Chao Li , Shaokang Dong , Shangdong Yang , Yujing Hu , Wenbin Li , Yang Gao
{"title":"Coordinating Multi-Agent Reinforcement Learning via Dual Collaborative Constraints","authors":"Chao Li , Shaokang Dong , Shangdong Yang , Yujing Hu , Wenbin Li , Yang Gao","doi":"10.1016/j.neunet.2024.106858","DOIUrl":"10.1016/j.neunet.2024.106858","url":null,"abstract":"<div><div>Many real-world multi-agent tasks exhibit a nearly decomposable structure, where interactions among agents within the same interaction set are strong while interactions between different sets are relatively weak. Efficiently modeling the nearly decomposable structure and leveraging it to coordinate agents can enhance the learning efficiency of multi-agent reinforcement learning algorithms for cooperative tasks, while existing works typically fail. To overcome this limitation, this paper proposes a novel algorithm named Dual Collaborative Constraints (DCC) that identifies the interaction sets as subtasks and achieves both intra-subtask and inter-subtask coordination. Specifically, DCC employs a bi-level structure to periodically distribute agents into multiple subtasks, and proposes both local and global collaborative constraints based on mutual information to facilitate both intra-subtask and inter-subtask coordination among agents. These two constraints ensure that agents within the same subtask reach a consensus on their local action selections and all of them select superior joint actions that maximize the overall task performance. Experimentally, we evaluate DCC on various cooperative multi-agent tasks, and its superior performance against multiple state-of-the-art baselines demonstrates its effectiveness.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"182 ","pages":"Article 106858"},"PeriodicalIF":6.0,"publicationDate":"2024-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142649523","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Neural NetworksPub Date : 2024-11-12DOI: 10.1016/j.neunet.2024.106885
Yuhang Wang , Bin Zou , Jie Xu , Chen Xu , Yuan Yan Tang
{"title":"ALR-HT: A fast and efficient Lasso regression without hyperparameter tuning","authors":"Yuhang Wang , Bin Zou , Jie Xu , Chen Xu , Yuan Yan Tang","doi":"10.1016/j.neunet.2024.106885","DOIUrl":"10.1016/j.neunet.2024.106885","url":null,"abstract":"<div><div>Lasso regression, known for its efficacy in high-dimensional data analysis and feature selection, stands as a cornerstone in the realm of supervised learning for regression estimation. However, hyperparameter tuning for Lasso regression is often time-consuming and susceptible to noisy data in big data scenarios. In this paper we introduce a new additive Lasso regression without Hyperparameter Tuning (ALR-HT) by integrating Markov resampling with additive models. We estimate the generalization bounds of the proposed ALR-HT and establish the fast learning rate. The experimental results for benchmark datasets confirm that the proposed ALR-HT algorithm has better performance in terms of sampling and training total time, mean squared error (MSE) compared to other algorithms. We present some discussions on the ALR-HT algorithm and apply it to Ridge regression, to show its versatility and effectiveness in regularized regression scenarios.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"181 ","pages":"Article 106885"},"PeriodicalIF":6.0,"publicationDate":"2024-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142639971","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Neural NetworksPub Date : 2024-11-10DOI: 10.1016/j.neunet.2024.106892
You Lu , Wenzhuo Song , Chidubem Arachie , Bert Huang
{"title":"Weakly supervised label learning flows","authors":"You Lu , Wenzhuo Song , Chidubem Arachie , Bert Huang","doi":"10.1016/j.neunet.2024.106892","DOIUrl":"10.1016/j.neunet.2024.106892","url":null,"abstract":"<div><div>Supervised learning usually requires a large amount of labeled data. However, attaining ground-truth labels is costly for many tasks. Alternatively, weakly supervised methods learn with cheap weak signals that only approximately label some data. Many existing weakly supervised learning methods learn a deterministic function that estimates labels given the input data and weak signals. In this paper, we develop label learning flows (LLF), a general framework for weakly supervised learning problems. Our method is a generative model based on normalizing flows. The main idea of LLF is to optimize the conditional likelihoods of all possible labelings of the data within a constrained space defined by weak signals. We develop a training method for LLF that trains the conditional flow inversely and avoids estimating the labels. Once a model is trained, we can make predictions with a sampling algorithm. We apply LLF to three weakly supervised learning problems. Experiment results show that our method outperforms many baselines we compare against.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"182 ","pages":"Article 106892"},"PeriodicalIF":6.0,"publicationDate":"2024-11-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142645086","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Outer synchronization and outer H<sub>∞</sub> synchronization for coupled fractional-order reaction-diffusion neural networks with multiweights.","authors":"Jin-Liang Wang, Si-Yang Wang, Yan-Ran Zhu, Tingwen Huang","doi":"10.1016/j.neunet.2024.106893","DOIUrl":"https://doi.org/10.1016/j.neunet.2024.106893","url":null,"abstract":"<p><p>This paper introduces multiple state or spatial-diffusion coupled fractional-order reaction-diffusion neural networks, and discusses the outer synchronization and outer H<sub>∞</sub> synchronization problems for these coupled fractional-order reaction-diffusion neural networks (CFRNNs). The Lyapunov functional method, Laplace transform and inequality techniques are utilized to obtain some outer synchronization conditions for CFRNNs. Moreover, some criteria are also provided to make sure the outer H<sub>∞</sub> synchronization of CFRNNs. Finally, the derived outer and outer H<sub>∞</sub> synchronization conditions are validated on the basis of two numerical examples.</p>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"181 ","pages":"106893"},"PeriodicalIF":6.0,"publicationDate":"2024-11-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142639979","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Neural NetworksPub Date : 2024-11-09DOI: 10.1016/j.neunet.2024.106883
Hiroki Yasumoto, Toshiyuki Tanaka
{"title":"Complexities of feature-based learning systems, with application to reservoir computing","authors":"Hiroki Yasumoto, Toshiyuki Tanaka","doi":"10.1016/j.neunet.2024.106883","DOIUrl":"10.1016/j.neunet.2024.106883","url":null,"abstract":"<div><div>This paper studies complexity measures of reservoir systems. For this purpose, a more general model that we call a feature-based learning system, which is the composition of a feature map and of a final estimator, is studied. We study complexity measures such as growth function, VC-dimension, pseudo-dimension and Rademacher complexity. On the basis of the results, we discuss how the unadjustability of reservoirs and the linearity of readouts can affect complexity measures of the reservoir systems. Furthermore, some of the results generalize or improve the existing results.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"182 ","pages":"Article 106883"},"PeriodicalIF":6.0,"publicationDate":"2024-11-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142645081","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}