{"title":"MASA-Net: Multi-Aspect Channel–Spatial Attention Network With Cross-Layer Feature Aggregation for Accurate Fungi Species Identification","authors":"Indranil Bera, Rajesh Mukherjee, Bidesh Chakraborty","doi":"10.1049/csy2.70029","DOIUrl":"https://doi.org/10.1049/csy2.70029","url":null,"abstract":"<p>Accurate identification of fungal species is essential for effective diagnosis and treatment. Traditional microscopy-based methods are often subjective and time-consuming. Deep learning has emerged as a promising tool in this domain. However, existing deep learning models often struggle to generalise in the presence of class imbalance and subtle morphological differences, which are common in fungal image datasets. This study proposes MASA-Net, a deep learning framework that combines a fine-tuned DenseNet201 backbone with a multi-aspect channel–spatial attention (MASA) module. The attention mechanism refines spatial and channel-wise features by capturing multi-scale spatial patterns and adaptively emphasising informative channels. This enhances the network's ability to focus on diagnostically relevant fungal structures while suppressing irrelevant features. The MASA-Net is evaluated on the DeFungi dataset and demonstrates superior performance in terms of accuracy, precision, recall and <i>F</i>1-score. It also outperforms established attention mechanisms such as squeeze-and-excitation networks (SE) and convolutional block attention module (CBAM) under identical conditions. These results highlight MASA-Net's robustness and effectiveness in addressing class imbalance and structural variability, offering a reliable solution for automated fungal species identification.</p>","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"7 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2025-09-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/csy2.70029","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145146841","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Bernardo Manuel Pirozzo, Mariano De Paula, Sebastián Aldo Villar, Carola de Benito, Gerardo Gabriel Acosta, Rodrigo Picos
{"title":"A Visual Odometry Artificial Intelligence-Based Method for Trajectory Learning and Tracking Applied to Mobile Robots","authors":"Bernardo Manuel Pirozzo, Mariano De Paula, Sebastián Aldo Villar, Carola de Benito, Gerardo Gabriel Acosta, Rodrigo Picos","doi":"10.1049/csy2.70028","DOIUrl":"10.1049/csy2.70028","url":null,"abstract":"<p>Autonomous systems have demonstrated high performance in several applications. One of the most important is localisation systems, which are necessary for the safe navigation of autonomous cars or mobile robots. However, despite significant advances in this field, there are still areas open for research and improvement. Two of the most important challenges include the precise traversal of a bounded route and emergencies arising from the breakdown or failure of one or more sensors, which can lead to malfunction or system localisation failure. To address these issues, auxiliary assistance systems are necessary, enabling localisation for a safe return to the starting point, completing the trajectory, or facilitating an emergency stop in a designated area for such situations. Motivated by the exploration of applying artificial intelligence to pose estimation in a navigation system, this article introduces a monocular visual odometry method that, through teach and repeat, learns and autonomously replicates trajectories. Our proposal can serve as either a primary localisation system or an auxiliary assistance system. In the first case, our approach is applicable in scenarios where the traversing route remains unchanged. In the second case, the goal is to achieve a safe return to the starting point or to reach the end point of the trajectory. We initially utilised a publicly available dataset to showcase the learning capability and robustness under different visibility conditions to validate our proposal. Subsequently, we compared our approach with other well-known methods to assess performance metrics. Finally, we evaluated real-time trajectory replication on a ground robot, both simulated and real, across multiple trajectories of increasing complexity.</p>","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"7 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2025-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/csy2.70028","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145038123","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Qiyuan Fu, Ping Liu, Qinglang Xie, Shidong Zhai, Mingjie Liu
{"title":"Clustering Path Optimisation-Based 2-Opt Rapid Wax-Drawing Trajectory Planning for Industrial 3D Wax-Drawing Robots","authors":"Qiyuan Fu, Ping Liu, Qinglang Xie, Shidong Zhai, Mingjie Liu","doi":"10.1049/csy2.70025","DOIUrl":"10.1049/csy2.70025","url":null,"abstract":"<p>The tool path trajectory serves as a cornerstone of three-dimensional (3D) printing robot technology, and path optimisation algorithms are instrumental in enabling faster, more precise and higher-quality prints. This work proposes a clustering path optimisation-based 2-opt rapid wax-drawing trajectory planning method for 3D drawing robots. Firstly, the input wax-drawing image is preprocessed to extract contour information, which is then simplified into polygons. Next, the spiral and filling trajectory algorithms are used to convert the polygons into corresponding spiral and filling paths, which are modelled as nodes in the travelling salesman problem (TSP). An improved k-means++ clustering algorithm is then designed to adaptively divide the nodes into multiple clusters. Each cluster is subsequently planned using the improved ant colony optimisation (ACO) algorithm to find the shortest path. Afterwards, the nearest-neighbour algorithm is employed to connect the shortest paths of each cluster, forming an initial tool path. Finally, the 2-opt optimisation algorithm is incorporated to optimise the preliminary path, resulting in the optimal motion trajectory for the wax-drawing tool. The verification tests show that the proposed method achieves an average reduction in path length of 30.75% compared with the parallel scanning method, traditional ant colony optimisation, Christofides with 2-opt algorithm. Meanwhile, the 3D robot wax-drawing experiments demonstrate a 17.9% reduction in drawing time, significantly improving the efficiency of large-scale production and highlighting the practical value of 3D drawing robots.</p>","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"7 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2025-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/csy2.70025","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145037520","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"RevFB-BEV: Memory-Efficient Network With Reversible Swin Transformer for 3D BEV Object Detection","authors":"Leilei Pan, Yingnan Guo, Yu Zhang","doi":"10.1049/csy2.70021","DOIUrl":"10.1049/csy2.70021","url":null,"abstract":"<p>The perception of Bird's Eye View (BEV) has become a widely adopted approach in 3D object detection due to its spatial and dimensional consistency. However, the increasing complexity of neural network architectures has resulted in higher training memory, thereby limiting the scalability of model training. To address these challenges, we propose a novel model, RevFB-BEV, which is based on the Reversible Swin Transformer (RevSwin) with Forward-Backward View Transformation (FBVT) and LiDAR Guided Back Projection (LGBP). This approach includes the RevSwin backbone network, which employs a reversible architecture to minimise training memory by recomputing intermediate parameters. Moreover, we introduce the FBVT module that refines BEV features extracted from forward projection, yielding denser and more precise camera BEV representations. The LGBP module further utilises LiDAR BEV guidance for back projection to achieve more accurate camera BEV features. Extensive experiments on the nuScenes dataset demonstrate notable performance improvements, with our model achieving over a <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <mn>4</mn>\u0000 <mo>×</mo>\u0000 </mrow>\u0000 <annotation> $4times $</annotation>\u0000 </semantics></math> reduction in training memory and a more than <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <mn>12</mn>\u0000 <mo>×</mo>\u0000 </mrow>\u0000 <annotation> $12times $</annotation>\u0000 </semantics></math> decrease in single-backbone training memory. These efficiency gains become even more pronounced with deeper network architectures. Additionally, RevFB-BEV achieves 68.1 mAP (mean Average Precision) on the validation set and 68.9 mAP on the test set, which is nearly on par with the baseline BEVFusion, underscoring its effectiveness in resource-constrained scenarios.</p>","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"7 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2025-09-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/csy2.70021","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145012633","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A General Optimisation-Based Framework for Global Pose Estimation With Multiple Sensors","authors":"Tong Qin, Shaozu Cao, Jie Pan, Shaojie Shen","doi":"10.1049/csy2.70023","DOIUrl":"10.1049/csy2.70023","url":null,"abstract":"<p>Accurate state estimation is a fundamental problem for autonomous robots. To achieve locally accurate and globally drift-free state estimation, multiple sensors with complementary properties are usually fused together. Local sensors (camera, IMU (inertial measurement unit), LiDAR, etc.) provide precise poses within a small region, whereas global sensors (GPS (global positioning system), magnetometer, barometer, etc.) supply noisy but globally drift-free localisation in a large-scale environment. In this paper, we propose a sensor fusion framework to fuse local states with global sensors, which achieves locally accurate and globally drift-free pose estimation. Local estimations, produced by existing visual odometry/visual-inertial odometry (VO/VIO) approaches, are fused with global sensors in a pose graph optimisation. Within the graph optimisation, local estimations are aligned into a global coordinate. Meanwhile, the accumulated drifts are eliminated. We evaluated the performance of our system on public datasets and with real-world experiments. The results are compared with those of other state-of-the-art algorithms. We highlight that our system is a general framework which can easily fuse various global sensors in a unified pose graph optimisation.</p>","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"7 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2025-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/csy2.70023","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144935142","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Conflict-Free Planning and Data-Driven Control of Large-Scale Nonlinear Multi-Robot Systems","authors":"You Wu, Yi Lei, Haoran Tan, Jin Guo, Yaonan Wang","doi":"10.1049/csy2.70027","DOIUrl":"10.1049/csy2.70027","url":null,"abstract":"<p>This paper addresses a crucial challenge in the domain of smart factories and intelligent warehouse logistics, focusing on conflict-free planning and the smooth operation of large-scale nonlinear mobile robots. To tackle the challenges associated with scheduling large-scale mobile robots, an improved space–time multi-robot planning algorithm is proposed. The cloud servers are adopted in this algorithm for computation, which enables faster response to the planning requirements of large-scale mobile robots. Furthermore, enhancements to a model-free adaptive predictive control method are proposed to enhance the networked control effectiveness of the nonlinear robots. The algorithm's capability to accommodate conflict-free path planning for large-scale mobile robots is demonstrated through simulation results. Experimental findings further validate the effectiveness of the cloud-based large-scale mobile robot planning and control system in achieving both conflict-free path planning and accurate path tracking. This research holds substantial implications for enhancing logistics transportation efficiency and driving advancements in the field of smart factories and intelligent warehouse logistics.</p>","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"7 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/csy2.70027","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144927354","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Tengyue Wang, Zhefan Lin, Yunze Shi, Songjie Xiao, Liangjing Yang
{"title":"Robotic Arm C-Space Trajectory Planning Using Large-Scale Digital Twin Parallelism and Safety-Prioritisable Optimal Search Algorithm","authors":"Tengyue Wang, Zhefan Lin, Yunze Shi, Songjie Xiao, Liangjing Yang","doi":"10.1049/csy2.70026","DOIUrl":"10.1049/csy2.70026","url":null,"abstract":"<p>This paper proposes a trajectory planning approach based on the configuration space (C-space) generated from large-scale digital twinning. Leveraging GPU-based parallelism, the C-space of a multi-degree-of-freedom (multi-DoF) manipulator in a complex task space with obstacles can be mapped out through extensive simulation of motion and collision of multiple virtual robot arms known as digital twins. An optimal search algorithm is incorporated with artificial potential field generated in the C-space to allow the prioritising of safety in accordance with the varying risks associated with the obstacles by means of variable repulsive potential. To extend the high-degree path to smooth and continuous joint trajectories, a spline operation is applied. Finally, a 7-DOF physical manipulator is deployed for the execution of the planned trajectory in a task space filled with obstacles. Results demonstrated a 16.3% improvement in success rate achieved by utilising the safety-prioritisable search algorithm. With this unified formulation of the control and planning problem in the C-space, the kinematics complexity of a large DOF manipulator in obstacle-present task space could be truly relieved from the joint control loop. This simplification, in turn, opens up prospective work in dynamic reconstruction of the C-space.</p>","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"7 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2025-08-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/csy2.70026","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144915120","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yang You, Zhimeng Chen, Jian Qiao, Huan Liu, Jing Fang, Jie Luo, Jiaxin Wei, Jiarong Xu
{"title":"Lightweight DETR for Small-Target Defect Detection in Stators and Rotors of Pumped Storage Generators","authors":"Yang You, Zhimeng Chen, Jian Qiao, Huan Liu, Jing Fang, Jie Luo, Jiaxin Wei, Jiarong Xu","doi":"10.1049/csy2.70022","DOIUrl":"10.1049/csy2.70022","url":null,"abstract":"<p>Real-time detection of micro-defects in pumped storage generator stators and rotors remains challenging due to small-target obscurity and edge deployment constraints. This paper proposes EdgeFault-detection transformer (DETR), a lightweight transformer model that integrates three innovations: (1) dynamic geometric-photometric augmentation for robustness, (2) a FasterNet backbone with Partial Convolution (PConv) that reduces the number of parameters by 40% (from 20 × 10<sup>6</sup> to 12 × 10<sup>6</sup>), and (3) cross-scale small-object head enhancing defect localisation. Experiments on 8763 industrial images demonstrate 75.38% [email protected] (+17.25% over RT-DETR) and 49.3% mAP<sub>small</sub> (+42.9% from baseline). The model achieves 22 FPS on NVIDIA RTX A4000 GPUs (640 × 640 resolution), validating real-time industrial applicability. Strategic computation allocation increases GFLOPs (giga floating-point operations) by 16.4% (from 58.6 to 68.2) to prioritise safety–critical precision, justifying the trade-off for detecting high-risk anomalies (e.g., insulation cracks).</p>","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"7 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2025-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/csy2.70022","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144894145","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Lightweight Hand Acupoint Recognition Based on Middle Finger Cun Measurement","authors":"Zili Meng, Minglang Lu, Guanci Yang, Tianyi Zhao, Donghua Zheng, Ling He, Zhi Shan","doi":"10.1049/csy2.70024","DOIUrl":"10.1049/csy2.70024","url":null,"abstract":"<p>Acupoint therapy plays a crucial role in the prevention and treatment of various diseases. Accurate and efficient intelligent acupoint recognition methods are essential for enhancing the operational capabilities of embodied intelligent robots in acupoint massage and related applications. This paper proposes a lightweight hand acupoint recognition (LHAR) method based on middle finger cun measurement. First, to obtain a lightweight model for rapid positioning of the hand area, on the basis of the design of the partially convolutional gated regularisation unit and the efficient shared convolutional detection head, an improved YOLO11 algorithm based on a lightweight efficient shared convolutional detection head (YOLO11-SH) was proposed. Second, according to the theory of traditional Chinese medicine, a method of positional relationship determination between acupoints based on middle finger cun measurement is established. The MediaPipe algorithm is subsequently used to obtain 21 keypoints of the hand and serves as a reference point for obtaining features of middle finger cun via positional relationship determination. Then, the offset-based localisation approach is adopted to achieve accurate recognition of acupoints by using the obtained feature of middle finger cun. Comparative experiments with five representative lightweight models demonstrate that YOLO11-SH achieves an [email protected] of 97.3%, with 1.59 × 10<sup>6</sup> parameters, 3.9 × 10<sup>9</sup> FLOPs, a model weight of 3.4 MB and an inference speed of 325.8 FPS, outperforming the comparison methods in terms of both recognition accuracy and model efficiency. The experimental results of acupoint recognition indicate that the overall recognition accuracy of LHAR has reached 94.49%. The average normalised displacement error for different acupoints ranges from 0.036 to 0.105, all within the error threshold of ≤ 0.15. Finally, LHAR is integrated into the robotic platform, and a robotic massage experiment is conducted to verify the effectiveness of LHAR.</p>","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"7 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2025-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/csy2.70024","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144892497","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Qingyang Xu, Siwei Huang, Yong Song, Bao Pang, Chengjin Zhang
{"title":"Hybrid Dynamic Point Removal and Ellipsoid Modelling of Object-Based Semantic SLAM","authors":"Qingyang Xu, Siwei Huang, Yong Song, Bao Pang, Chengjin Zhang","doi":"10.1049/csy2.70020","DOIUrl":"10.1049/csy2.70020","url":null,"abstract":"<p>For the issue of low positioning accuracy in dynamic environments with traditional simultaneous localisation and mapping (SLAM), a dynamic point removal strategy combining object detection and optical flow tracking has been proposed. To fully utilise the semantic information, an ellipsoid model of the detected semantic objects was first constructed based on the plane and point cloud constraints, which assists in loop closure detection. Bilateral semantic map matching was achieved through the Kuhn–Munkres (KM) algorithm maximum weight assignment, and the pose transformation between local and global maps was determined by the random sample consensus (RANSAC) algorithm. Finally, a stable semantic SLAM system suitable for dynamic environments was constructed. The effectiveness of achieving the system's positioning accuracy under dynamic interference and large visual-inertial loop closure was verified by the experiment.</p>","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"7 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2025-08-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/csy2.70020","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144881343","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}