{"title":"Development of an efficient method for object detection and localization in 3D space using RGBD cameras for autonomous systems","authors":"Nataliya Boyko","doi":"10.1016/j.ijcce.2025.04.005","DOIUrl":null,"url":null,"abstract":"<div><div>The work presents an efficient algorithm for object detection, orientation estimation, and isometric positioning in 3D space using RGBD camera data. The goal of the study is to improve the accuracy and processing speed of autonomous navigation and manipulation systems under conditions of limited computational resources. The proposed approach combines heuristic isometry estimation with segmentation methods (DBSCAN), plane estimation (RANSAC), and orientation analysis, enabling effective processing of scenes with planar backgrounds. The main advantage of the algorithm lies in its ability to operate in real time: the processing time for a single frame is only 20 ms, achieving object positioning accuracy up to 5.48 cm. The results of experimental research confirm a high level of accuracy and stability even under challenging conditions. The algorithm outperforms existing models in terms of processing speed while demonstrating comparable or superior positioning accuracy. The practical significance of the proposed method lies in its potential application in mobile robotics, automated warehouse systems, and machine vision systems where high autonomy and precision are required. The algorithm can also be adapted to a broader range of tasks due to its flexible hyperparameter tuning. A key limitation remains the requirement for object placement on a planar surface and the use of a depth camera, which necessitates a specific environmental setup. The proposed method makes a significant contribution to the advancement of computer vision and autonomous robotics technologies, opening prospects for its implementation in next-generation systems.</div></div>","PeriodicalId":100694,"journal":{"name":"International Journal of Cognitive Computing in Engineering","volume":"6 ","pages":"Pages 537-551"},"PeriodicalIF":0.0000,"publicationDate":"2025-04-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Cognitive Computing in Engineering","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2666307425000245","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
The work presents an efficient algorithm for object detection, orientation estimation, and isometric positioning in 3D space using RGBD camera data. The goal of the study is to improve the accuracy and processing speed of autonomous navigation and manipulation systems under conditions of limited computational resources. The proposed approach combines heuristic isometry estimation with segmentation methods (DBSCAN), plane estimation (RANSAC), and orientation analysis, enabling effective processing of scenes with planar backgrounds. The main advantage of the algorithm lies in its ability to operate in real time: the processing time for a single frame is only 20 ms, achieving object positioning accuracy up to 5.48 cm. The results of experimental research confirm a high level of accuracy and stability even under challenging conditions. The algorithm outperforms existing models in terms of processing speed while demonstrating comparable or superior positioning accuracy. The practical significance of the proposed method lies in its potential application in mobile robotics, automated warehouse systems, and machine vision systems where high autonomy and precision are required. The algorithm can also be adapted to a broader range of tasks due to its flexible hyperparameter tuning. A key limitation remains the requirement for object placement on a planar surface and the use of a depth camera, which necessitates a specific environmental setup. The proposed method makes a significant contribution to the advancement of computer vision and autonomous robotics technologies, opening prospects for its implementation in next-generation systems.