{"title":"Characterizing dynamic swimming behaviors of three-particle magnetic microswimmer near a solid surface","authors":"Qianqian Wang, Lidong Yang, Jiangfan Yu, Li Zhang","doi":"10.1109/ROBIO.2017.8324620","DOIUrl":"https://doi.org/10.1109/ROBIO.2017.8324620","url":null,"abstract":"Particle-based magnetically actuated microswimmers have potential to be used as microrobotic tools for biomedical applications. In this paper, we report the dynamic swimming behaviors of a magnetic microswimmer near a solid surface. This microswimmer consists of three paramagnetic microparticles and is actuated using a rotating magnetic field. The microswimmer exhibits simple rotation and propulsion with varied dynamic poses by tuning the input frequency of applied field. When the input frequency is less than 8 Hz, the microswimmer performs simple rotation and has no translational displacement. When subjected to higher input frequency (8–15 Hz), it performs propulsion, resulting in dynamic swimming behaviors. Moreover, our results indicate that higher swimming velocity is realized if the microswimmer swims near a solid surface because of the induced pressure difference in surrounding fluid of the microagent.","PeriodicalId":197159,"journal":{"name":"2017 IEEE International Conference on Robotics and Biomimetics (ROBIO)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125659169","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Simulations for time-optimal trajectory planning along parametric polynomial lane-change curves for a unicycle","authors":"Chien-Sheng Wu, Zih-Yun Chiu, Jing-Sin Liu","doi":"10.1109/ROBIO.2017.8324741","DOIUrl":"https://doi.org/10.1109/ROBIO.2017.8324741","url":null,"abstract":"G2 lane-change path imposes symmetry boundary conditions on the path geometry for autonomous point-to-point movement. This paper presents the comparative study of three planar parametric polynomial curves as lane-change paths followed by an autonomous vehicle, assuming that the neighboring lane is free. A simulated model based on unicycle that accounts for acceleration constraints, velocity bounds and the time-optimal parameterization of each path is adopted. We base the timeoptimal trajectory simulations on numerical integration for each family of lane-change curves under two different end conditions representing loose and hard curvature scenarios to highlight the effect of path curvature on achievable speed or minimal travel time. Simulations show that the dominating factor for travel time of following a lane change curve is different for loose and hard curvature condition. Conclusions for fastest lane-change performed by unicycle could provide a reference for more complex and complete dynamic model of autonomous vehicle considering tire and friction forces.","PeriodicalId":197159,"journal":{"name":"2017 IEEE International Conference on Robotics and Biomimetics (ROBIO)","volume":"60 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121946225","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Accuracy analysis for robotized assembly system","authors":"Fei Zhao, H. Gu, Cheng Li, Chanyuan Chen","doi":"10.1109/ROBIO.2017.8324688","DOIUrl":"https://doi.org/10.1109/ROBIO.2017.8324688","url":null,"abstract":"Robotized assembly with high precision is increasingly required in critical applications for high-end products. Considering the complexity of a robotized assembly system, a systematic analysis of the error chain is important to assess the success of an assembly case and to propose a proper assembly strategy. In this paper, a robotized assembly system is studied by defining the key problems, identifying the key consisting, and addressing the key procedures. In particular, this paper proposes an assembly accuracy analysis model for analyzing the error chain of misalignment, which provides a system level assembly accuracy estimation and strategy planning. A practical case is studied with the proposed accuracy analysis method, which is further verified via assembly experiments.","PeriodicalId":197159,"journal":{"name":"2017 IEEE International Conference on Robotics and Biomimetics (ROBIO)","volume":"146 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132562741","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Dynamic objects detection and tracking for a laser scanner and camera system","authors":"Cheng Zou, B. He, Liwei Zhang, Jianwei Zhang","doi":"10.1109/ROBIO.2017.8324442","DOIUrl":"https://doi.org/10.1109/ROBIO.2017.8324442","url":null,"abstract":"Dynamic object detection and tracking from complex scenes is a challenge in the field of robot vision. In this paper, we exploit a laser scanner associated with a camera to achieve the map reconstruction, removing and tracking dynamic objects. Laser-based and image-based dynamic detection method are proposed, and their quality is analyzed respectively. Efficiency and accuracy are enhanced by a hybrid method associated 3D and 2D information. A 3D Gaussian mixture probability hypothesis density-based (GM-PHD) filter is designed to track the motion trajectories of multiple dynamic objects. The experimental result conducted on KITTI public dataset show the effectiveness of our method.","PeriodicalId":197159,"journal":{"name":"2017 IEEE International Conference on Robotics and Biomimetics (ROBIO)","volume":"435 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134051763","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Camera-odometer calibration and fusion using graph based optimization","authors":"Yijia He, Yue Guo, Aixue Ye, Kui Yuan","doi":"10.1109/ROBIO.2017.8324650","DOIUrl":"https://doi.org/10.1109/ROBIO.2017.8324650","url":null,"abstract":"Monocular visual odometry (vo) estimates the camera motion only up to a scale which is prone to localization failure when the light is changing. The wheel encoders can provide metric information and accurate local localization. Fusing camera information with wheel odometer data is a good way to estimate robot motion. In such methods, calibrating camera-odometer extrinsic parameters and fusing sensor information to perform localization are key problems. We solve these problems by transforming the wheel odometry measurement to the camera frame that can construct a factor-graph edge between every two keyframes. By building factor graph, we can use graph-based optimization technology to estimate cameraodometer extrinsic parameters and fuse sensor information to estimate robot motion. We also derive the covariance matrix of the wheel odometry edges which is important when using graph-based optimization. Simulation experiments are used to validate the extrinsic calibration. For real-world experiments, we use our method to fuse the semi-direct visual odometry (SVO) with wheel encoder data, and the results show the fusion approach is effective.","PeriodicalId":197159,"journal":{"name":"2017 IEEE International Conference on Robotics and Biomimetics (ROBIO)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131743634","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Sining Yang, Shaowu Yang, Xiaodong Yi, Wenjing Yang
{"title":"Real-time globally consistent 3D grid mapping","authors":"Sining Yang, Shaowu Yang, Xiaodong Yi, Wenjing Yang","doi":"10.1109/ROBIO.2017.8324536","DOIUrl":"https://doi.org/10.1109/ROBIO.2017.8324536","url":null,"abstract":"For path planning of mobile robots in complex unknown three-dimensional (3D) environments, an accurate 3D volumetric representation of the environment is usually required. In this paper, we present a novel method to produce globally consistent 3D grid maps in real-time, through a grid-map update strategy and an efficient data structure. We transform point clouds into the world frame provided by a visual SLAM module, then process them by the grid mapping module to produce globally consistent 3D grid maps. Experimental results in an indoor scenario demonstrate that our approach is able to update the 3D grid map efficiently and can cope with loop closures while operating in real-time.","PeriodicalId":197159,"journal":{"name":"2017 IEEE International Conference on Robotics and Biomimetics (ROBIO)","volume":"146-147 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132988361","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Improved FDCM in laser scanning inspection system for workpiece deformation","authors":"Yibin Huang, Yue Guo, Kui Yuan","doi":"10.1109/ROBIO.2017.8324484","DOIUrl":"https://doi.org/10.1109/ROBIO.2017.8324484","url":null,"abstract":"Three-dimensional (3D) inspection based on machine vision is high-precision and efficient. In this paper, a laser scanning system using the triangulation measurement is designed to obtain dense point clouds of the workpiece surface, and the point clouds are projected to two-dimensional (2D) range images. Fast Directional Chamfer Matching (FDCM) is a reliable algorithm for object detection and localization, and it is improved to accelerate the time-consuming 3D registration. Although the author of FDCM greatly improved the directional chamfer matching, it remains very slow in practice. We mainly improved the line fitting process and distance transform in this algorithm, and they greatly accelerate the scanning process. Experimental results show that the scanning system with the improved algorithm can highlight the deformation of the workpiece in real time.","PeriodicalId":197159,"journal":{"name":"2017 IEEE International Conference on Robotics and Biomimetics (ROBIO)","volume":"112 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133524164","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ahmet Furkan Guc, Mert Ali Ihsan Kalin, Cem Karakadioglu, Onur Özcan
{"title":"C-Quad: A miniature, foldable quadruped with C-shaped compliant legs","authors":"Ahmet Furkan Guc, Mert Ali Ihsan Kalin, Cem Karakadioglu, Onur Özcan","doi":"10.1109/ROBIO.2017.8324389","DOIUrl":"https://doi.org/10.1109/ROBIO.2017.8324389","url":null,"abstract":"C-Quad is an origami-inspired, foldable, miniature robot whose legs and body are all machined from one PET sheet each. The already famous compliant legs are modified such that they can be manufactured from a flat PET sheet and folded into the C-shape wanted. The compliant legs enable the miniature robot to run fast and scale obstacles with ease due to the geometry of the legs. C-Quad has four legs that are manufactured separately from the main body frame, which is also manufactured from a single PET sheet. All of its legs are actuated individually with a total of four DC motors. Despite the thin PET film, the structural rigidity and robustness of the body frame is increased by using specialized folds and locks. The manufacturing and assembly of the robot takes approximately 2.5 hours. C-Quad carries a battery, an Arduino Pro Micro control board, a bluetooth communication module, custom made encoders and commercially available IR sensors for motor speed control and motor drivers, all of which weighs 38 grams. By using very simple control strategies, it can achieve the speed of 2.7 Bodylengths/sec, can perform in-place turns and can climb over obstacles more than half of its height.","PeriodicalId":197159,"journal":{"name":"2017 IEEE International Conference on Robotics and Biomimetics (ROBIO)","volume":"68 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132259727","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Vertical motion control of underwater robot based on hydrodynamics and kinematics analysis","authors":"Mingjie Dong, Wusheng Chou, Bin Fang","doi":"10.1109/ROBIO.2017.8324474","DOIUrl":"https://doi.org/10.1109/ROBIO.2017.8324474","url":null,"abstract":"Depth control of underwater robot is of vital importance, especially when it needs to work at specific depths under the water. However, vertical motion control of underwater robot is nonlinear and it is difficult to make the robot stop at the specific depth for the first time it reaches the depth for the existence of inertia. The paper proposes a vertical motion control strategy combining hydrodynamics with kinematics analysis, to make sure that the underwater robot can just stop at the specific depth with its velocity nearly zero at the same time. At first, the hydrodynamics analysis of underwater robot is finished using FLUENT to obtain the relationship between water resistance and speed of the underwater robot in heave direction. Then, the error of the hydrodynamics analysis is compensated through field experiments using least square method with the help of depth gauge. After that, the control strategy of thrust from the two vertical propellers is proposed using kinematical analysis of underwater robot, according to the difference between the current depth and the target depth. Simulation and field experiments verify the effectiveness of the proposed vertical motion control strategy.","PeriodicalId":197159,"journal":{"name":"2017 IEEE International Conference on Robotics and Biomimetics (ROBIO)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132627636","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Real-time stereo-vision localization system for safe landing of unmanned aerial vehicles","authors":"Weiwei Kong, Tianjiang Hu, Jianwei Zhang","doi":"10.1109/ROBIO.2017.8324773","DOIUrl":"https://doi.org/10.1109/ROBIO.2017.8324773","url":null,"abstract":"Orchestrating a safe landing is one of the greatest challenges for Unmanned Aerial Vehicles (UAVs). This paper aims at the autonomous localization and landing bottleneck by developing a real-time ground-based stereo visual system. This novel architecture consists of two separate perception components which are mounted with a pan-and-tilt unit (PTU) and optical sensors. Furthermore, a tracking-inspired stereo detection algorithm is proposed to improve localization accuracy. The algorithm synthesizes a Bounding Box Shrinking (BBS) approach into the Generic Object Tracking Using Regression Networks (GOTURN) method. Both datasets driven offline simulation, and online flight experiments are conducted to validate effectiveness as well as better performance of the novel system and the overall accuracy during the landing process. Also, this autonomous landing system caters for different UAV systems in operation, such as fixed-wing and rotary wing, particularly in GNSS-denied or-impaired environments.","PeriodicalId":197159,"journal":{"name":"2017 IEEE International Conference on Robotics and Biomimetics (ROBIO)","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132687369","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}