{"title":"Pre-diagnosis of pelvic floor disorders-based image registration and clustering","authors":"Cicero L. Costa, Túlia A. A. Macedo, C. Barcelos","doi":"10.1109/ICAR46387.2019.8981575","DOIUrl":"https://doi.org/10.1109/ICAR46387.2019.8981575","url":null,"abstract":"Pelvic dysfunction mainly affects adult women, it is estimated that 15% of multiparous women suffer from the problem. Dysfunctions can be diagnosed by defecography, a dynamic MRI scan. Images are used by specialists to diagnose organ dysfunction such as the bladder and the early rectum. This paper presents an automated classification system that uses a non-rigid registration based on a variational model to create automatic markings from initial markings made by an expert. The classification is based on simple average and the centroids of the K-means grouping technique. The classification made by the system is evaluated by confusion matrix based metrics. The obtained results using 21 defecography exams from 21 different patients indicate that the proposed technique is a promising tool in the diagnosis of pelvic floor disorders and can assist the physician in the diagnostic process.","PeriodicalId":6606,"journal":{"name":"2019 19th International Conference on Advanced Robotics (ICAR)","volume":"25 1","pages":"572-577"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82230441","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Model-based Dynamic Pose Graph SLAM in Unstructured Dynamic Environments","authors":"Amy Deeb, M. Seto, Yajun Pan","doi":"10.1109/ICAR46387.2019.8981632","DOIUrl":"https://doi.org/10.1109/ICAR46387.2019.8981632","url":null,"abstract":"Navigation in dynamic environments is a challenge for autonomous vehicles operating without prior maps or global position references. This poses high risk to vehicles that perform scientific studies and monitoring missions in marine Arctic environments characterized by slowly moving sea ice with few truly static landmarks. Whereas mature simultaneous localization and mapping (SLAM) approaches assume a static environment, this work extends pose graph SLAM to spatiotemporally evolving environments. A novel model-based dynamic factor is proposed to capture a landmark's state transition model - whether the state be kinematic, appearance or otherwise. The structure of the state transition model is assumed to be known a priori, while the parameters are estimated on-line. Expectation maximization is used to avoid adding variables to the graph. Proof-of-concept results are shown in small- and medium-scale simulation, and small-scale laboratory environments for a small quadrotor. Preliminary laboratory validation results shows the effect of mechanical limitations of the quadrotor platform and increased uncertainties associated with the model-based dynamic factors on the SLAM estimate. Simulation results are encouraging for the application of model-based dynamic factors to dynamic landmarks with a constant-velocity kinematic model.","PeriodicalId":6606,"journal":{"name":"2019 19th International Conference on Advanced Robotics (ICAR)","volume":"78 1","pages":"123-128"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84967194","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Héctor Azpúrua, Filipe A. S. Rocha, G. Garcia, Alexandre Souza Santos, Eduardo Cota, Luiz Guilherme Barros, Alexandre S Thiago, G. Pessin, G. Freitas
{"title":"EspeleoRobô - a robotic device to inspect confined environments","authors":"Héctor Azpúrua, Filipe A. S. Rocha, G. Garcia, Alexandre Souza Santos, Eduardo Cota, Luiz Guilherme Barros, Alexandre S Thiago, G. Pessin, G. Freitas","doi":"10.1109/ICAR46387.2019.8981627","DOIUrl":"https://doi.org/10.1109/ICAR46387.2019.8981627","url":null,"abstract":"Robotic devices operating in confined environments such as underground, tunnel systems, and cave networks are currently receiving particular attention from research and industry. One example is the Brazilian mining company Vale S.A., which is employing a robot - Espeleolcobô - to access restricted areas. The robot was designed initially to inspect natural caves during teleoperated missions and is now being used to monitor dam galleries and other confined or dangerous environments. This paper describes recent developments regarding locomotion mechanisms and mobility analyses, localization, and path planning strategies aiming autonomous robot operation. Preliminary results from simulation and field experiments validate the robotic device concept. Lessons learned from multiple field tests in industrial scenarios are also described.","PeriodicalId":6606,"journal":{"name":"2019 19th International Conference on Advanced Robotics (ICAR)","volume":"75 1","pages":"17-23"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83380279","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Edison Kleiber Titito Concha, Diego Pittol, Ricardo Westhauser, M. Kolberg, R. Maffei, Edson Prestes e Silva
{"title":"Map Point Optimization in Keyframe-Based SLAM using Covisibility Graph and Information Fusion","authors":"Edison Kleiber Titito Concha, Diego Pittol, Ricardo Westhauser, M. Kolberg, R. Maffei, Edson Prestes e Silva","doi":"10.1109/ICAR46387.2019.8981653","DOIUrl":"https://doi.org/10.1109/ICAR46387.2019.8981653","url":null,"abstract":"Keyframe-based monocular SLAM (Simultaneous Localization and Mapping) is one of the main visual SLAM approaches, used to estimate the camera motion together with the map reconstruction over selected frames. These techniques represent the environment by map points located in the three-dimensional space, that can be recognized and located in the frame. However, these techniques usually cannot decide when a map point is an outlier or obsolete information and can be discarded. Another problem is to decide when combining map points corresponding to the same three-dimensional point. In this paper, we present a robust method to maintain a refined map. This approach uses the covisibility graph and an algorithm based on information fusion to build a probabilistic map, that explicitly models outlier measurements. In addition, we incorporate a pruning mechanism to reduce redundant information and remove outliers. In this way, our approach manages to reduce the map size maintaining essential information of the environment. Finally, in order to evaluate the performance of our method, we incorporate it into an ORB-SLAM system and measure the accuracy achieved on publicly available benchmark datasets which contain indoor images sequences recorded with a hand-held monocular camera.","PeriodicalId":6606,"journal":{"name":"2019 19th International Conference on Advanced Robotics (ICAR)","volume":"24 1","pages":"129-134"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88606538","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Fast Convolutional Neural Network for Real-Time Robotic Grasp Detection","authors":"E. G. Ribeiro, V. Grassi","doi":"10.1109/ICAR46387.2019.8981651","DOIUrl":"https://doi.org/10.1109/ICAR46387.2019.8981651","url":null,"abstract":"The development of the robotics field has not yet allowed robots to execute, with dexterity, simple actions performed by humans. One of them is the grasping of objects by robotic manipulators. Aiming to explore the use of deep learning algorithms, specifically Convolutional Neural Networks (CNN), to approach the robotic grasping problem, this work addresses the visual perception phase involved in the task. To achieve this goal, the dataset “Cornell Grasp” was used to train a CNN capable of predicting the most suitable place to grasp the object. It does this by obtaining a grasping rectangle that symbolizes the position, orientation, and opening of the robot's parallel grippers just before the grippers are closed. The proposed system works in real-time due to the small number of network parameters. This is possible by means of the data augmentation strategy used. The efficiency of the detection is in accordance with the state of the art and the speed of prediction, to the best of our knowledge, is the highest in the literature.","PeriodicalId":6606,"journal":{"name":"2019 19th International Conference on Advanced Robotics (ICAR)","volume":"129 1","pages":"49-54"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75266096","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
J. C. Vendrichoski, T. L. Costa, E. S. Elyoussef, E. D. Pieri
{"title":"Mathematical modeling and control of a quadrotor aerial vehicle with tiltrotors aimed for interaction tasks","authors":"J. C. Vendrichoski, T. L. Costa, E. S. Elyoussef, E. D. Pieri","doi":"10.1109/ICAR46387.2019.8981636","DOIUrl":"https://doi.org/10.1109/ICAR46387.2019.8981636","url":null,"abstract":"With a wide spectrum of applications ranging from entertainment to military use, quadrotors, in their conventional construction with four fixed rotors, have proven to be sufficiently capable of robustly performing numerous tasks. However, the capability to generate thrust just in one direction, i.e., normal to its main plane, is very restrictive and drastically reduces the maneuverability and agility of the vehicle. In this paper, an alternative model of the quadrotor is presented. Constructively, the difference lies in the addition of a mechanism that tilts the rotors in the longitudinal direction, which in practice adds maneuverability by enabling the longitudinal translation uncoupled from the pitching movement. In addition to this new thrust component in the longitudinal direction, this configuration also yields a significant increase in the yaw torque. These are highly desired features in a UAV used to execute tasks that require physical interaction with the surrounding environment. The mathematical model of the entire system is obtained by employing the Euler-Lagrange formalism and a multi-body approach. In addition, a basic control scheme is used to verify, through simulation, the obtained model.","PeriodicalId":6606,"journal":{"name":"2019 19th International Conference on Advanced Robotics (ICAR)","volume":"38 1","pages":"161-166"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79388832","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
N. Dhanaraj, Julietta Maffeo, G. Pereira, Jason N. Gross, Yu Gu, Nathan Hewitt, Casey Edmonds-Estes, Rachel Jarman, Jeongwoo Seo, Henry Gunner, Alexandra Hatfield, Tucker Johnson, Lunet Yifru
{"title":"Adaptable Platform for Interactive Swarm Robotics (APIS): A Human-Swarm Interaction Research Testbed","authors":"N. Dhanaraj, Julietta Maffeo, G. Pereira, Jason N. Gross, Yu Gu, Nathan Hewitt, Casey Edmonds-Estes, Rachel Jarman, Jeongwoo Seo, Henry Gunner, Alexandra Hatfield, Tucker Johnson, Lunet Yifru","doi":"10.1109/ICAR46387.2019.8981628","DOIUrl":"https://doi.org/10.1109/ICAR46387.2019.8981628","url":null,"abstract":"This paper describes the Adaptable Platform for Interactive Swarm robotics (APIS) - a testbed designed to accelerate development in human-swarm interaction (HSI) research. Specifically, this paper presents the design of a swarm robot platform composed of fifty low cost robots coupled with a testing field and a software architecture that allows for modular and versatile development of swarm algorithms. The motivation behind developing this platform is that the emergence of a swarm's collective behavior can be difficult to predict and control. However, human-swarm interaction can measurably increase a swarm's performance as the human operator may have intuition or knowledge unavailable to the swarm. The development of APIS allows researchers to focus on HSI research, without being constrained to a fixed ruleset or interface. A short survey is presented that offers a taxonomy of swarm platforms and provides conclusions that contextualize the development of APIS. Next, the motivations, design and functionality of the APIS testbed are described. Finally, the operation and potential of the platform are demonstrated through two experimental evaluations.","PeriodicalId":6606,"journal":{"name":"2019 19th International Conference on Advanced Robotics (ICAR)","volume":"31 1","pages":"720-726"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75015681","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Kyle Lindgren, Sarah Leung, W. Nothwang, E. J. Shamwell
{"title":"BooM-Vio: Bootstrapped Monocular Visual-Inertial Odometry with Absolute Trajectory Estimation through Unsupervised Deep Learning","authors":"Kyle Lindgren, Sarah Leung, W. Nothwang, E. J. Shamwell","doi":"10.1109/ICAR46387.2019.8981570","DOIUrl":"https://doi.org/10.1109/ICAR46387.2019.8981570","url":null,"abstract":"Machine learning has emerged as an extraordinary tool for solving many computer vision tasks by extracting and correlating meaningful features from high dimensional inputs in ways that often exceed the best human-derived modeling efforts. However, the area of vision-aided localization remains diverse with many traditional, model-based approaches (i.e. filtering- or nonlinear least- squares- based) often outperforming deep, model-free approaches. In this work, we present Bootstrapped Monocular VIO (BooM), a scaled monocular visual-inertial odometry (VIO) solution that leverages the complex data association ability of model-free approaches with the ability to exploit known geometric dynamics with model-based approaches. Our end-to-end, unsupervised deep neural network simultaneously learns to perform visual-inertial odometry and estimate scene depth while scale is enforced through a loss signal computed from position change magnitude estimates from traditional methods. We evaluate our network against a state-of-the-art (SoA) approach on the KITTI driving dataset as well as a micro aerial vehicle (MAV) dataset that we collected in the AirSim simulation environment. We further demonstrate the benefits of our combined approach through robustness tests on degraded trajectories.","PeriodicalId":6606,"journal":{"name":"2019 19th International Conference on Advanced Robotics (ICAR)","volume":"71 1","pages":"516-522"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89708563","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yuchi Ishikawa, Haruya Ishikawa, S. Akizuki, Masaki Yamazaki, Y. Taniguchi, Y. Aoki
{"title":"Task-oriented Function Detection Based on Operational Tasks","authors":"Yuchi Ishikawa, Haruya Ishikawa, S. Akizuki, Masaki Yamazaki, Y. Taniguchi, Y. Aoki","doi":"10.1109/ICAR46387.2019.8981633","DOIUrl":"https://doi.org/10.1109/ICAR46387.2019.8981633","url":null,"abstract":"We propose novel representations for functions of an object, namely Task-oriented Function, which is improved upon the idea of Afforadance in the field of Robotics Vision. We also propose a convolutional neural network to detect task-oriented functions. This network takes as input an operational task as well as an RGB image and assign each pixel an appropriate label for every task. Task-oriented funciton makes it possible to descibe various ways to use an object because the outputs from the network differ depending on operational tasks. We introduce a new dataset for task-oriented function detection, which contains about 1200 RGB images and 6000 pixel-level annotations assuming five tasks. Our proposed method reached 0.80 mean IOU in our dataset.","PeriodicalId":6606,"journal":{"name":"2019 19th International Conference on Advanced Robotics (ICAR)","volume":"36 1","pages":"635-640"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89736537","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
J. Smet, E. V. Poorten, V. Poliakov, Kenan Niu, Frédérique Chesterman, J. Fornier, M. Ahmad, M. Ourak, Viktor Vörös, J. Deprest
{"title":"Evaluating the Potential Benefit of Autostereoscopy in Laparoscopic Sacrocolpopexy through VR Simulation","authors":"J. Smet, E. V. Poorten, V. Poliakov, Kenan Niu, Frédérique Chesterman, J. Fornier, M. Ahmad, M. Ourak, Viktor Vörös, J. Deprest","doi":"10.1109/ICAR46387.2019.8981553","DOIUrl":"https://doi.org/10.1109/ICAR46387.2019.8981553","url":null,"abstract":"During laparoscopic sacrocolpopexy, pelvic organ prolapse is repaired by suturing one side of a synthetic mesh around the vaginal vault while stapling the other end to the sacrum, restoring the anatomical position of the vagina. A perineal assistant positions and tensions the vault with a vaginal manipulator instrument to properly expose the vaginal tissue to the laparoscopic surgeon. A technical difficulty during this surgery is the loss of depth perception due to visualization of the patient's internals on a 2D screen. Especially during precise surgical tasks, a more natural way to understand the distance between the laparoscopic instruments and the surgical region of interest could be advantageous. This work describes an exploratory study to investigate the potential of introducing 3D visualization into this surgical intervention. More in particular, experimentation is conducted with autostereoscopic display technology. A mixed reality setup was constructed featuring a virtual reality model of the vagina, 2D and 3D visualization, a physical interface representing the tissue of the body wall and a tracking system to track instrument motion. An experiment was conducted whereby the participants had to navigate the instrument to a number of pre-defined locations under 2D or 3D visualization. Compared to 2D, a considerable reduction in average task time (-42.9 %), travelled path lenght (-31.8 %) and errors (-52.2 %) was observed when performing the experiment in 3D. Where this work demonstrated a potential benefit of autostereoscopic visualization with respect to 2D visualization, in future work we wish to investigate if there also exists a benefit when comparing this technology with conventional stereoscopic visualization and whether stereoscopy can be used for (semi-) automated guidance during robotic laparoscopy.","PeriodicalId":6606,"journal":{"name":"2019 19th International Conference on Advanced Robotics (ICAR)","volume":"26 1","pages":"566-571"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91029830","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}