2018 Second IEEE International Conference on Robotic Computing (IRC)最新文献

筛选
英文 中文
A Pleliminary Study on Human Chewing Action Counter 人体咀嚼动作计数器的初步研究
2018 Second IEEE International Conference on Robotic Computing (IRC) Pub Date : 1900-01-01 DOI: 10.1109/IRC.2018.00070
Hyun-Mo Yang, Y. Son, Young-One Cho, Jin-Woo Jung
{"title":"A Pleliminary Study on Human Chewing Action Counter","authors":"Hyun-Mo Yang, Y. Son, Young-One Cho, Jin-Woo Jung","doi":"10.1109/IRC.2018.00070","DOIUrl":"https://doi.org/10.1109/IRC.2018.00070","url":null,"abstract":"This paper deals with a novel method which can estimate the occurrence number of human chewing actions by the help of image processing technique. At first, the user's mouth is recognized by the help of Haar cascade classifiers for human face and mouth. And then, this mouth image is processed with our proposed algorithm which can counter the occurrence number of human chewing action and can also reset the counter by confirming the mouth openness for new meal consumption. The experimental results show that it can be applied to improve chewing habits for kids.","PeriodicalId":416113,"journal":{"name":"2018 Second IEEE International Conference on Robotic Computing (IRC)","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131239523","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Collaborative Goal Distribution in Distributed Multiagent Systems 分布式多智能体系统中的协同目标分配
2018 Second IEEE International Conference on Robotic Computing (IRC) Pub Date : 1900-01-01 DOI: 10.1109/IRC.2018.00066
Sujin Park, Sang-Gyu Park, Hyeonggun Lee, Minji Hyun, Eunsuh Lee, Jeonghyeon Ahn, Lauren Featherstun, Yongho Kim, E. Matson
{"title":"Collaborative Goal Distribution in Distributed Multiagent Systems","authors":"Sujin Park, Sang-Gyu Park, Hyeonggun Lee, Minji Hyun, Eunsuh Lee, Jeonghyeon Ahn, Lauren Featherstun, Yongho Kim, E. Matson","doi":"10.1109/IRC.2018.00066","DOIUrl":"https://doi.org/10.1109/IRC.2018.00066","url":null,"abstract":"Distributed multiagent systems consist of multiple agents which perform related tasks. In this kind of system, the tasks are distributed amongst the agents by an operator based on shared information. The information used to assign tasks includes not only agent's capability, but also agent's state, the goal's state, and conditions from the surrounding environments. Distributed multi agent systems are usually constrained by uncertain information about nearby agents, and by limited network availability to transfer information to the operator. Given these constraints of using an operator, a better designed system might allow agents to distribute tasks on their own. This paper proposes a goal distribution strategy for collaborative distributed multi agent systems where agents distribute tasks amongst themselves. In this strategy, a goal model is shared amongst all participating agents, enabling them to synchronize in order to achieve complex goals that require sequential executions. Agents in this system are capable of transferring information over the network where all others belong to. The approach was tested and verified using StarCraft II APIs, introduced by Blizzard and Google Deepmind.","PeriodicalId":416113,"journal":{"name":"2018 Second IEEE International Conference on Robotic Computing (IRC)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132021435","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Environment-Dependent Depth Enhancement with Multi-modal Sensor Fusion Learning 基于多模态传感器融合学习的环境依赖深度增强
2018 Second IEEE International Conference on Robotic Computing (IRC) Pub Date : 1900-01-01 DOI: 10.1109/IRC.2018.00049
Kuya Takami, Taeyoung Lee
{"title":"Environment-Dependent Depth Enhancement with Multi-modal Sensor Fusion Learning","authors":"Kuya Takami, Taeyoung Lee","doi":"10.1109/IRC.2018.00049","DOIUrl":"https://doi.org/10.1109/IRC.2018.00049","url":null,"abstract":"This paper presents a new learning based multimodal sensing paradigm within a probabilistic framework to improve the depth image measurements of an RGB-D camera. The proposed approach uses an RGB-D camera and laser range finder to provide an improved depth image using convolutional neural network (CNN) approximation within a probabilistic inference framework. Synchronized RGB-D and laser measurements are collected in an environment to train a model, which is then used for depth image accuracy improvements and sensor range extension. The model exploits additional RGB information, which contains depth cues, to enhance the accuracy of pixel level measurements. A computationally efficient implementation of the CNN allows the model to train while exploring an unknown area to provide improved depth image measurements. The approach yields depth images containing spatial information far beyond the suggested operational limits. We demonstrate a nearly three-fold depth range extension (3:5m to 10m) while maintaining similar camera accuracy at the maximum range. The mean absolute error is also reduced from the original depth image by a factor of six. The efficacy of this approach is demonstrated in an unstructured office space.","PeriodicalId":416113,"journal":{"name":"2018 Second IEEE International Conference on Robotic Computing (IRC)","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114290498","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Autonomous Quadrotor 3D Mapping and Exploration Using Exact Occupancy Probabilities 自主四旋翼三维映射和探索使用精确的占用概率
2018 Second IEEE International Conference on Robotic Computing (IRC) Pub Date : 1900-01-01 DOI: 10.1109/IRC.2018.00016
Evan Kaufman, Kuya Takami, Zhuming Ai, Taeyoung Lee
{"title":"Autonomous Quadrotor 3D Mapping and Exploration Using Exact Occupancy Probabilities","authors":"Evan Kaufman, Kuya Takami, Zhuming Ai, Taeyoung Lee","doi":"10.1109/IRC.2018.00016","DOIUrl":"https://doi.org/10.1109/IRC.2018.00016","url":null,"abstract":"This paper deals with the aerial exploration for an unknown three-dimensional environment, where Bayesian probabilistic mapping is integrated with a stochastic motion planning scheme to minimize the map uncertainties in an optimal fashion. We utilize the popular occupancy grid mapping representation, with the goal of determining occupancy probabilities of evenly-spaced grid cells in 3D with sensor fusion from multiple depth sensors with realistic sensor capabilities. The 3D exploration problem is decomposed into 3D mapping and 2D motion planning for efficient real-time implementation. This is achieved by projecting important aspects of the 3D map onto 2D maps, where a predicted level of map uncertainty, known as Shannon's entropy, provides an exploration policy that governs robotic motion. Both mapping and exploration algorithms are demonstrated with both numerical simulations and quadrotor flight experiments.","PeriodicalId":416113,"journal":{"name":"2018 Second IEEE International Conference on Robotic Computing (IRC)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122207717","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Open-Finger: Mobile Application Platform Enhanced by Physical Finger Open-Finger:物理手指增强的移动应用平台
2018 Second IEEE International Conference on Robotic Computing (IRC) Pub Date : 1900-01-01 DOI: 10.1109/IRC.2018.00041
Hiroaki Tobita, Hirotaka Saitoh
{"title":"Open-Finger: Mobile Application Platform Enhanced by Physical Finger","authors":"Hiroaki Tobita, Hirotaka Saitoh","doi":"10.1109/IRC.2018.00041","DOIUrl":"https://doi.org/10.1109/IRC.2018.00041","url":null,"abstract":"We introduce our Open-Finger that integrates the smartphone with a physical finger. Smartphones are widely used for communication and entertainment, and have characteristic features such an even surface and a few buttons. Our interaction with them is quite simple and really limited. In contrast, we have found a way to use a physical finger attached to a smartphone. A real finger has many capabilities such as pointing and touching. For example, we use our finger to point at something or someone, to move something, or to count a number. We can also use such features for interactions between us and our smartphones. Thus, the finger approach makes smartphone more intuitive and familiar for novice and elderly users who are not good at manipulating smartphone. In this paper, we describe our design concepts, prototype implementation and application possibilities.","PeriodicalId":416113,"journal":{"name":"2018 Second IEEE International Conference on Robotic Computing (IRC)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122706589","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Towards a Well-Founded Software Component Model for Cyber-Physical Control Systems 面向网络物理控制系统的良好软件组件模型
2018 Second IEEE International Conference on Robotic Computing (IRC) Pub Date : 1900-01-01 DOI: 10.1109/IRC.2018.00055
J. Malenfant
{"title":"Towards a Well-Founded Software Component Model for Cyber-Physical Control Systems","authors":"J. Malenfant","doi":"10.1109/IRC.2018.00055","DOIUrl":"https://doi.org/10.1109/IRC.2018.00055","url":null,"abstract":"Cyber-physical control systems (CPCS), and their instantiation as autonomous robotic control architectures, are notoriously difficult to specify, implement, test, validate and verify. In this paper, we propose to integrate hybrid systems and their declension as hybrid automata and DEVS simulation models within a full-fledged and well-founded software component model tailored for CPCS. We present how the resulting comprehensive modeling tool can support the different phases of the software development to provide more reliable, more robust and more adaptable CPCS. The key concept is to provide components with a modeling and simulation capability that seamlessly support the software development process, from model-in-the-loop initial validations, until deployment time actual system verification.","PeriodicalId":416113,"journal":{"name":"2018 Second IEEE International Conference on Robotic Computing (IRC)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132626457","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Collective Behavior Acquisition of Real Robotic Swarms Using Deep Reinforcement Learning 基于深度强化学习的真实机器人群体集体行为获取
2018 Second IEEE International Conference on Robotic Computing (IRC) Pub Date : 1900-01-01 DOI: 10.1109/IRC.2018.00038
T. Yasuda, K. Ohkura
{"title":"Collective Behavior Acquisition of Real Robotic Swarms Using Deep Reinforcement Learning","authors":"T. Yasuda, K. Ohkura","doi":"10.1109/IRC.2018.00038","DOIUrl":"https://doi.org/10.1109/IRC.2018.00038","url":null,"abstract":"Swarm robotic systems are a type of multi-robot systems, in which robots operate without any form of centralized control. The most popular approach for SRS is the so-called ad hoc or behavior-based approach; desired collective behavior is obtained by manually by designing the behavior of individual robot in advance. On the other hand, in the principled or automatic design approach, a certain general methodology for developing appropriate collective behavior is adopted. This paper investigates a deep reinforcement learning approach to collective behavior acquisition of swarm robotics systems. Robots are expected to collect information in parallel and share their experience for accelerating the learning. We conduct real swarm robot experiments and evaluate the learning performance in a scenario where robots consecutively travel between two landmarks.","PeriodicalId":416113,"journal":{"name":"2018 Second IEEE International Conference on Robotic Computing (IRC)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128047552","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Improving Student Surveys with Natural Language Processing 用自然语言处理改进学生调查
2018 Second IEEE International Conference on Robotic Computing (IRC) Pub Date : 1900-01-01 DOI: 10.1109/IRC.2018.00079
Karoline Hood, Patrick K. Kuiper
{"title":"Improving Student Surveys with Natural Language Processing","authors":"Karoline Hood, Patrick K. Kuiper","doi":"10.1109/IRC.2018.00079","DOIUrl":"https://doi.org/10.1109/IRC.2018.00079","url":null,"abstract":"Stakeholders from academic institutions across the world employ surveys to assess the quality of their work. With surveys these stakeholders attempt to obtain quantified, structured, and directed data in order to make decisions. Often these stakeholders employ long, directed Likert scaled surveys to gain this information. We propose an alternate construction for academic surveys, where stakeholders provide 1-3 open ended \"free text\" questions, allowing students to lead the discussion. We call this survey methodology \"Student Directed Discussion Surveys\" (SDDS). SDDS retain the ability to provide quantified, structured, and directed results by employing Natural Language Processing (NLP). We confirm the accuracy of SDDS in relation to traditional Likert scaled surveys with a permutation test, assessing a negligible statistical difference between SDDS and Likert surveys using real data. We then show the utility of SDDS by employing word frequency and sentiment analysis, providing important unbiased decision making information, which is limited when traditional Likert scaled surveys are administered.","PeriodicalId":416113,"journal":{"name":"2018 Second IEEE International Conference on Robotic Computing (IRC)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125587215","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Sound Identification for Fire-Fighting Mobile Robots 消防移动机器人的声音识别
2018 Second IEEE International Conference on Robotic Computing (IRC) Pub Date : 1900-01-01 DOI: 10.1109/IRC.2018.00020
Eli M. Baum, Mario Harper, Ryan Alicea, Camilo Ordonez
{"title":"Sound Identification for Fire-Fighting Mobile Robots","authors":"Eli M. Baum, Mario Harper, Ryan Alicea, Camilo Ordonez","doi":"10.1109/IRC.2018.00020","DOIUrl":"https://doi.org/10.1109/IRC.2018.00020","url":null,"abstract":"A structure engulfed in flames can pose an extreme danger for fire-fighting personnel as well as any people trapped inside. A companion robot to assist the fire-fighters could potentially help speed up the search for humans while reducing risk for the fire-fighters. However, robots operating in these environments need to be able to operate in very low visibility conditions because of the heavy smoke, debris and unstructured terrain. This paper develops an audio classification algorithm to identify sounds relevant to fire-fighting such as people in distress (baby cries, screams, coughs), structural failure (wood snapping, glass breaking), fire, fire trucks, and crowds. The outputs of the classifier are then used as alerts for the fire-fighter or to modify the configuration of a robot capable of navigating unstructured terrain. The approach used extracts an array of features from audio recordings and employs a single hidden layer, feed forward neural network for classification. The simplicity in network structure enables performance on limited hardware and obtains classification results with an overall accuracy of 85.7%.","PeriodicalId":416113,"journal":{"name":"2018 Second IEEE International Conference on Robotic Computing (IRC)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114988196","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Internet of Things: Technology to Enable the Elderly 物联网:为老年人服务的技术
2018 Second IEEE International Conference on Robotic Computing (IRC) Pub Date : 1900-01-01 DOI: 10.1109/IRC.2018.00075
Chan-Gun Lee, S. Park, Yoonha Jung, Youngji Lee, Mariah Mathews
{"title":"Internet of Things: Technology to Enable the Elderly","authors":"Chan-Gun Lee, S. Park, Yoonha Jung, Youngji Lee, Mariah Mathews","doi":"10.1109/IRC.2018.00075","DOIUrl":"https://doi.org/10.1109/IRC.2018.00075","url":null,"abstract":"The purpose of this project is to integrate IoT technology into the homes of the elderly that live alone using simple, inexpensive, accessible devices and open source software. Using technology such as Raspberry Pi (RPi), Open Source Computer Vision (OpenCV), and Node.js web server, actions can be controlled to supervise an unaccompanied elderly person. There are five services in this paper: opening the door via facial recognition with a servo motor, detecting motion and sending alarms to their family members, getting real-time indoor temperatures, remotely toggling the light switch on or off, and measuring the amount of trash in a selected trash bin. All functions are controlled by an Android application that can be customized depending on the specific visual needs of the user. This project proposes solutions to help the elderly benefit from user-friendly IoT technology. The solutions allow for notifications to be shared with family members, which can provide peace of mind.","PeriodicalId":416113,"journal":{"name":"2018 Second IEEE International Conference on Robotic Computing (IRC)","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124105057","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信