2015 IEEE International Symposium on Mixed and Augmented Reality最新文献

筛选
英文 中文
[POSTER] An Adaptive Augmented Reality Interface for Hand Based on Probabilistic Approach 一种基于概率方法的手部自适应增强现实界面
2015 IEEE International Symposium on Mixed and Augmented Reality Pub Date : 2015-09-29 DOI: 10.1109/ISMAR.2015.44
Jinki Jung, Hyeopwoo Lee, H. Yang
{"title":"[POSTER] An Adaptive Augmented Reality Interface for Hand Based on Probabilistic Approach","authors":"Jinki Jung, Hyeopwoo Lee, H. Yang","doi":"10.1109/ISMAR.2015.44","DOIUrl":"https://doi.org/10.1109/ISMAR.2015.44","url":null,"abstract":"In this paper we propose an adaptive Augmented Reality interface for hand gestures based on a probabilistic model. The proposed method provides an in-situ interface and the corresponding functionalities by recognizing a context of hand shape and gesture which requires the accurate recognition of static and dynamic hand states. We present an appearance-based hand feature representation that yields robustness against hand shape variations, and a feature extraction method based on the fingertip likelihood from a GMM model. Experimental results show that both context-sensitivity and accurate hand gesture recognition are achieved throughout the quantitative evaluation and its implementation as a three-in-one virtual interface.","PeriodicalId":240196,"journal":{"name":"2015 IEEE International Symposium on Mixed and Augmented Reality","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-09-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115765970","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
[POSTER] Marker Identification Using IR LEDs and RGB Color Descriptors [海报]利用红外发光二极管和RGB颜色描述符识别标记
2015 IEEE International Symposium on Mixed and Augmented Reality Pub Date : 2015-09-29 DOI: 10.1109/ISMAR.2015.30
G. Koutaki, Shodai Hirata, Hiromu Sato, K. Uchimura
{"title":"[POSTER] Marker Identification Using IR LEDs and RGB Color Descriptors","authors":"G. Koutaki, Shodai Hirata, Hiromu Sato, K. Uchimura","doi":"10.1109/ISMAR.2015.30","DOIUrl":"https://doi.org/10.1109/ISMAR.2015.30","url":null,"abstract":"In optical motion capture systems, it is difficult to correctly recognize markers based on their unique identifiers (IDs) in a single frame. In this paper, we propose two types of light-emitting diodes (LEDs) and cameras, infrared (IR) and RGB, in order to correctly detect and identify all markers tracking objects in a given system. To detect and estimate the three-dimensional (3D) position of the marker, we measure IR LEDs using IR stereo cameras. Furthermore, in order to identify each marker, we calculate and compare the RGB color descriptor in the vicinity of its center. Our system consists of general IR and RGB cameras, and is easy to extend by increasing the number of cameras. We implemented an IR/RGB LED marker circuit and constructed a simple motion capture system to test the effectiveness of our system. The results show that our system can detect the 3D positions and unique IDs of markers in one frame.","PeriodicalId":240196,"journal":{"name":"2015 IEEE International Symposium on Mixed and Augmented Reality","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-09-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117354851","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
[POSTER] Augmented Wire Routing Navigation for Wire Assembly [海报]用于导线装配的增强导线路由导航
2015 IEEE International Symposium on Mixed and Augmented Reality Pub Date : 2015-09-29 DOI: 10.1109/ISMAR.2015.28
M. Rice, Hong Huei Tay, Jamie Ng, C. Lim, S. Selvaraj, Ellick Wu
{"title":"[POSTER] Augmented Wire Routing Navigation for Wire Assembly","authors":"M. Rice, Hong Huei Tay, Jamie Ng, C. Lim, S. Selvaraj, Ellick Wu","doi":"10.1109/ISMAR.2015.28","DOIUrl":"https://doi.org/10.1109/ISMAR.2015.28","url":null,"abstract":"Within modern manufacturing, digital solutions are needed to optimize and aid shop floor processes. This includes user-centered technologies that can be appropriately integrated into factory environments to assist in the efficiency of manufacturing tasks. In this paper, we present a dynamic system to support the electrical wiring assembly of commercial aircraft. Specifically, we describe the system design, which aims to improve the productivity of factory operators through the integration of wearable and mobile solutions. An evaluation of the augmented reality component of our system using a pair of smart glasses is reported with 12 participants, as we describe important interaction issues in the ongoing development of this work.","PeriodicalId":240196,"journal":{"name":"2015 IEEE International Symposium on Mixed and Augmented Reality","volume":"57 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-09-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123505863","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Introducing Augmented Reality to Optical Coherence Tomography in Ophthalmic Microsurgery 在眼科显微外科光学相干断层扫描中引入增强现实技术
2015 IEEE International Symposium on Mixed and Augmented Reality Pub Date : 2015-09-29 DOI: 10.1109/ISMAR.2015.15
Hessam Roodaki, K. Filippatos, A. Eslami, Nassir Navab
{"title":"Introducing Augmented Reality to Optical Coherence Tomography in Ophthalmic Microsurgery","authors":"Hessam Roodaki, K. Filippatos, A. Eslami, Nassir Navab","doi":"10.1109/ISMAR.2015.15","DOIUrl":"https://doi.org/10.1109/ISMAR.2015.15","url":null,"abstract":"Augmented Reality (AR) in microscopic surgery has been subject of several studies in the past two decades. Nevertheless, AR has not found its way into everyday microsurgical workflows. The introduction of new surgical microscopes equipped with Optical Coherence Tomography (OCT) enables the surgeons to perform multimodal (optical and OCT) imaging in the operating room. Taking full advantage of such elaborate source of information requires sophisticated intraoperative image fusion, information extraction, guidance and visualization methods. Medical AR is a unique approach to facilitate utilization of multimodal medical imaging devices. Here we propose a novel medical AR solution to the long-known problem of determining the distance between the surgical instrument tip and the underlying tissue in ophthalmic surgery to further pave the way of AR into the surgical theater. Our method brings augmented reality to OCT for the first time by augmenting the surgeon's view of the OCT images with an estimated instrument cross-section shape and distance to the retinal surface using only information from the shadow of the instrument in intraoperative OCT images. We demonstrate the applicability of our method in retinal surgery using a phantom eye and evaluate the accuracy of the augmented information using a micromanipulator.","PeriodicalId":240196,"journal":{"name":"2015 IEEE International Symposium on Mixed and Augmented Reality","volume":"54 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-09-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123643750","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
[POSTER] Natural 3D Interaction Using a See-Through Mobile AR System [海报]使用透明移动AR系统的自然3D交互
2015 IEEE International Symposium on Mixed and Augmented Reality Pub Date : 2015-09-29 DOI: 10.1109/ISMAR.2015.27
Yuko Unuma, T. Komuro
{"title":"[POSTER] Natural 3D Interaction Using a See-Through Mobile AR System","authors":"Yuko Unuma, T. Komuro","doi":"10.1109/ISMAR.2015.27","DOIUrl":"https://doi.org/10.1109/ISMAR.2015.27","url":null,"abstract":"In this paper, we propose an interaction system in which the appearance of the image displayed on a mobile display is consistent with that of the real space and that enables a user to interact with virtual objects overlaid on the image using the user's hand. The three-dimensional scene obtained by a depth camera is projected according to the user's viewpoint position obtained by face tracking, and the see-through image whose appearance is consistent with that outside the mobile display is generated. Interaction with virtual objects is realized by using the depth information obtained by the depth camera. To move virtual objects as if they were in real space, virtual objects are rendered in the world coordinate system that is fixed to a real scene even if the mobile display moves, and the direction of gravitational force added to virtual objects is made consistent with that of the world coordinate system. The former is realized by using the ICP (Iterative Closest Point) algorithm and the latter is realized by using the information obtained by an accelerometer. Thus, natural interaction with virtual objects using the user's hand is realized.","PeriodicalId":240196,"journal":{"name":"2015 IEEE International Symposium on Mixed and Augmented Reality","volume":"76 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-09-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128126037","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
[POSTER] On-site AR Interface with Web-Based 3D Archiving System for Archaeological Project [海报]考古项目现场AR界面与基于web的3D归档系统
2015 IEEE International Symposium on Mixed and Augmented Reality Pub Date : 2015-09-29 DOI: 10.1109/ISMAR.2015.68
Ryosuke Matsushita, Tokihisa Higo, H. Suita, Y. Yasumuro
{"title":"[POSTER] On-site AR Interface with Web-Based 3D Archiving System for Archaeological Project","authors":"Ryosuke Matsushita, Tokihisa Higo, H. Suita, Y. Yasumuro","doi":"10.1109/ISMAR.2015.68","DOIUrl":"https://doi.org/10.1109/ISMAR.2015.68","url":null,"abstract":"This paper proposes an AR (augmented reality) interface for on-site use in an archaeological project. We have already been developing a web-based 3D archiving system for supporting the diverse specialties and nationalities needed for carrying out the survey and restoration work of the archaeological project. Our 3D archiving system is designed for the spontaneous updating and sharing of information on findings in order to better enable frequent discussions, through a 3D virtual copy of the field site that a user can visit, explore, and embed information into, over the Internet. Here we present an AR user interface to enhance access from mobile devices at the actual site to the archiving system. Using SFM (structure from motion) and solving the Perspective-n-Point (PnP) problem, a photo taken at the site can be stably matched to the pre-registered photo sets in the archiving system and the archived information is automatically overlaid on the photo, just in the same manner as exploring the virtual version of the site on desktop PCs. Our implementation effectively works on an on-going project in Saqqara, Egypt.","PeriodicalId":240196,"journal":{"name":"2015 IEEE International Symposium on Mixed and Augmented Reality","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-09-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121713015","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
[POSTER] Authoring Tools in Augmented Reality: An Analysis and Classification of Content Design Tools 增强现实中的创作工具:内容设计工具的分析与分类
2015 IEEE International Symposium on Mixed and Augmented Reality Pub Date : 2015-09-29 DOI: 10.1109/ISMAR.2015.47
Roberta Cabral Mota, R. Roberto, V. Teichrieb
{"title":"[POSTER] Authoring Tools in Augmented Reality: An Analysis and Classification of Content Design Tools","authors":"Roberta Cabral Mota, R. Roberto, V. Teichrieb","doi":"10.1109/ISMAR.2015.47","DOIUrl":"https://doi.org/10.1109/ISMAR.2015.47","url":null,"abstract":"Augmented Reality Authoring Tools are important instruments that can help a widespread use of AR. They can be classified as programming or content design tools in which the latter completely removes the necessity of programming skills to develop an AR solution. Several solutions have been developed in the past years, however there are few works aiming to identify patterns and general models for such tools. This work aims to perform a trend analysis on content design tools in order to identify their functionalities regarding AR, authoring paradigms, deployment strategies and general dataflow models. This work is aimed to assist developers willing to create authoring tools, therefore, it focus on the last three aspects. Thus, 19 tools were analyzed and through this evaluation it were identified two authoring paradigms and two deployment strategies. Moreover, from their combination it was possible to elaborate four generic dataflow models in which every tool could be fit into.","PeriodicalId":240196,"journal":{"name":"2015 IEEE International Symposium on Mixed and Augmented Reality","volume":"90 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-09-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131982832","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
[POSTER] Road Maintenance MR System Using LRF and PDR [海报]基于LRF和PDR的道路养护MR系统
2015 IEEE International Symposium on Mixed and Augmented Reality Pub Date : 2015-09-29 DOI: 10.1109/ISMAR.2015.66
Ching-Tzun Chang, Ryosuke Ichikari, Koji Makita, T. Okuma, T. Kurata
{"title":"[POSTER] Road Maintenance MR System Using LRF and PDR","authors":"Ching-Tzun Chang, Ryosuke Ichikari, Koji Makita, T. Okuma, T. Kurata","doi":"10.1109/ISMAR.2015.66","DOIUrl":"https://doi.org/10.1109/ISMAR.2015.66","url":null,"abstract":"We have been developing a mixed reality system to support road maintenance using overlaid visual aids. Such a system requires a positioning method that can provide sub-meter accuracy and function even if the appearance of the road surface changes significantly caused by many factors such as construction phase, time and weather. Therefore, we are developing a real-time worker positioning method that can be applied to these situation by integrating laser range finder (LRF) and pedestrian dead-reckoning (PDR) data. In the field, multiple workers move around the workspace. Therefore, it is necessary to determine corresponding pairs of PDR-based and LRF-based trajectories by identifying similar trajectories. In this study, we propose a method to calculate the similarity between trajectories and a procedure to integrate corresponding pairs of trajectories to acquire the position and movement direction of a worker.","PeriodicalId":240196,"journal":{"name":"2015 IEEE International Symposium on Mixed and Augmented Reality","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-09-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131539238","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
[POSTER] Natural User Interface for Ambient Objects [海报]环境对象的自然用户界面
2015 IEEE International Symposium on Mixed and Augmented Reality Pub Date : 2015-09-29 DOI: 10.1109/ISMAR.2015.25
Meng Ma, Kevin Merckx, P. Fallavollita, Nassir Navab
{"title":"[POSTER] Natural User Interface for Ambient Objects","authors":"Meng Ma, Kevin Merckx, P. Fallavollita, Nassir Navab","doi":"10.1109/ISMAR.2015.25","DOIUrl":"https://doi.org/10.1109/ISMAR.2015.25","url":null,"abstract":"To help the computing device always understand the spacial relationship between the user's gesture and the ambient objects, a methodology is proposed to find the user's virtual eye center in the wearable camera coordinate system and then calculate accurately where a user is pointing at to perform the natural interaction. First, the wearable RGB-D sensor is affixed around the user forehead. A tool-free calibration is done by having the user move their fingers along their lines of sight from his eye center to the random selected targets. The fingertips are detected in the depth camera and then the interaction of these lines of sight is calculated. Then we present how to find where the user is pointing at in different scenarios with a depth map, a detected object and a controlled virtual element. To validate our methods, we perform a point-to-screen experiment. Results demonstrate that when a user is interacting with a display up to 1.5 meters away, our natural gesture interface has an average error of 2.1cm. In conclusion, the presented technique is a viable option for a reliable user interaction.","PeriodicalId":240196,"journal":{"name":"2015 IEEE International Symposium on Mixed and Augmented Reality","volume":"199 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-09-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115685181","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A Framework to Evaluate Omnidirectional Video Coding Schemes 一个评价全向视频编码方案的框架
2015 IEEE International Symposium on Mixed and Augmented Reality Pub Date : 2015-09-29 DOI: 10.1109/ISMAR.2015.12
Matt C. Yu, H. Lakshman, B. Girod
{"title":"A Framework to Evaluate Omnidirectional Video Coding Schemes","authors":"Matt C. Yu, H. Lakshman, B. Girod","doi":"10.1109/ISMAR.2015.12","DOIUrl":"https://doi.org/10.1109/ISMAR.2015.12","url":null,"abstract":"Omnidirectional videos of real world environments viewed on head-mounted displays with real-time head motion tracking can offer immersive visual experiences. For live streaming applications, compression is critical to reduce the bitrate. Omnidirectional videos, which are spherical in nature, are mapped onto one or more planes before encoding to interface with modern video coding standards. In this paper, we consider the problem of evaluating the coding efficiency in the context of viewing with a head-mounted display. We extract viewport based head motion trajectories, and compare the original and coded videos on the viewport. With this approach, we compare different sphere-to-plane mappings. We show that the average viewport quality can be approximated by a weighted spherical PSNR.","PeriodicalId":240196,"journal":{"name":"2015 IEEE International Symposium on Mixed and Augmented Reality","volume":"216 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-09-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115067818","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 344
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信