HoloLabel: Augmented reality user-in-the-loop online annotation tool for as-is building information

Agrawal Dhruv, Lobsiger Janik, Bo Jessica Yi, Kaufmann Véronique, A. Iro
{"title":"HoloLabel: Augmented reality user-in-the-loop online annotation tool for as-is building information","authors":"Agrawal Dhruv, Lobsiger Janik, Bo Jessica Yi, Kaufmann Véronique, A. Iro","doi":"10.35490/ec3.2022.174","DOIUrl":null,"url":null,"abstract":"As-is building models are becoming increasingly common in the Architecture, Engineering, and Construction industry, with many stakeholders requesting this information throughout the lifecycle of a building. Devices equipped with RGB cameras and depth sensors being readily available simplifies the task of capturing and reconstructing an environment (scene) as a spatial 3D mesh or point cloud. However, the task of converting this purely geometric information to a semantically meaningful as-is building model is non-trivial. State-of-the-art practice follows a first step of acquiring the spatial 3D mesh on site and subsequently resorts in manual or assisted semantic labeling in the of-fice, where experts often have to work for many hours using non-intuitive and error-prone tools. To address this inef-ficiency, we develop HoloLabel , an Augmented Reality application on HoloLens that allows users to directly and on-site annotate a scene in 3D with rich semantic information while simultaneously capturing its spatial 3D mesh. Our tool follows a user-in-the-loop protocol to perform the task of 3D semantic segmentation, i.e., each face of the 3D mesh should be annotated with a semantic label. We leverage the HoloLens’s Spatial Mapping feature and build a 3D mesh of the scene while the user is walking around; at intervals, we apply an automatic geometry-based segmentation algorithm to generate segmentation proposals. The user then assigns predefined semantic labels to the proposals and – if necessary – uses a virtual paintbrush to refine the proposed segments or create new ones. Finally, the user has the option to add rich semantic descriptions (e.g., material, shape, or relationship to another object) to segments using voice-to-text technology. We aim to lay the groundwork to leverage upcoming mixed reality devices for intuitive synchronous as-is semantic building model generation directly in the real world.","PeriodicalId":142381,"journal":{"name":"Proceedings of the 2022 European Conference on Computing in Construction","volume":"5 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2022 European Conference on Computing in Construction","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.35490/ec3.2022.174","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

As-is building models are becoming increasingly common in the Architecture, Engineering, and Construction industry, with many stakeholders requesting this information throughout the lifecycle of a building. Devices equipped with RGB cameras and depth sensors being readily available simplifies the task of capturing and reconstructing an environment (scene) as a spatial 3D mesh or point cloud. However, the task of converting this purely geometric information to a semantically meaningful as-is building model is non-trivial. State-of-the-art practice follows a first step of acquiring the spatial 3D mesh on site and subsequently resorts in manual or assisted semantic labeling in the of-fice, where experts often have to work for many hours using non-intuitive and error-prone tools. To address this inef-ficiency, we develop HoloLabel , an Augmented Reality application on HoloLens that allows users to directly and on-site annotate a scene in 3D with rich semantic information while simultaneously capturing its spatial 3D mesh. Our tool follows a user-in-the-loop protocol to perform the task of 3D semantic segmentation, i.e., each face of the 3D mesh should be annotated with a semantic label. We leverage the HoloLens’s Spatial Mapping feature and build a 3D mesh of the scene while the user is walking around; at intervals, we apply an automatic geometry-based segmentation algorithm to generate segmentation proposals. The user then assigns predefined semantic labels to the proposals and – if necessary – uses a virtual paintbrush to refine the proposed segments or create new ones. Finally, the user has the option to add rich semantic descriptions (e.g., material, shape, or relationship to another object) to segments using voice-to-text technology. We aim to lay the groundwork to leverage upcoming mixed reality devices for intuitive synchronous as-is semantic building model generation directly in the real world.
HoloLabel:用于现有建筑信息的增强现实用户在线注释工具
原样建筑模型在体系结构、工程和建筑行业中变得越来越普遍,许多涉众在建筑的整个生命周期中都需要这些信息。配备有RGB相机和深度传感器的设备可以方便地简化捕捉和重建环境(场景)作为空间3D网格或点云的任务。然而,将这种纯粹的几何信息转换为语义上有意义的现有构建模型的任务并不简单。最先进的实践遵循在现场获取空间3D网格的第一步,随后在办公室中诉诸手动或辅助语义标记,专家通常必须使用非直观且容易出错的工具工作数小时。为了解决这个低效率问题,我们开发了HoloLabel,这是一款基于HoloLens的增强现实应用程序,允许用户直接在现场用丰富的语义信息注释3D场景,同时捕获其空间3D网格。我们的工具遵循用户在环协议来执行3D语义分割任务,即3D网格的每个面都应该用语义标签进行注释。我们利用HoloLens的空间映射功能,在用户走动时构建场景的3D网格;每隔一段时间,我们应用基于几何的自动分割算法来生成分割建议。然后,用户将预定义的语义标签分配给建议,如果有必要,可以使用虚拟画笔来改进建议的部分或创建新的部分。最后,用户可以选择使用语音到文本技术向片段添加丰富的语义描述(例如,材料,形状或与另一个对象的关系)。我们的目标是奠定基础,利用即将到来的混合现实设备,直接在现实世界中直观地同步生成语义构建模型。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信