{"title":"用于自动估算建筑物最低层标高的综合视觉语言和基础模型","authors":"Yu‐Hsuan Ho, Longxiang Li, Ali Mostafavi","doi":"10.1111/mice.13310","DOIUrl":null,"url":null,"abstract":"Street view imagery has emerged as a valuable resource for urban analytics research. Recent studies have explored its potential for estimating lowest floor elevation (LFE), offering a scalable alternative to traditional on‐site measurements, crucial for assessing properties' flood risk and damage extent. While existing methods rely on object detection, the introduction of image segmentation has expanded the utility of street view images for LFE estimation, although challenges still remain in segmentation quality and capability to distinguish front doors from other doors. To address these challenges in LFE estimation, this study integrates the Segment Anything model, a segmentation foundation model, with vision language models (VLMs) to conduct text‐prompt image segmentation on street view images for LFE estimation. By evaluating various VLMs, integration methods, and text prompts, the most suitable model was identified for street view image analytics and LFE estimation tasks, thereby improving the coverage of the current LFE estimation model based on image segmentation from 33% to 56% of properties. Remarkably, our proposed method, ELEV‐VISION‐SAM, significantly enhances the availability of LFE estimation to almost all properties in which the front door is visible in the street view image. In addition, the findings present the first baseline and quantified comparison of various vision models for street view image‐based LFE estimation. The model and findings not only contribute to advancing street view image segmentation for urban analytics but also provide a novel approach for image segmentation tasks for other civil engineering and infrastructure analytics tasks.","PeriodicalId":156,"journal":{"name":"Computer-Aided Civil and Infrastructure Engineering","volume":"109 1","pages":""},"PeriodicalIF":8.5000,"publicationDate":"2024-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Integrated vision language and foundation model for automated estimation of building lowest floor elevation\",\"authors\":\"Yu‐Hsuan Ho, Longxiang Li, Ali Mostafavi\",\"doi\":\"10.1111/mice.13310\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Street view imagery has emerged as a valuable resource for urban analytics research. Recent studies have explored its potential for estimating lowest floor elevation (LFE), offering a scalable alternative to traditional on‐site measurements, crucial for assessing properties' flood risk and damage extent. While existing methods rely on object detection, the introduction of image segmentation has expanded the utility of street view images for LFE estimation, although challenges still remain in segmentation quality and capability to distinguish front doors from other doors. To address these challenges in LFE estimation, this study integrates the Segment Anything model, a segmentation foundation model, with vision language models (VLMs) to conduct text‐prompt image segmentation on street view images for LFE estimation. By evaluating various VLMs, integration methods, and text prompts, the most suitable model was identified for street view image analytics and LFE estimation tasks, thereby improving the coverage of the current LFE estimation model based on image segmentation from 33% to 56% of properties. Remarkably, our proposed method, ELEV‐VISION‐SAM, significantly enhances the availability of LFE estimation to almost all properties in which the front door is visible in the street view image. In addition, the findings present the first baseline and quantified comparison of various vision models for street view image‐based LFE estimation. The model and findings not only contribute to advancing street view image segmentation for urban analytics but also provide a novel approach for image segmentation tasks for other civil engineering and infrastructure analytics tasks.\",\"PeriodicalId\":156,\"journal\":{\"name\":\"Computer-Aided Civil and Infrastructure Engineering\",\"volume\":\"109 1\",\"pages\":\"\"},\"PeriodicalIF\":8.5000,\"publicationDate\":\"2024-07-26\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Computer-Aided Civil and Infrastructure Engineering\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://doi.org/10.1111/mice.13310\",\"RegionNum\":1,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computer-Aided Civil and Infrastructure Engineering","FirstCategoryId":"5","ListUrlMain":"https://doi.org/10.1111/mice.13310","RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS","Score":null,"Total":0}
Integrated vision language and foundation model for automated estimation of building lowest floor elevation
Street view imagery has emerged as a valuable resource for urban analytics research. Recent studies have explored its potential for estimating lowest floor elevation (LFE), offering a scalable alternative to traditional on‐site measurements, crucial for assessing properties' flood risk and damage extent. While existing methods rely on object detection, the introduction of image segmentation has expanded the utility of street view images for LFE estimation, although challenges still remain in segmentation quality and capability to distinguish front doors from other doors. To address these challenges in LFE estimation, this study integrates the Segment Anything model, a segmentation foundation model, with vision language models (VLMs) to conduct text‐prompt image segmentation on street view images for LFE estimation. By evaluating various VLMs, integration methods, and text prompts, the most suitable model was identified for street view image analytics and LFE estimation tasks, thereby improving the coverage of the current LFE estimation model based on image segmentation from 33% to 56% of properties. Remarkably, our proposed method, ELEV‐VISION‐SAM, significantly enhances the availability of LFE estimation to almost all properties in which the front door is visible in the street view image. In addition, the findings present the first baseline and quantified comparison of various vision models for street view image‐based LFE estimation. The model and findings not only contribute to advancing street view image segmentation for urban analytics but also provide a novel approach for image segmentation tasks for other civil engineering and infrastructure analytics tasks.
期刊介绍:
Computer-Aided Civil and Infrastructure Engineering stands as a scholarly, peer-reviewed archival journal, serving as a vital link between advancements in computer technology and civil and infrastructure engineering. The journal serves as a distinctive platform for the publication of original articles, spotlighting novel computational techniques and inventive applications of computers. Specifically, it concentrates on recent progress in computer and information technologies, fostering the development and application of emerging computing paradigms.
Encompassing a broad scope, the journal addresses bridge, construction, environmental, highway, geotechnical, structural, transportation, and water resources engineering. It extends its reach to the management of infrastructure systems, covering domains such as highways, bridges, pavements, airports, and utilities. The journal delves into areas like artificial intelligence, cognitive modeling, concurrent engineering, database management, distributed computing, evolutionary computing, fuzzy logic, genetic algorithms, geometric modeling, internet-based technologies, knowledge discovery and engineering, machine learning, mobile computing, multimedia technologies, networking, neural network computing, optimization and search, parallel processing, robotics, smart structures, software engineering, virtual reality, and visualization techniques.