相机陷阱图像中动物距离估计的摄影测量方法

IF 5.8 2区 环境科学与生态学 Q1 ECOLOGY
B. Mirka , C.D. Lippitt , G.M. Harris , R. Converse , M. Gurule , S.E. Sesnie , M.J. Butler , D.R. Stewart , Z. Rossman
{"title":"相机陷阱图像中动物距离估计的摄影测量方法","authors":"B. Mirka ,&nbsp;C.D. Lippitt ,&nbsp;G.M. Harris ,&nbsp;R. Converse ,&nbsp;M. Gurule ,&nbsp;S.E. Sesnie ,&nbsp;M.J. Butler ,&nbsp;D.R. Stewart ,&nbsp;Z. Rossman","doi":"10.1016/j.ecoinf.2025.103120","DOIUrl":null,"url":null,"abstract":"<div><div>The ability to assess accurate distance measurements, meaning the distance from an animal to the observer, is critical for wildlife monitoring methods such as distance sampling. This process has typically been done through manual measurements or using reference points, making fieldwork and analysis labor intensive and error prone. Using Structure from Motion (SfM) photogrammetry and geotagged imagery captured from smartphones offers a simple and cost-effective solution to distance estimation from camera traps. This research evaluates the potential of SfM to create 3D models of the areas within a camera trap field-of-view to produce depth maps that can be used to automate the distance estimation process from bounding box-labeled imagery. These methods have the potential, when paired with deep learning algorithms or crowdsourcing, to automate the process of wildlife identification and distance estimation, reducing the need for manual fieldwork or analysis while increasing the precision and replicability of population estimates. To test the accuracy of these methods and evaluate best practices, three types of GNSS integration into the SfM model were assessed at three camera trap sites placed within the Sevilleta National Wildlife Refuge in New Mexico, USA. The evaluation focused on the geometric accuracy of the resulting model, quantified using Root Mean Square Error (RMSE) against known ground control points, the accuracy of distance estimates relative to field-measured distances, and the comparison of modeled distances of detected animals to ground control points. The first method used only the GNSS data embedded (geotagged) cell phone images (ImageGNSS), the second used ground control points (GCPGNSS), and the third used both GNSS sources (AllGNSS). Models created using ImageGNSS had the highest accuracy, with an overall RMSE of 1.25 pixels and a mean absolute error (MAE) of 1.76 m for distance estimation from GCPs. A comparison of the modeled distance for 10 animals to known distance points produced minimal error (MAE of 3.31 m), demonstrating the potential of photogrammetric approaches to make accurate distance estimations from camera trap imagery. Unlike other digital distance estimation techniques, error rates do not increase significantly (<em>p</em> = 0.36) at distances greater than 15 m to at least a distance of 35 m. The biggest driver of distance error was scene complexity (complex topography, dense vegetation, etc.)</div></div>","PeriodicalId":51024,"journal":{"name":"Ecological Informatics","volume":"89 ","pages":"Article 103120"},"PeriodicalIF":5.8000,"publicationDate":"2025-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A photogrammetric approach to the estimation of distance to animals in camera trap images\",\"authors\":\"B. Mirka ,&nbsp;C.D. Lippitt ,&nbsp;G.M. Harris ,&nbsp;R. Converse ,&nbsp;M. Gurule ,&nbsp;S.E. Sesnie ,&nbsp;M.J. Butler ,&nbsp;D.R. Stewart ,&nbsp;Z. Rossman\",\"doi\":\"10.1016/j.ecoinf.2025.103120\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>The ability to assess accurate distance measurements, meaning the distance from an animal to the observer, is critical for wildlife monitoring methods such as distance sampling. This process has typically been done through manual measurements or using reference points, making fieldwork and analysis labor intensive and error prone. Using Structure from Motion (SfM) photogrammetry and geotagged imagery captured from smartphones offers a simple and cost-effective solution to distance estimation from camera traps. This research evaluates the potential of SfM to create 3D models of the areas within a camera trap field-of-view to produce depth maps that can be used to automate the distance estimation process from bounding box-labeled imagery. These methods have the potential, when paired with deep learning algorithms or crowdsourcing, to automate the process of wildlife identification and distance estimation, reducing the need for manual fieldwork or analysis while increasing the precision and replicability of population estimates. To test the accuracy of these methods and evaluate best practices, three types of GNSS integration into the SfM model were assessed at three camera trap sites placed within the Sevilleta National Wildlife Refuge in New Mexico, USA. The evaluation focused on the geometric accuracy of the resulting model, quantified using Root Mean Square Error (RMSE) against known ground control points, the accuracy of distance estimates relative to field-measured distances, and the comparison of modeled distances of detected animals to ground control points. The first method used only the GNSS data embedded (geotagged) cell phone images (ImageGNSS), the second used ground control points (GCPGNSS), and the third used both GNSS sources (AllGNSS). Models created using ImageGNSS had the highest accuracy, with an overall RMSE of 1.25 pixels and a mean absolute error (MAE) of 1.76 m for distance estimation from GCPs. A comparison of the modeled distance for 10 animals to known distance points produced minimal error (MAE of 3.31 m), demonstrating the potential of photogrammetric approaches to make accurate distance estimations from camera trap imagery. Unlike other digital distance estimation techniques, error rates do not increase significantly (<em>p</em> = 0.36) at distances greater than 15 m to at least a distance of 35 m. The biggest driver of distance error was scene complexity (complex topography, dense vegetation, etc.)</div></div>\",\"PeriodicalId\":51024,\"journal\":{\"name\":\"Ecological Informatics\",\"volume\":\"89 \",\"pages\":\"Article 103120\"},\"PeriodicalIF\":5.8000,\"publicationDate\":\"2025-03-26\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Ecological Informatics\",\"FirstCategoryId\":\"93\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1574954125001293\",\"RegionNum\":2,\"RegionCategory\":\"环境科学与生态学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ECOLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Ecological Informatics","FirstCategoryId":"93","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1574954125001293","RegionNum":2,"RegionCategory":"环境科学与生态学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ECOLOGY","Score":null,"Total":0}
引用次数: 0

摘要

评估准确距离测量的能力,即动物到观察者的距离,对于距离采样等野生动物监测方法至关重要。这个过程通常是通过人工测量或使用参考点来完成的,这使得现场工作和分析变得劳动密集型,而且容易出错。使用运动结构(SfM)摄影测量和从智能手机捕获的地理标记图像为相机陷阱的距离估计提供了一种简单且经济有效的解决方案。本研究评估了SfM在相机陷阱视场内创建区域3D模型以生成深度图的潜力,该深度图可用于从边界框标记图像中自动估计距离过程。当与深度学习算法或众包相结合时,这些方法具有自动化野生动物识别和距离估计过程的潜力,减少了手工实地工作或分析的需要,同时提高了种群估计的精度和可复制性。为了测试这些方法的准确性并评估最佳实践,在美国新墨西哥州塞维利亚国家野生动物保护区的三个相机陷阱站点对三种集成到SfM模型中的GNSS进行了评估。评估的重点是最终模型的几何精度,使用已知地面控制点的均方根误差(RMSE)进行量化,相对于现场测量距离的距离估计精度,以及检测到的动物与地面控制点的模型距离的比较。第一种方法仅使用嵌入(地理标记)手机图像的GNSS数据(ImageGNSS),第二种方法使用地面控制点(GCPGNSS),第三种方法使用两个GNSS源(AllGNSS)。使用ImageGNSS创建的模型具有最高的精度,总体RMSE为1.25像素,平均绝对误差(MAE)为1.76 m。将10只动物的模型距离与已知距离点进行比较,误差最小(MAE为3.31 m),证明了摄影测量方法在从相机陷阱图像中进行准确距离估计方面的潜力。与其他数字距离估计技术不同,在距离大于15米至至少35米时,错误率不会显著增加(p = 0.36)。距离误差的最大驱动因素是场景复杂性(复杂地形、茂密植被等)。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
A photogrammetric approach to the estimation of distance to animals in camera trap images
The ability to assess accurate distance measurements, meaning the distance from an animal to the observer, is critical for wildlife monitoring methods such as distance sampling. This process has typically been done through manual measurements or using reference points, making fieldwork and analysis labor intensive and error prone. Using Structure from Motion (SfM) photogrammetry and geotagged imagery captured from smartphones offers a simple and cost-effective solution to distance estimation from camera traps. This research evaluates the potential of SfM to create 3D models of the areas within a camera trap field-of-view to produce depth maps that can be used to automate the distance estimation process from bounding box-labeled imagery. These methods have the potential, when paired with deep learning algorithms or crowdsourcing, to automate the process of wildlife identification and distance estimation, reducing the need for manual fieldwork or analysis while increasing the precision and replicability of population estimates. To test the accuracy of these methods and evaluate best practices, three types of GNSS integration into the SfM model were assessed at three camera trap sites placed within the Sevilleta National Wildlife Refuge in New Mexico, USA. The evaluation focused on the geometric accuracy of the resulting model, quantified using Root Mean Square Error (RMSE) against known ground control points, the accuracy of distance estimates relative to field-measured distances, and the comparison of modeled distances of detected animals to ground control points. The first method used only the GNSS data embedded (geotagged) cell phone images (ImageGNSS), the second used ground control points (GCPGNSS), and the third used both GNSS sources (AllGNSS). Models created using ImageGNSS had the highest accuracy, with an overall RMSE of 1.25 pixels and a mean absolute error (MAE) of 1.76 m for distance estimation from GCPs. A comparison of the modeled distance for 10 animals to known distance points produced minimal error (MAE of 3.31 m), demonstrating the potential of photogrammetric approaches to make accurate distance estimations from camera trap imagery. Unlike other digital distance estimation techniques, error rates do not increase significantly (p = 0.36) at distances greater than 15 m to at least a distance of 35 m. The biggest driver of distance error was scene complexity (complex topography, dense vegetation, etc.)
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Ecological Informatics
Ecological Informatics 环境科学-生态学
CiteScore
8.30
自引率
11.80%
发文量
346
审稿时长
46 days
期刊介绍: The journal Ecological Informatics is devoted to the publication of high quality, peer-reviewed articles on all aspects of computational ecology, data science and biogeography. The scope of the journal takes into account the data-intensive nature of ecology, the growing capacity of information technology to access, harness and leverage complex data as well as the critical need for informing sustainable management in view of global environmental and climate change. The nature of the journal is interdisciplinary at the crossover between ecology and informatics. It focuses on novel concepts and techniques for image- and genome-based monitoring and interpretation, sensor- and multimedia-based data acquisition, internet-based data archiving and sharing, data assimilation, modelling and prediction of ecological data.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信