María-José Zurita, Daniel Riofrío, Noel Pérez, David Romo, D. Benítez, Ricardo Flores Moyano, Felipe Grijalva, Maria Baldeon-Calisto
{"title":"Towards Automatic Animal Classification in Wildlife Environments for Native Species Monitoring in the Amazon","authors":"María-José Zurita, Daniel Riofrío, Noel Pérez, David Romo, D. Benítez, Ricardo Flores Moyano, Felipe Grijalva, Maria Baldeon-Calisto","doi":"10.1109/ColCACI59285.2023.10226093","DOIUrl":null,"url":null,"abstract":"Although critical for habitat and species conservation, camera trap image analysis is often manual, time-consuming, and expensive. Thus, automating this process would allow large-scale research on biodiversity hotspots of large conspicuous mammals and bird species. This paper explores the use of deep learning species-level object detection and classification models for this task, using two state-of-the-art architectures, YOLOv5 and Faster R-CNN, for two species: white-lipped peccary and collared peccary. The dataset contains 7,733 images obtained after data augmentation from the Tiputini Biodiversity Station. The models were trained in 70% of the dataset, validated in 20%, and tested in 10% of the available data. The YOLOv5 model proved to be more robust, having lower losses and a higher overall mAP (Mean Average Precision) value than Faster-RCNN. This is one of the first steps towards developing an automated camera trap analysis tool, allowing a large-scale analysis of population and habitat trends to benefit their conservation. The results suggest that hyperparameter fine-tuning would improve our models and allow us to extend this tool to other native species.","PeriodicalId":206196,"journal":{"name":"2023 IEEE Colombian Conference on Applications of Computational Intelligence (ColCACI)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE Colombian Conference on Applications of Computational Intelligence (ColCACI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ColCACI59285.2023.10226093","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Although critical for habitat and species conservation, camera trap image analysis is often manual, time-consuming, and expensive. Thus, automating this process would allow large-scale research on biodiversity hotspots of large conspicuous mammals and bird species. This paper explores the use of deep learning species-level object detection and classification models for this task, using two state-of-the-art architectures, YOLOv5 and Faster R-CNN, for two species: white-lipped peccary and collared peccary. The dataset contains 7,733 images obtained after data augmentation from the Tiputini Biodiversity Station. The models were trained in 70% of the dataset, validated in 20%, and tested in 10% of the available data. The YOLOv5 model proved to be more robust, having lower losses and a higher overall mAP (Mean Average Precision) value than Faster-RCNN. This is one of the first steps towards developing an automated camera trap analysis tool, allowing a large-scale analysis of population and habitat trends to benefit their conservation. The results suggest that hyperparameter fine-tuning would improve our models and allow us to extend this tool to other native species.
虽然对栖息地和物种保护至关重要,但相机陷阱图像分析通常是手动的,耗时且昂贵。因此,自动化这一过程将允许对大型显眼哺乳动物和鸟类物种的生物多样性热点进行大规模研究。本文探讨了在该任务中使用深度学习物种级目标检测和分类模型,使用两种最先进的架构,YOLOv5和Faster R-CNN,用于两种物种:白唇pecary和项圈pecary。该数据集包含7733幅来自蒂普蒂尼生物多样性站的图像。这些模型在70%的数据集上进行了训练,在20%的数据集上进行了验证,在10%的可用数据中进行了测试。事实证明,YOLOv5模型比Faster-RCNN更具鲁棒性,具有更低的损失和更高的总体mAP (Mean Average Precision)值。这是开发自动相机陷阱分析工具的第一步,允许对种群和栖息地趋势进行大规模分析,以有利于它们的保护。结果表明,超参数微调将改进我们的模型,并允许我们将该工具扩展到其他本地物种。