Khalif Amir Zakry, Mohamad Syahiran Soria, Irwandi Hipni Mohamad Hipiny, Hamimah Ujir, Ruhana Hassan
{"title":"Chelonia mydas detection and image extraction from noisy field recordings","authors":"Khalif Amir Zakry, Mohamad Syahiran Soria, Irwandi Hipni Mohamad Hipiny, Hamimah Ujir, Ruhana Hassan","doi":"10.11591/ijai.v13.i2.pp2354-2363","DOIUrl":null,"url":null,"abstract":"Wildlife videography is an essential data collection method for conducting research on animals. The video recording process of an animal like the Chelonia Mydas turtle in its natural habitat requires the setting up of special camera traps or by performing complex camera movement to capture the animal in frame whilst the cameraman maneuvers over uneven terrain while filming. The result is hours of footage that only have the presence of the intended subject in it for seconds whilst the rest is background footage; or noisy and blurry footage that has only several usable frames among thousands of noisy and unusable ones. This presents a problem that deep learning models can help to assist, especially in detecting a wildlife subject and extracting usable data from hours of noise and background footage. This paper proposes the use of machine learning models to detect and extract wildlife images of Chelonia Mydas turtles to help prune through hundreds and thousands of frames from several video footages. Our paper shows that utilizing a custom model with various confidence scores can label and crop out images in noisy field video recordings of Chelonia Mydas turtles with up to 99.89% of output images correctly cropped and labeled.","PeriodicalId":507934,"journal":{"name":"IAES International Journal of Artificial Intelligence (IJ-AI)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IAES International Journal of Artificial Intelligence (IJ-AI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.11591/ijai.v13.i2.pp2354-2363","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Wildlife videography is an essential data collection method for conducting research on animals. The video recording process of an animal like the Chelonia Mydas turtle in its natural habitat requires the setting up of special camera traps or by performing complex camera movement to capture the animal in frame whilst the cameraman maneuvers over uneven terrain while filming. The result is hours of footage that only have the presence of the intended subject in it for seconds whilst the rest is background footage; or noisy and blurry footage that has only several usable frames among thousands of noisy and unusable ones. This presents a problem that deep learning models can help to assist, especially in detecting a wildlife subject and extracting usable data from hours of noise and background footage. This paper proposes the use of machine learning models to detect and extract wildlife images of Chelonia Mydas turtles to help prune through hundreds and thousands of frames from several video footages. Our paper shows that utilizing a custom model with various confidence scores can label and crop out images in noisy field video recordings of Chelonia Mydas turtles with up to 99.89% of output images correctly cropped and labeled.