GAN-based motion blur elimination as a preprocessing step for enhanced AI-driven computer-aided camera monitoring in poultry and free-range farming in low-resource settings

IF 6.3 Q1 AGRICULTURAL ENGINEERING
Shwetha V , Maddodi B S , Sheikh Adil , Vijaya Laxmi , Sakshi Shrivastava
{"title":"GAN-based motion blur elimination as a preprocessing step for enhanced AI-driven computer-aided camera monitoring in poultry and free-range farming in low-resource settings","authors":"Shwetha V ,&nbsp;Maddodi B S ,&nbsp;Sheikh Adil ,&nbsp;Vijaya Laxmi ,&nbsp;Sakshi Shrivastava","doi":"10.1016/j.atech.2025.100915","DOIUrl":null,"url":null,"abstract":"<div><div>Accurately determining poultry gender ratios is essential for assessing the economic value of free-range farming, particularly in resource-limited settings. Traditional and manual methods for gender identification are often labor-intensive, time-consuming, and prone to errors, especially in the early stages of poultry development. To address these challenges, this study introduces an automated approach that uses advanced machine learning techniques. Specifically, we propose a classification framework that integrates Convolutional Neural Networks (CNNs) with Generative Adversarial Networks (GANs) to enhance the accuracy of poultry gender identification. Our framework incorporates a novel GAN-based motion blur elimination method, which is broadly applicable to detect and classify moving subjects, including poultry. The proposed approach demonstrates a 98% accuracy in distinguishing between male and female birds at early growth stages by analyzing key features such as crown pixel measurements, feather gap analysis, and leg measurements. Furthermore, we conduct a comprehensive comparison of four segmentation models—UNet, ResUNet, ResUNet+, and a novel GAN-enhanced UNet—under varying motion blur conditions (80%, 50%, 30%, and 10%). Our results highlight the superiority of ResUNet+ over conventional models, achieving a peak Dice Coefficient of 91.2%, an Intersection over Union (IoU) of 86.7%, and the highest segmentation accuracy at reduced blur levels. These findings underscore the efficacy of deep learning-based approaches in advancing poultry gender classification while improving image quality in dynamic environments.</div></div>","PeriodicalId":74813,"journal":{"name":"Smart agricultural technology","volume":"11 ","pages":"Article 100915"},"PeriodicalIF":6.3000,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Smart agricultural technology","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2772375525001480","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AGRICULTURAL ENGINEERING","Score":null,"Total":0}
引用次数: 0

Abstract

Accurately determining poultry gender ratios is essential for assessing the economic value of free-range farming, particularly in resource-limited settings. Traditional and manual methods for gender identification are often labor-intensive, time-consuming, and prone to errors, especially in the early stages of poultry development. To address these challenges, this study introduces an automated approach that uses advanced machine learning techniques. Specifically, we propose a classification framework that integrates Convolutional Neural Networks (CNNs) with Generative Adversarial Networks (GANs) to enhance the accuracy of poultry gender identification. Our framework incorporates a novel GAN-based motion blur elimination method, which is broadly applicable to detect and classify moving subjects, including poultry. The proposed approach demonstrates a 98% accuracy in distinguishing between male and female birds at early growth stages by analyzing key features such as crown pixel measurements, feather gap analysis, and leg measurements. Furthermore, we conduct a comprehensive comparison of four segmentation models—UNet, ResUNet, ResUNet+, and a novel GAN-enhanced UNet—under varying motion blur conditions (80%, 50%, 30%, and 10%). Our results highlight the superiority of ResUNet+ over conventional models, achieving a peak Dice Coefficient of 91.2%, an Intersection over Union (IoU) of 86.7%, and the highest segmentation accuracy at reduced blur levels. These findings underscore the efficacy of deep learning-based approaches in advancing poultry gender classification while improving image quality in dynamic environments.
基于gan的运动模糊消除作为预处理步骤,用于增强人工智能驱动的计算机辅助摄像机监控,用于低资源环境下的家禽和散养养殖
准确确定家禽性别比例对于评估自由放养的经济价值至关重要,特别是在资源有限的情况下。传统的和手工的性别鉴定方法往往是劳动密集型的,耗时的,而且容易出错,特别是在家禽发育的早期阶段。为了应对这些挑战,本研究引入了一种使用先进机器学习技术的自动化方法。具体而言,我们提出了一个集成卷积神经网络(cnn)和生成对抗网络(gan)的分类框架,以提高家禽性别识别的准确性。我们的框架结合了一种新的基于gan的运动模糊消除方法,该方法广泛适用于检测和分类运动对象,包括家禽。该方法通过分析诸如冠像素测量、羽毛间隙分析和腿部测量等关键特征,证明了在早期生长阶段区分雄性和雌性鸟类的准确率为98%。此外,我们在不同的运动模糊条件下(80%,50%,30%和10%)对四种分割模型- unet, ResUNet, ResUNet+和一种新的gan增强unet进行了全面的比较。我们的研究结果突出了ResUNet+相对于传统模型的优势,达到了91.2%的峰值Dice系数,86.7%的交集/联合(Intersection over Union, IoU),以及在降低模糊水平下的最高分割精度。这些发现强调了基于深度学习的方法在推进家禽性别分类的同时提高动态环境下的图像质量的有效性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
4.20
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信