An Advanced Chicken Face Detection Network Based on GAN and MAE.

Xiaoxiao Ma, Xinai Lu, Yihong Huang, Xinyi Yang, Ziyin Xu, Guozhao Mo, Yufei Ren, Lin Li
{"title":"An Advanced Chicken Face Detection Network Based on GAN and MAE.","authors":"Xiaoxiao Ma,&nbsp;Xinai Lu,&nbsp;Yihong Huang,&nbsp;Xinyi Yang,&nbsp;Ziyin Xu,&nbsp;Guozhao Mo,&nbsp;Yufei Ren,&nbsp;Lin Li","doi":"10.3390/ani12213055","DOIUrl":null,"url":null,"abstract":"<p><p>Achieving high-accuracy chicken face detection is a significant breakthrough for smart poultry agriculture in large-scale farming and precision management. However, the current dataset of chicken faces based on accurate data is scarce, detection models possess low accuracy and slow speed, and the related detection algorithm is ineffective for small object detection. To tackle these problems, an object detection network based on GAN-MAE (generative adversarial network-masked autoencoders) data augmentation is proposed in this paper for detecting chickens of different ages. First, the images were generated using GAN and MAE to augment the dataset. Afterward, CSPDarknet53 was used as the backbone network to enhance the receptive field in the object detection network to detect different sizes of objects in the same image. The 128×128 feature map output was added to three feature map outputs of this paper, thus changing the feature map output of eightfold downsampling to fourfold downsampling, which provided smaller object features for subsequent feature fusion. Secondly, the feature fusion module was improved based on the idea of dense connection. Then the module achieved feature reuse so that the YOLO head classifier could combine features from different levels of feature layers to capture greater classification and detection results. Ultimately, the comparison experiments' outcomes showed that the mAP (mean average Precision) of the suggested method was up to 0.84, which was 29.2% higher than other networks', and the detection speed was the same, up to 37 frames per second. Better detection accuracy can be obtained while meeting the actual scenario detection requirements. Additionally, an end-to-end web system was designed to apply the algorithm to practical applications.</p>","PeriodicalId":519482,"journal":{"name":"Animals : an Open Access Journal from MDPI","volume":" ","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2022-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9655765/pdf/","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Animals : an Open Access Journal from MDPI","FirstCategoryId":"97","ListUrlMain":"https://doi.org/10.3390/ani12213055","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

Abstract

Achieving high-accuracy chicken face detection is a significant breakthrough for smart poultry agriculture in large-scale farming and precision management. However, the current dataset of chicken faces based on accurate data is scarce, detection models possess low accuracy and slow speed, and the related detection algorithm is ineffective for small object detection. To tackle these problems, an object detection network based on GAN-MAE (generative adversarial network-masked autoencoders) data augmentation is proposed in this paper for detecting chickens of different ages. First, the images were generated using GAN and MAE to augment the dataset. Afterward, CSPDarknet53 was used as the backbone network to enhance the receptive field in the object detection network to detect different sizes of objects in the same image. The 128×128 feature map output was added to three feature map outputs of this paper, thus changing the feature map output of eightfold downsampling to fourfold downsampling, which provided smaller object features for subsequent feature fusion. Secondly, the feature fusion module was improved based on the idea of dense connection. Then the module achieved feature reuse so that the YOLO head classifier could combine features from different levels of feature layers to capture greater classification and detection results. Ultimately, the comparison experiments' outcomes showed that the mAP (mean average Precision) of the suggested method was up to 0.84, which was 29.2% higher than other networks', and the detection speed was the same, up to 37 frames per second. Better detection accuracy can be obtained while meeting the actual scenario detection requirements. Additionally, an end-to-end web system was designed to apply the algorithm to practical applications.

基于GAN和MAE的鸡脸检测网络。
实现高精度的鸡脸检测是智能家禽农业在规模化养殖和精细化管理方面的重大突破。然而,目前基于准确数据的鸡脸数据集很少,检测模型精度低、速度慢,相关检测算法对小目标检测效果不佳。为了解决这些问题,本文提出了一种基于GAN-MAE(生成对抗网络掩码自编码器)数据增强的目标检测网络,用于检测不同年龄的鸡。首先,使用GAN和MAE生成图像以增强数据集。随后,使用CSPDarknet53作为骨干网络,增强目标检测网络中的感受野,检测同幅图像中不同大小的目标。将128×128特征图输出加入到本文的三个特征图输出中,将八次降采样的特征图输出改为四次降采样,为后续的特征融合提供了更小的目标特征。其次,基于密集连接思想对特征融合模块进行改进;然后模块实现特征重用,使得YOLO头部分类器可以将不同层次特征层的特征组合起来,获得更大的分类检测结果。最终,对比实验结果表明,该方法的mAP (mean average Precision)高达0.84,比其他网络提高了29.2%,并且检测速度相同,高达37帧/秒。在满足实际场景检测需求的同时,可以获得更好的检测精度。此外,还设计了一个端到端web系统,将该算法应用于实际应用。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信