{"title":"基于生成式人工智能的管道架构,提高智能杂草控制系统的训练效率","authors":"Sourav Modak, Anthony Stein","doi":"10.1016/j.sysarc.2025.103464","DOIUrl":null,"url":null,"abstract":"<div><div>In automated crop protection tasks, deep learning has demonstrated significant potential. However, these advanced models rely heavily on high-quality, diverse datasets, which are often scarce and costly to obtain in agricultural settings. Traditional data augmentation techniques, while useful for increasing the volume of the dataset, often fail to capture the real-world variability needed for robust model training. In this paper, we present a novel method for generating synthetic images to enhance the training of deep learning-based object detection models for intelligent weed control, aiming to improve data efficiency. The architecture of our GenAI-based image generation pipeline integrates the Segment Anything Model (SAM) for zero-shot domain adaptation with a text-to-image Stable Diffusion Model, enabling the creation of synthetic images that can accurately reflect the idiosyncratic properties and appearances of a variety of real-world conditions. We further assess the application of these synthetic datasets on edge devices by evaluating state-of-the-art lightweight YOLO models, measuring data efficiency by comparing mAP50 and mAP50-95 scores among different proportions of real and synthetic training data. Incorporating these synthetic datasets into the training process has been found to result in notable improvements in terms of data efficiency. For instance, most YOLO models that are trained on a dataset consisting of 10% synthetic images and 90% real-world images typically demonstrate superior scores on mAP50 and mAP50-95 metrics compared to those trained solely on real-world images. The integration of this approach opens opportunities for achieving continual self-improvement of perception modules in intelligent technical systems.</div></div>","PeriodicalId":50027,"journal":{"name":"Journal of Systems Architecture","volume":"167 ","pages":"Article 103464"},"PeriodicalIF":3.7000,"publicationDate":"2025-06-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Generative AI-based pipeline architecture for increasing training efficiency in intelligent weed control systems\",\"authors\":\"Sourav Modak, Anthony Stein\",\"doi\":\"10.1016/j.sysarc.2025.103464\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>In automated crop protection tasks, deep learning has demonstrated significant potential. However, these advanced models rely heavily on high-quality, diverse datasets, which are often scarce and costly to obtain in agricultural settings. Traditional data augmentation techniques, while useful for increasing the volume of the dataset, often fail to capture the real-world variability needed for robust model training. In this paper, we present a novel method for generating synthetic images to enhance the training of deep learning-based object detection models for intelligent weed control, aiming to improve data efficiency. The architecture of our GenAI-based image generation pipeline integrates the Segment Anything Model (SAM) for zero-shot domain adaptation with a text-to-image Stable Diffusion Model, enabling the creation of synthetic images that can accurately reflect the idiosyncratic properties and appearances of a variety of real-world conditions. We further assess the application of these synthetic datasets on edge devices by evaluating state-of-the-art lightweight YOLO models, measuring data efficiency by comparing mAP50 and mAP50-95 scores among different proportions of real and synthetic training data. Incorporating these synthetic datasets into the training process has been found to result in notable improvements in terms of data efficiency. For instance, most YOLO models that are trained on a dataset consisting of 10% synthetic images and 90% real-world images typically demonstrate superior scores on mAP50 and mAP50-95 metrics compared to those trained solely on real-world images. The integration of this approach opens opportunities for achieving continual self-improvement of perception modules in intelligent technical systems.</div></div>\",\"PeriodicalId\":50027,\"journal\":{\"name\":\"Journal of Systems Architecture\",\"volume\":\"167 \",\"pages\":\"Article 103464\"},\"PeriodicalIF\":3.7000,\"publicationDate\":\"2025-06-02\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Systems Architecture\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1383762125001365\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Systems Architecture","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1383762125001365","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE","Score":null,"Total":0}
Generative AI-based pipeline architecture for increasing training efficiency in intelligent weed control systems
In automated crop protection tasks, deep learning has demonstrated significant potential. However, these advanced models rely heavily on high-quality, diverse datasets, which are often scarce and costly to obtain in agricultural settings. Traditional data augmentation techniques, while useful for increasing the volume of the dataset, often fail to capture the real-world variability needed for robust model training. In this paper, we present a novel method for generating synthetic images to enhance the training of deep learning-based object detection models for intelligent weed control, aiming to improve data efficiency. The architecture of our GenAI-based image generation pipeline integrates the Segment Anything Model (SAM) for zero-shot domain adaptation with a text-to-image Stable Diffusion Model, enabling the creation of synthetic images that can accurately reflect the idiosyncratic properties and appearances of a variety of real-world conditions. We further assess the application of these synthetic datasets on edge devices by evaluating state-of-the-art lightweight YOLO models, measuring data efficiency by comparing mAP50 and mAP50-95 scores among different proportions of real and synthetic training data. Incorporating these synthetic datasets into the training process has been found to result in notable improvements in terms of data efficiency. For instance, most YOLO models that are trained on a dataset consisting of 10% synthetic images and 90% real-world images typically demonstrate superior scores on mAP50 and mAP50-95 metrics compared to those trained solely on real-world images. The integration of this approach opens opportunities for achieving continual self-improvement of perception modules in intelligent technical systems.
期刊介绍:
The Journal of Systems Architecture: Embedded Software Design (JSA) is a journal covering all design and architectural aspects related to embedded systems and software. It ranges from the microarchitecture level via the system software level up to the application-specific architecture level. Aspects such as real-time systems, operating systems, FPGA programming, programming languages, communications (limited to analysis and the software stack), mobile systems, parallel and distributed architectures as well as additional subjects in the computer and system architecture area will fall within the scope of this journal. Technology will not be a main focus, but its use and relevance to particular designs will be. Case studies are welcome but must contribute more than just a design for a particular piece of software.
Design automation of such systems including methodologies, techniques and tools for their design as well as novel designs of software components fall within the scope of this journal. Novel applications that use embedded systems are also central in this journal. While hardware is not a part of this journal hardware/software co-design methods that consider interplay between software and hardware components with and emphasis on software are also relevant here.