{"title":"减少农业数据的注释工作量:使用DINOv2和K-means进行简单快速的无监督核心集选择。","authors":"Laura Gómez-Zamanillo, Nagore Portilla, Artzai Picón, Itziar Egusquiza, Ramón Navarra-Mestre, Andoni Elola, Arantza Bereciartua-Perez","doi":"10.3389/fpls.2025.1546756","DOIUrl":null,"url":null,"abstract":"<p><p>The need for large amounts of annotated data is a major obstacle to adopting deep learning in agricultural applications, where annotation is typically time-consuming and requires expert knowledge. To address this issue, methods have been developed to select data for manual annotation that represents the existing variability in the dataset, thereby avoiding redundant information. Coreset selection methods aim to choose a small subset of data samples that best represents the entire dataset. These methods can therefore be used to select a reduced set of samples for annotation, optimizing the training of a deep learning model for the best possible performance. In this work, we propose a simple yet effective coreset selection method that combines the recent foundation model DINOv2 as a powerful feature selector with the well-known K-Means clustering method. Samples are selected from each calculated cluster to form the final coreset. The proposed method is validated by comparing the performance metrics of a multiclass classification model trained on datasets reduced randomly and using the proposed method. This validation is conducted on two different datasets, and in both cases, the proposed method achieves better results, with improvements of up to 0.15 in the F1 score for significant reductions in the training datasets. Additionally, the importance of using DINOv2 as a feature extractor to achieve these good results is studied.</p>","PeriodicalId":12632,"journal":{"name":"Frontiers in Plant Science","volume":"16 ","pages":"1546756"},"PeriodicalIF":4.1000,"publicationDate":"2025-05-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12116677/pdf/","citationCount":"0","resultStr":"{\"title\":\"Reducing annotation effort in agricultural data: simple and fast unsupervised coreset selection with DINOv2 and K-means.\",\"authors\":\"Laura Gómez-Zamanillo, Nagore Portilla, Artzai Picón, Itziar Egusquiza, Ramón Navarra-Mestre, Andoni Elola, Arantza Bereciartua-Perez\",\"doi\":\"10.3389/fpls.2025.1546756\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>The need for large amounts of annotated data is a major obstacle to adopting deep learning in agricultural applications, where annotation is typically time-consuming and requires expert knowledge. To address this issue, methods have been developed to select data for manual annotation that represents the existing variability in the dataset, thereby avoiding redundant information. Coreset selection methods aim to choose a small subset of data samples that best represents the entire dataset. These methods can therefore be used to select a reduced set of samples for annotation, optimizing the training of a deep learning model for the best possible performance. In this work, we propose a simple yet effective coreset selection method that combines the recent foundation model DINOv2 as a powerful feature selector with the well-known K-Means clustering method. Samples are selected from each calculated cluster to form the final coreset. The proposed method is validated by comparing the performance metrics of a multiclass classification model trained on datasets reduced randomly and using the proposed method. This validation is conducted on two different datasets, and in both cases, the proposed method achieves better results, with improvements of up to 0.15 in the F1 score for significant reductions in the training datasets. Additionally, the importance of using DINOv2 as a feature extractor to achieve these good results is studied.</p>\",\"PeriodicalId\":12632,\"journal\":{\"name\":\"Frontiers in Plant Science\",\"volume\":\"16 \",\"pages\":\"1546756\"},\"PeriodicalIF\":4.1000,\"publicationDate\":\"2025-05-14\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12116677/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Frontiers in Plant Science\",\"FirstCategoryId\":\"99\",\"ListUrlMain\":\"https://doi.org/10.3389/fpls.2025.1546756\",\"RegionNum\":2,\"RegionCategory\":\"生物学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2025/1/1 0:00:00\",\"PubModel\":\"eCollection\",\"JCR\":\"Q1\",\"JCRName\":\"PLANT SCIENCES\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Frontiers in Plant Science","FirstCategoryId":"99","ListUrlMain":"https://doi.org/10.3389/fpls.2025.1546756","RegionNum":2,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/1/1 0:00:00","PubModel":"eCollection","JCR":"Q1","JCRName":"PLANT SCIENCES","Score":null,"Total":0}
Reducing annotation effort in agricultural data: simple and fast unsupervised coreset selection with DINOv2 and K-means.
The need for large amounts of annotated data is a major obstacle to adopting deep learning in agricultural applications, where annotation is typically time-consuming and requires expert knowledge. To address this issue, methods have been developed to select data for manual annotation that represents the existing variability in the dataset, thereby avoiding redundant information. Coreset selection methods aim to choose a small subset of data samples that best represents the entire dataset. These methods can therefore be used to select a reduced set of samples for annotation, optimizing the training of a deep learning model for the best possible performance. In this work, we propose a simple yet effective coreset selection method that combines the recent foundation model DINOv2 as a powerful feature selector with the well-known K-Means clustering method. Samples are selected from each calculated cluster to form the final coreset. The proposed method is validated by comparing the performance metrics of a multiclass classification model trained on datasets reduced randomly and using the proposed method. This validation is conducted on two different datasets, and in both cases, the proposed method achieves better results, with improvements of up to 0.15 in the F1 score for significant reductions in the training datasets. Additionally, the importance of using DINOv2 as a feature extractor to achieve these good results is studied.
期刊介绍:
In an ever changing world, plant science is of the utmost importance for securing the future well-being of humankind. Plants provide oxygen, food, feed, fibers, and building materials. In addition, they are a diverse source of industrial and pharmaceutical chemicals. Plants are centrally important to the health of ecosystems, and their understanding is critical for learning how to manage and maintain a sustainable biosphere. Plant science is extremely interdisciplinary, reaching from agricultural science to paleobotany, and molecular physiology to ecology. It uses the latest developments in computer science, optics, molecular biology and genomics to address challenges in model systems, agricultural crops, and ecosystems. Plant science research inquires into the form, function, development, diversity, reproduction, evolution and uses of both higher and lower plants and their interactions with other organisms throughout the biosphere. Frontiers in Plant Science welcomes outstanding contributions in any field of plant science from basic to applied research, from organismal to molecular studies, from single plant analysis to studies of populations and whole ecosystems, and from molecular to biophysical to computational approaches.
Frontiers in Plant Science publishes articles on the most outstanding discoveries across a wide research spectrum of Plant Science. The mission of Frontiers in Plant Science is to bring all relevant Plant Science areas together on a single platform.