Mehdi Seifi, Damian Dalle Nogare, Juan Manuel Battagliotti, Vera Galinova, Ananya Kedige Rao, Pierre-Henri Jouneau, Anwai Archit, Constantin Pape, Johan Decelle, Florian Jug, Joran Deschamps
{"title":"FeatureForest:基础模型的力量,随机森林的可用性。","authors":"Mehdi Seifi, Damian Dalle Nogare, Juan Manuel Battagliotti, Vera Galinova, Ananya Kedige Rao, Pierre-Henri Jouneau, Anwai Archit, Constantin Pape, Johan Decelle, Florian Jug, Joran Deschamps","doi":"10.1038/s44303-025-00089-9","DOIUrl":null,"url":null,"abstract":"<p><p>Analysis of biological images relies heavily on segmenting the biological objects of interest in the image before performing quantitative analysis. Deep learning (DL) is ubiquitous in such segmentation tasks, but can be cumbersome to apply, as it often requires a large amount of manual labeling to produce ground-truth data, and expert knowledge to train the models. More recently, large foundation models, such as SAM, have shown promising results on scientific images. They, however, require manual prompting for each object or tedious post-processing to selectively segment these objects. Here, we present FeatureForest, a method that leverages the feature embeddings of large foundation models to train a random forest classifier, thereby providing users with a rapid way of semantically segmenting complex images using only a few labeling strokes. We demonstrate the improvement in performance over a variety of datasets and provide an open-source implementation in napari that can be extended to new models.</p>","PeriodicalId":501709,"journal":{"name":"npj Imaging","volume":"3 1","pages":"32"},"PeriodicalIF":0.0000,"publicationDate":"2025-07-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12238231/pdf/","citationCount":"0","resultStr":"{\"title\":\"FeatureForest: the power of foundation models, the usability of random forests.\",\"authors\":\"Mehdi Seifi, Damian Dalle Nogare, Juan Manuel Battagliotti, Vera Galinova, Ananya Kedige Rao, Pierre-Henri Jouneau, Anwai Archit, Constantin Pape, Johan Decelle, Florian Jug, Joran Deschamps\",\"doi\":\"10.1038/s44303-025-00089-9\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Analysis of biological images relies heavily on segmenting the biological objects of interest in the image before performing quantitative analysis. Deep learning (DL) is ubiquitous in such segmentation tasks, but can be cumbersome to apply, as it often requires a large amount of manual labeling to produce ground-truth data, and expert knowledge to train the models. More recently, large foundation models, such as SAM, have shown promising results on scientific images. They, however, require manual prompting for each object or tedious post-processing to selectively segment these objects. Here, we present FeatureForest, a method that leverages the feature embeddings of large foundation models to train a random forest classifier, thereby providing users with a rapid way of semantically segmenting complex images using only a few labeling strokes. We demonstrate the improvement in performance over a variety of datasets and provide an open-source implementation in napari that can be extended to new models.</p>\",\"PeriodicalId\":501709,\"journal\":{\"name\":\"npj Imaging\",\"volume\":\"3 1\",\"pages\":\"32\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2025-07-08\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12238231/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"npj Imaging\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1038/s44303-025-00089-9\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"npj Imaging","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1038/s44303-025-00089-9","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
FeatureForest: the power of foundation models, the usability of random forests.
Analysis of biological images relies heavily on segmenting the biological objects of interest in the image before performing quantitative analysis. Deep learning (DL) is ubiquitous in such segmentation tasks, but can be cumbersome to apply, as it often requires a large amount of manual labeling to produce ground-truth data, and expert knowledge to train the models. More recently, large foundation models, such as SAM, have shown promising results on scientific images. They, however, require manual prompting for each object or tedious post-processing to selectively segment these objects. Here, we present FeatureForest, a method that leverages the feature embeddings of large foundation models to train a random forest classifier, thereby providing users with a rapid way of semantically segmenting complex images using only a few labeling strokes. We demonstrate the improvement in performance over a variety of datasets and provide an open-source implementation in napari that can be extended to new models.