Zicheng Liu, Jiahui Li, Siyuan Li, Zelin Zang, Cheng Tan, Yufei Huang, Yajing Bai, Stan Z. Li
{"title":"GenBench: A Benchmarking Suite for Systematic Evaluation of Genomic Foundation Models","authors":"Zicheng Liu, Jiahui Li, Siyuan Li, Zelin Zang, Cheng Tan, Yufei Huang, Yajing Bai, Stan Z. Li","doi":"arxiv-2406.01627","DOIUrl":null,"url":null,"abstract":"The Genomic Foundation Model (GFM) paradigm is expected to facilitate the\nextraction of generalizable representations from massive genomic data, thereby\nenabling their application across a spectrum of downstream applications.\nDespite advancements, a lack of evaluation framework makes it difficult to\nensure equitable assessment due to experimental settings, model intricacy,\nbenchmark datasets, and reproducibility challenges. In the absence of\nstandardization, comparative analyses risk becoming biased and unreliable. To\nsurmount this impasse, we introduce GenBench, a comprehensive benchmarking\nsuite specifically tailored for evaluating the efficacy of Genomic Foundation\nModels. GenBench offers a modular and expandable framework that encapsulates a\nvariety of state-of-the-art methodologies. Through systematic evaluations of\ndatasets spanning diverse biological domains with a particular emphasis on both\nshort-range and long-range genomic tasks, firstly including the three most\nimportant DNA tasks covering Coding Region, Non-Coding Region, Genome\nStructure, etc. Moreover, We provide a nuanced analysis of the interplay\nbetween model architecture and dataset characteristics on task-specific\nperformance. Our findings reveal an interesting observation: independent of the\nnumber of parameters, the discernible difference in preference between the\nattention-based and convolution-based models on short- and long-range tasks may\nprovide insights into the future design of GFM.","PeriodicalId":501070,"journal":{"name":"arXiv - QuanBio - Genomics","volume":"25 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - QuanBio - Genomics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2406.01627","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
The Genomic Foundation Model (GFM) paradigm is expected to facilitate the
extraction of generalizable representations from massive genomic data, thereby
enabling their application across a spectrum of downstream applications.
Despite advancements, a lack of evaluation framework makes it difficult to
ensure equitable assessment due to experimental settings, model intricacy,
benchmark datasets, and reproducibility challenges. In the absence of
standardization, comparative analyses risk becoming biased and unreliable. To
surmount this impasse, we introduce GenBench, a comprehensive benchmarking
suite specifically tailored for evaluating the efficacy of Genomic Foundation
Models. GenBench offers a modular and expandable framework that encapsulates a
variety of state-of-the-art methodologies. Through systematic evaluations of
datasets spanning diverse biological domains with a particular emphasis on both
short-range and long-range genomic tasks, firstly including the three most
important DNA tasks covering Coding Region, Non-Coding Region, Genome
Structure, etc. Moreover, We provide a nuanced analysis of the interplay
between model architecture and dataset characteristics on task-specific
performance. Our findings reveal an interesting observation: independent of the
number of parameters, the discernible difference in preference between the
attention-based and convolution-based models on short- and long-range tasks may
provide insights into the future design of GFM.