{"title":"基于机器学习的贝叶斯超参数优化eQTL分析","authors":"Andrew Quitadamo, James Johnson, Xinghua Shi","doi":"10.1145/3107411.3107434","DOIUrl":null,"url":null,"abstract":"Machine learning methods are being applied to a wide range of problems in biology and bioinformatics. These methods often rely on configuring high level parameters, or hyperparameters, such as regularization hyperparameters in sparse learning models like graph-guided multitask Lasso methods. Different choices for these hyperparameters will lead to different results, which makes finding good hyperparameter combinations an important task when using these hyperparameter dependent methods. There are several different ways to tune hyperparameters including manual tuning, grid search, random search, and Bayesian optimization. In this paper, we apply three hyperparameter tuning strategies to eQTL analysis including grid and random search in addition to Bayesian optimization. Experiments show that the Bayesian optimization strategy outperforms the other strategies in modeling eQTL associations. Applying this strategy to assess eQTL associations using the 1000 Genomes structural variation genotypes and RNAseq data in gEUVADIS, we identify a set of new SVs associated with gene expression changes in a human population.","PeriodicalId":246388,"journal":{"name":"Proceedings of the 8th ACM International Conference on Bioinformatics, Computational Biology,and Health Informatics","volume":"88 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2017-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"Bayesian Hyperparameter Optimization for Machine Learning Based eQTL Analysis\",\"authors\":\"Andrew Quitadamo, James Johnson, Xinghua Shi\",\"doi\":\"10.1145/3107411.3107434\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Machine learning methods are being applied to a wide range of problems in biology and bioinformatics. These methods often rely on configuring high level parameters, or hyperparameters, such as regularization hyperparameters in sparse learning models like graph-guided multitask Lasso methods. Different choices for these hyperparameters will lead to different results, which makes finding good hyperparameter combinations an important task when using these hyperparameter dependent methods. There are several different ways to tune hyperparameters including manual tuning, grid search, random search, and Bayesian optimization. In this paper, we apply three hyperparameter tuning strategies to eQTL analysis including grid and random search in addition to Bayesian optimization. Experiments show that the Bayesian optimization strategy outperforms the other strategies in modeling eQTL associations. Applying this strategy to assess eQTL associations using the 1000 Genomes structural variation genotypes and RNAseq data in gEUVADIS, we identify a set of new SVs associated with gene expression changes in a human population.\",\"PeriodicalId\":246388,\"journal\":{\"name\":\"Proceedings of the 8th ACM International Conference on Bioinformatics, Computational Biology,and Health Informatics\",\"volume\":\"88 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2017-08-20\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 8th ACM International Conference on Bioinformatics, Computational Biology,and Health Informatics\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3107411.3107434\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 8th ACM International Conference on Bioinformatics, Computational Biology,and Health Informatics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3107411.3107434","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Bayesian Hyperparameter Optimization for Machine Learning Based eQTL Analysis
Machine learning methods are being applied to a wide range of problems in biology and bioinformatics. These methods often rely on configuring high level parameters, or hyperparameters, such as regularization hyperparameters in sparse learning models like graph-guided multitask Lasso methods. Different choices for these hyperparameters will lead to different results, which makes finding good hyperparameter combinations an important task when using these hyperparameter dependent methods. There are several different ways to tune hyperparameters including manual tuning, grid search, random search, and Bayesian optimization. In this paper, we apply three hyperparameter tuning strategies to eQTL analysis including grid and random search in addition to Bayesian optimization. Experiments show that the Bayesian optimization strategy outperforms the other strategies in modeling eQTL associations. Applying this strategy to assess eQTL associations using the 1000 Genomes structural variation genotypes and RNAseq data in gEUVADIS, we identify a set of new SVs associated with gene expression changes in a human population.