Sikun Guo, Amir Hassan Shariatmadari, Guangzhi Xiong, Aidong Zhang
{"title":"拥抱推进科学发现的基础模型。","authors":"Sikun Guo, Amir Hassan Shariatmadari, Guangzhi Xiong, Aidong Zhang","doi":"10.1109/bigdata62323.2024.10825618","DOIUrl":null,"url":null,"abstract":"<p><p>Machine learning foundation models, particularly large language models (LLMs) such as GPT-4o, have revolutionized traditional applications in computer vision and natural language processing, marking a significant shift in recent years. Building on these advancements, recent efforts have explored the potential of foundation models in hypothesis generation, highlighting their possibility in aiding human researchers in scientific discovery. In this paper, we envision a future where academia increasingly integrates foundation models to accelerate and enhance the process of scientific discovery. Motivated by potential application scenarios of foundation models in scientific research, our vision is anchored in a central question: How can we accelerate scientific discovery with the aid of foundation models? To address this overarching question, we raise two key challenges that need to be addressed: (1) how to effectively harness the parametric knowledge embedded in foundation models to propel scientific discovery? and (2) how to develop rigorous yet scalable methods to evaluate the effectiveness of foundation models in supporting scientific research? To tackle these two challenges, we propose our approaches, termed knowledge-grounded Chain-of-Idea (KG-CoI) hypothesis generation and IdeaBench - Benchmarking LLM hypothesis generators in a customizable manner. Through addressing these challenges, we outline our vision in hope to inspire new ideas and innovations in harnessing foundation models for advancing scientific discovery, paving the way for a new era of research collaboration between humans and artificial intelligence.</p>","PeriodicalId":520404,"journal":{"name":"Proceedings : ... IEEE International Conference on Big Data. IEEE International Conference on Big Data","volume":"2024 ","pages":"1746-1755"},"PeriodicalIF":0.0000,"publicationDate":"2024-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11923747/pdf/","citationCount":"0","resultStr":"{\"title\":\"Embracing Foundation Models for Advancing Scientific Discovery.\",\"authors\":\"Sikun Guo, Amir Hassan Shariatmadari, Guangzhi Xiong, Aidong Zhang\",\"doi\":\"10.1109/bigdata62323.2024.10825618\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Machine learning foundation models, particularly large language models (LLMs) such as GPT-4o, have revolutionized traditional applications in computer vision and natural language processing, marking a significant shift in recent years. Building on these advancements, recent efforts have explored the potential of foundation models in hypothesis generation, highlighting their possibility in aiding human researchers in scientific discovery. In this paper, we envision a future where academia increasingly integrates foundation models to accelerate and enhance the process of scientific discovery. Motivated by potential application scenarios of foundation models in scientific research, our vision is anchored in a central question: How can we accelerate scientific discovery with the aid of foundation models? To address this overarching question, we raise two key challenges that need to be addressed: (1) how to effectively harness the parametric knowledge embedded in foundation models to propel scientific discovery? and (2) how to develop rigorous yet scalable methods to evaluate the effectiveness of foundation models in supporting scientific research? To tackle these two challenges, we propose our approaches, termed knowledge-grounded Chain-of-Idea (KG-CoI) hypothesis generation and IdeaBench - Benchmarking LLM hypothesis generators in a customizable manner. Through addressing these challenges, we outline our vision in hope to inspire new ideas and innovations in harnessing foundation models for advancing scientific discovery, paving the way for a new era of research collaboration between humans and artificial intelligence.</p>\",\"PeriodicalId\":520404,\"journal\":{\"name\":\"Proceedings : ... IEEE International Conference on Big Data. IEEE International Conference on Big Data\",\"volume\":\"2024 \",\"pages\":\"1746-1755\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-12-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11923747/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings : ... IEEE International Conference on Big Data. IEEE International Conference on Big Data\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/bigdata62323.2024.10825618\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings : ... IEEE International Conference on Big Data. IEEE International Conference on Big Data","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/bigdata62323.2024.10825618","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Embracing Foundation Models for Advancing Scientific Discovery.
Machine learning foundation models, particularly large language models (LLMs) such as GPT-4o, have revolutionized traditional applications in computer vision and natural language processing, marking a significant shift in recent years. Building on these advancements, recent efforts have explored the potential of foundation models in hypothesis generation, highlighting their possibility in aiding human researchers in scientific discovery. In this paper, we envision a future where academia increasingly integrates foundation models to accelerate and enhance the process of scientific discovery. Motivated by potential application scenarios of foundation models in scientific research, our vision is anchored in a central question: How can we accelerate scientific discovery with the aid of foundation models? To address this overarching question, we raise two key challenges that need to be addressed: (1) how to effectively harness the parametric knowledge embedded in foundation models to propel scientific discovery? and (2) how to develop rigorous yet scalable methods to evaluate the effectiveness of foundation models in supporting scientific research? To tackle these two challenges, we propose our approaches, termed knowledge-grounded Chain-of-Idea (KG-CoI) hypothesis generation and IdeaBench - Benchmarking LLM hypothesis generators in a customizable manner. Through addressing these challenges, we outline our vision in hope to inspire new ideas and innovations in harnessing foundation models for advancing scientific discovery, paving the way for a new era of research collaboration between humans and artificial intelligence.