Mark S. Pleasants, Michael N. Fienen, Hedeff I. Essaid, Joel D. Blomquist, Jing Yang, Ming Ye
{"title":"迈向一个新的框架,以评估过程为基础的模型配置和量化数据价值之前的校准","authors":"Mark S. Pleasants, Michael N. Fienen, Hedeff I. Essaid, Joel D. Blomquist, Jing Yang, Ming Ye","doi":"10.1029/2025wr040323","DOIUrl":null,"url":null,"abstract":"Model criticism, discrimination, and selection methods often rely on calibrated model outputs. Because calibration can be computationally expensive, model criticism can first be undertaken by assessing model outputs obtained from limited prior parameter ensembles. However, such prior‐based methods are often heuristic and do not formalize the notion of balancing model consistency with data and model complexity (i.e., model adequacy). We present a new framework to discriminate among candidate models prior to calibration that formalizes prior‐to‐calibration model adequacy into a metric to implicitly balance prior model output data coverage with model complexity represented by prior output (co)variance. The prior model adequacy metric “Mahalanobis distance deviation” quantifies the deviation of (a) the set of squared Mahalanobis distances of data from a prior model output distribution from (b) the set of squared Mahalanobis distances of data from their own distribution. A new data worth metric “discernment value” is also presented which quantifies the value of data for screening less‐adequate models prior to calibration. Discernment value is calculated from the change in variance of a weighted average of prior model outputs from all candidate models due to less‐adequate model outputs receiving lower weight. The framework is demonstrated using a one‐dimensional groundwater flow model with eight possible configurations. A synthetic data network is used to test the framework. Results show the framework identifies the candidate models most similar to the true model used to create the synthetic data. Discernment values show variation in the value of different data types and locations for screening less‐adequate models.","PeriodicalId":23799,"journal":{"name":"Water Resources Research","volume":"24 1","pages":""},"PeriodicalIF":5.0000,"publicationDate":"2025-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Toward a New Framework to Evaluate Process‐Based Model Configurations and Quantify Data Worth Prior to Calibration\",\"authors\":\"Mark S. Pleasants, Michael N. Fienen, Hedeff I. Essaid, Joel D. Blomquist, Jing Yang, Ming Ye\",\"doi\":\"10.1029/2025wr040323\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Model criticism, discrimination, and selection methods often rely on calibrated model outputs. Because calibration can be computationally expensive, model criticism can first be undertaken by assessing model outputs obtained from limited prior parameter ensembles. However, such prior‐based methods are often heuristic and do not formalize the notion of balancing model consistency with data and model complexity (i.e., model adequacy). We present a new framework to discriminate among candidate models prior to calibration that formalizes prior‐to‐calibration model adequacy into a metric to implicitly balance prior model output data coverage with model complexity represented by prior output (co)variance. The prior model adequacy metric “Mahalanobis distance deviation” quantifies the deviation of (a) the set of squared Mahalanobis distances of data from a prior model output distribution from (b) the set of squared Mahalanobis distances of data from their own distribution. A new data worth metric “discernment value” is also presented which quantifies the value of data for screening less‐adequate models prior to calibration. Discernment value is calculated from the change in variance of a weighted average of prior model outputs from all candidate models due to less‐adequate model outputs receiving lower weight. The framework is demonstrated using a one‐dimensional groundwater flow model with eight possible configurations. A synthetic data network is used to test the framework. Results show the framework identifies the candidate models most similar to the true model used to create the synthetic data. Discernment values show variation in the value of different data types and locations for screening less‐adequate models.\",\"PeriodicalId\":23799,\"journal\":{\"name\":\"Water Resources Research\",\"volume\":\"24 1\",\"pages\":\"\"},\"PeriodicalIF\":5.0000,\"publicationDate\":\"2025-09-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Water Resources Research\",\"FirstCategoryId\":\"89\",\"ListUrlMain\":\"https://doi.org/10.1029/2025wr040323\",\"RegionNum\":1,\"RegionCategory\":\"地球科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"ENVIRONMENTAL SCIENCES\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Water Resources Research","FirstCategoryId":"89","ListUrlMain":"https://doi.org/10.1029/2025wr040323","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENVIRONMENTAL SCIENCES","Score":null,"Total":0}
Toward a New Framework to Evaluate Process‐Based Model Configurations and Quantify Data Worth Prior to Calibration
Model criticism, discrimination, and selection methods often rely on calibrated model outputs. Because calibration can be computationally expensive, model criticism can first be undertaken by assessing model outputs obtained from limited prior parameter ensembles. However, such prior‐based methods are often heuristic and do not formalize the notion of balancing model consistency with data and model complexity (i.e., model adequacy). We present a new framework to discriminate among candidate models prior to calibration that formalizes prior‐to‐calibration model adequacy into a metric to implicitly balance prior model output data coverage with model complexity represented by prior output (co)variance. The prior model adequacy metric “Mahalanobis distance deviation” quantifies the deviation of (a) the set of squared Mahalanobis distances of data from a prior model output distribution from (b) the set of squared Mahalanobis distances of data from their own distribution. A new data worth metric “discernment value” is also presented which quantifies the value of data for screening less‐adequate models prior to calibration. Discernment value is calculated from the change in variance of a weighted average of prior model outputs from all candidate models due to less‐adequate model outputs receiving lower weight. The framework is demonstrated using a one‐dimensional groundwater flow model with eight possible configurations. A synthetic data network is used to test the framework. Results show the framework identifies the candidate models most similar to the true model used to create the synthetic data. Discernment values show variation in the value of different data types and locations for screening less‐adequate models.
期刊介绍:
Water Resources Research (WRR) is an interdisciplinary journal that focuses on hydrology and water resources. It publishes original research in the natural and social sciences of water. It emphasizes the role of water in the Earth system, including physical, chemical, biological, and ecological processes in water resources research and management, including social, policy, and public health implications. It encompasses observational, experimental, theoretical, analytical, numerical, and data-driven approaches that advance the science of water and its management. Submissions are evaluated for their novelty, accuracy, significance, and broader implications of the findings.