大尺度历史匹配问题的低自由度高斯混合模型拟合

G. Gao, Hao Jiang, Chaohui Chen, J. Vink, Y. E. Khamra, J. Ita, F. Saaf
{"title":"大尺度历史匹配问题的低自由度高斯混合模型拟合","authors":"G. Gao, Hao Jiang, Chaohui Chen, J. Vink, Y. E. Khamra, J. Ita, F. Saaf","doi":"10.2118/193916-ms","DOIUrl":null,"url":null,"abstract":"\n Gaussian-mixture-model (GMM) fitting has been proved a robust method to generate high quality, independent conditional samples of the posterior probability density function (PDF) by conditioning reservoir models to production data. However, the number of degrees-of-freedom (DOF) for all unknown GMM parameters may become huge for large-scale history-matching problems. A new formulation of GMM fitting with reduced number of DOF is proposed in this paper, to save memory-usage and reduce computational cost. Its performance is compared with other methods of GMM.\n The GMM fitting method can significantly improve the accuracy of the GMM approximation by adding more Gaussian components. In the full-rank GMM fitting formulation, both memory-usage and computational cost are proportional to the number of Gaussian components. In the reduced DOF GMM fitting formulation, the covariance matrix of the newly added Gaussian component is efficiently parameterized, using products of a low number of vectors and their transposes, whereas the other Gaussian components are simply modified by multipliers. Thus, memory usage and computational cost increase only modestly as the number of Gaussian components increases.\n Unknown GMM parameters, including the parameterized covariance matrix and mixture weighting factor for each Gaussian component, are first determined by minimizing the error that measures the distance between the GMM approximation and the actual posterior PDF. Then, performance of the new method is benchmarked against other methods using test problems with different numbers of uncertain parameters. The new method is found to perform more efficiently than the full-rank GMM fitting formulation, e.g., it further reduces the memory usage and computational cost by a factor of 5 to 10, while it achieves comparable accuracy. Although it is less efficient than the L-GMM approximation based on local linearization, it achieves much higher accuracy, e.g., it manages to further reduce the error by a factor of 20 to 600.\n Finally, the new method together with the parallelized acceptance-rejection (AR) algorithm is applied to a history matching problem. It is found to reduce the computational cost (i.e., the number of simulations required to generate an accepted conditional realization on average) by a factor of 200 when compared with the Markov chain Monte Carlo (MCMC) method, while the quality of accepted GMM samples is comparable to the MCMC samples. Uncertainty of reservoir model parameters and production forecasts can be properly quantified with accepted GMM samples by conditioning to production data.","PeriodicalId":246878,"journal":{"name":"Day 2 Thu, April 11, 2019","volume":"71 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Reduced Degrees of Freedom Gaussian Mixture Model Fitting for Large Scale History Matching Problems\",\"authors\":\"G. Gao, Hao Jiang, Chaohui Chen, J. Vink, Y. E. Khamra, J. Ita, F. Saaf\",\"doi\":\"10.2118/193916-ms\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"\\n Gaussian-mixture-model (GMM) fitting has been proved a robust method to generate high quality, independent conditional samples of the posterior probability density function (PDF) by conditioning reservoir models to production data. However, the number of degrees-of-freedom (DOF) for all unknown GMM parameters may become huge for large-scale history-matching problems. A new formulation of GMM fitting with reduced number of DOF is proposed in this paper, to save memory-usage and reduce computational cost. Its performance is compared with other methods of GMM.\\n The GMM fitting method can significantly improve the accuracy of the GMM approximation by adding more Gaussian components. In the full-rank GMM fitting formulation, both memory-usage and computational cost are proportional to the number of Gaussian components. In the reduced DOF GMM fitting formulation, the covariance matrix of the newly added Gaussian component is efficiently parameterized, using products of a low number of vectors and their transposes, whereas the other Gaussian components are simply modified by multipliers. Thus, memory usage and computational cost increase only modestly as the number of Gaussian components increases.\\n Unknown GMM parameters, including the parameterized covariance matrix and mixture weighting factor for each Gaussian component, are first determined by minimizing the error that measures the distance between the GMM approximation and the actual posterior PDF. Then, performance of the new method is benchmarked against other methods using test problems with different numbers of uncertain parameters. The new method is found to perform more efficiently than the full-rank GMM fitting formulation, e.g., it further reduces the memory usage and computational cost by a factor of 5 to 10, while it achieves comparable accuracy. Although it is less efficient than the L-GMM approximation based on local linearization, it achieves much higher accuracy, e.g., it manages to further reduce the error by a factor of 20 to 600.\\n Finally, the new method together with the parallelized acceptance-rejection (AR) algorithm is applied to a history matching problem. It is found to reduce the computational cost (i.e., the number of simulations required to generate an accepted conditional realization on average) by a factor of 200 when compared with the Markov chain Monte Carlo (MCMC) method, while the quality of accepted GMM samples is comparable to the MCMC samples. Uncertainty of reservoir model parameters and production forecasts can be properly quantified with accepted GMM samples by conditioning to production data.\",\"PeriodicalId\":246878,\"journal\":{\"name\":\"Day 2 Thu, April 11, 2019\",\"volume\":\"71 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-03-29\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Day 2 Thu, April 11, 2019\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.2118/193916-ms\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Day 2 Thu, April 11, 2019","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.2118/193916-ms","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

高斯混合模型(GMM)拟合已被证明是一种鲁棒的方法,通过将储层模型与生产数据相适应,可以生成高质量的独立后验概率密度函数(PDF)条件样本。然而,对于大规模的历史匹配问题,所有未知GMM参数的自由度(DOF)数量可能会变得非常大。为了节省内存和降低计算成本,本文提出了一种减少自由度的GMM拟合新公式。并与其它GMM方法进行了性能比较。GMM拟合方法通过加入更多的高斯分量,可以显著提高GMM逼近的精度。在全秩GMM拟合公式中,内存使用和计算成本都与高斯分量的数量成正比。在降自由度GMM拟合公式中,新加入的高斯分量的协方差矩阵通过使用少量向量及其转置的乘积有效地参数化,而其他高斯分量则通过乘法器简单地修改。因此,随着高斯分量数量的增加,内存使用量和计算成本只会适度增加。未知的GMM参数,包括参数化协方差矩阵和每个高斯分量的混合加权因子,首先通过最小化测量GMM近似值与实际后验PDF之间距离的误差来确定。然后,利用具有不同数量不确定参数的测试问题,将新方法的性能与其他方法进行基准测试。新方法比全秩GMM拟合公式更有效,例如,它进一步减少了5到10倍的内存使用和计算成本,同时达到了相当的精度。虽然它的效率低于基于局部线性化的L-GMM近似,但它达到了更高的精度,例如,它成功地将误差进一步减少了20到600倍。最后,将该方法与并行接受-拒绝(AR)算法一起应用于历史匹配问题。研究发现,与马尔可夫链蒙特卡罗(MCMC)方法相比,计算成本(即平均生成可接受的条件实现所需的模拟次数)降低了200倍,而可接受的GMM样本的质量与MCMC样本相当。通过对生产数据的调节,可以用可接受的GMM样本适当地量化储层模型参数和产量预测的不确定性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Reduced Degrees of Freedom Gaussian Mixture Model Fitting for Large Scale History Matching Problems
Gaussian-mixture-model (GMM) fitting has been proved a robust method to generate high quality, independent conditional samples of the posterior probability density function (PDF) by conditioning reservoir models to production data. However, the number of degrees-of-freedom (DOF) for all unknown GMM parameters may become huge for large-scale history-matching problems. A new formulation of GMM fitting with reduced number of DOF is proposed in this paper, to save memory-usage and reduce computational cost. Its performance is compared with other methods of GMM. The GMM fitting method can significantly improve the accuracy of the GMM approximation by adding more Gaussian components. In the full-rank GMM fitting formulation, both memory-usage and computational cost are proportional to the number of Gaussian components. In the reduced DOF GMM fitting formulation, the covariance matrix of the newly added Gaussian component is efficiently parameterized, using products of a low number of vectors and their transposes, whereas the other Gaussian components are simply modified by multipliers. Thus, memory usage and computational cost increase only modestly as the number of Gaussian components increases. Unknown GMM parameters, including the parameterized covariance matrix and mixture weighting factor for each Gaussian component, are first determined by minimizing the error that measures the distance between the GMM approximation and the actual posterior PDF. Then, performance of the new method is benchmarked against other methods using test problems with different numbers of uncertain parameters. The new method is found to perform more efficiently than the full-rank GMM fitting formulation, e.g., it further reduces the memory usage and computational cost by a factor of 5 to 10, while it achieves comparable accuracy. Although it is less efficient than the L-GMM approximation based on local linearization, it achieves much higher accuracy, e.g., it manages to further reduce the error by a factor of 20 to 600. Finally, the new method together with the parallelized acceptance-rejection (AR) algorithm is applied to a history matching problem. It is found to reduce the computational cost (i.e., the number of simulations required to generate an accepted conditional realization on average) by a factor of 200 when compared with the Markov chain Monte Carlo (MCMC) method, while the quality of accepted GMM samples is comparable to the MCMC samples. Uncertainty of reservoir model parameters and production forecasts can be properly quantified with accepted GMM samples by conditioning to production data.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信