{"title":"Normalizing Basis Functions: Approximate Stationary Models for Large Spatial Data","authors":"Antony Sikorski, Daniel McKenzie, Douglas Nychka","doi":"arxiv-2405.13821","DOIUrl":null,"url":null,"abstract":"In geostatistics, traditional spatial models often rely on the Gaussian\nProcess (GP) to fit stationary covariances to data. It is well known that this\napproach becomes computationally infeasible when dealing with large data\nvolumes, necessitating the use of approximate methods. A powerful class of\nmethods approximate the GP as a sum of basis functions with random\ncoefficients. Although this technique offers computational efficiency, it does\nnot inherently guarantee a stationary covariance. To mitigate this issue, the\nbasis functions can be \"normalized\" to maintain a constant marginal variance,\navoiding unwanted artifacts and edge effects. This allows for the fitting of\nnearly stationary models to large, potentially non-stationary datasets,\nproviding a rigorous base to extend to more complex problems. Unfortunately,\nthe process of normalizing these basis functions is computationally demanding.\nTo address this, we introduce two fast and accurate algorithms to the\nnormalization step, allowing for efficient prediction on fine grids. The\npractical value of these algorithms is showcased in the context of a spatial\nanalysis on a large dataset, where significant computational speedups are\nachieved. While implementation and testing are done specifically within the\nLatticeKrig framework, these algorithms can be adapted to other basis function\nmethods operating on regular grids.","PeriodicalId":501215,"journal":{"name":"arXiv - STAT - Computation","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-05-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - STAT - Computation","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2405.13821","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In geostatistics, traditional spatial models often rely on the Gaussian
Process (GP) to fit stationary covariances to data. It is well known that this
approach becomes computationally infeasible when dealing with large data
volumes, necessitating the use of approximate methods. A powerful class of
methods approximate the GP as a sum of basis functions with random
coefficients. Although this technique offers computational efficiency, it does
not inherently guarantee a stationary covariance. To mitigate this issue, the
basis functions can be "normalized" to maintain a constant marginal variance,
avoiding unwanted artifacts and edge effects. This allows for the fitting of
nearly stationary models to large, potentially non-stationary datasets,
providing a rigorous base to extend to more complex problems. Unfortunately,
the process of normalizing these basis functions is computationally demanding.
To address this, we introduce two fast and accurate algorithms to the
normalization step, allowing for efficient prediction on fine grids. The
practical value of these algorithms is showcased in the context of a spatial
analysis on a large dataset, where significant computational speedups are
achieved. While implementation and testing are done specifically within the
LatticeKrig framework, these algorithms can be adapted to other basis function
methods operating on regular grids.