{"title":"域自适应中带噪声样本最小二乘逼近的误差保证","authors":"Felix Bartel","doi":"10.5802/smai-jcm.96","DOIUrl":null,"url":null,"abstract":"Given $n$ samples of a function $f\\colon D\\to\\mathbb C$ in random points drawn with respect to a measure $\\varrho_S$ we develop theoretical analysis of the $L_2(D, \\varrho_T)$-approximation error. For a parituclar choice of $\\varrho_S$ depending on $\\varrho_T$, it is known that the weighted least squares method from finite dimensional function spaces $V_m$, $\\dim(V_m) = m<\\infty$ has the same error as the best approximation in $V_m$ up to a multiplicative constant when given exact samples with logarithmic oversampling. If the source measure $\\varrho_S$ and the target measure $\\varrho_T$ differ we are in the domain adaptation setting, a subfield of transfer learning. We model the resulting deterioration of the error in our bounds. Further, for noisy samples, our bounds describe the bias-variance trade off depending on the dimension $m$ of the approximation space $V_m$. All results hold with high probability. For demonstration, we consider functions defined on the $d$-dimensional cube given in unifom random samples. We analyze polynomials, the half-period cosine, and a bounded orthonormal basis of the non-periodic Sobolev space $H_{\\mathrm{mix}}^2$. Overcoming numerical issues of this $H_{\\text{mix}}^2$ basis, this gives a novel stable approximation method with quadratic error decay. Numerical experiments indicate the applicability of our results.","PeriodicalId":376888,"journal":{"name":"The SMAI journal of computational mathematics","volume":"46 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-04-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Error Guarantees for Least Squares Approximation with Noisy Samples in Domain Adaptation\",\"authors\":\"Felix Bartel\",\"doi\":\"10.5802/smai-jcm.96\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Given $n$ samples of a function $f\\\\colon D\\\\to\\\\mathbb C$ in random points drawn with respect to a measure $\\\\varrho_S$ we develop theoretical analysis of the $L_2(D, \\\\varrho_T)$-approximation error. For a parituclar choice of $\\\\varrho_S$ depending on $\\\\varrho_T$, it is known that the weighted least squares method from finite dimensional function spaces $V_m$, $\\\\dim(V_m) = m<\\\\infty$ has the same error as the best approximation in $V_m$ up to a multiplicative constant when given exact samples with logarithmic oversampling. If the source measure $\\\\varrho_S$ and the target measure $\\\\varrho_T$ differ we are in the domain adaptation setting, a subfield of transfer learning. We model the resulting deterioration of the error in our bounds. Further, for noisy samples, our bounds describe the bias-variance trade off depending on the dimension $m$ of the approximation space $V_m$. All results hold with high probability. For demonstration, we consider functions defined on the $d$-dimensional cube given in unifom random samples. We analyze polynomials, the half-period cosine, and a bounded orthonormal basis of the non-periodic Sobolev space $H_{\\\\mathrm{mix}}^2$. Overcoming numerical issues of this $H_{\\\\text{mix}}^2$ basis, this gives a novel stable approximation method with quadratic error decay. Numerical experiments indicate the applicability of our results.\",\"PeriodicalId\":376888,\"journal\":{\"name\":\"The SMAI journal of computational mathematics\",\"volume\":\"46 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-04-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"The SMAI journal of computational mathematics\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.5802/smai-jcm.96\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"The SMAI journal of computational mathematics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.5802/smai-jcm.96","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Error Guarantees for Least Squares Approximation with Noisy Samples in Domain Adaptation
Given $n$ samples of a function $f\colon D\to\mathbb C$ in random points drawn with respect to a measure $\varrho_S$ we develop theoretical analysis of the $L_2(D, \varrho_T)$-approximation error. For a parituclar choice of $\varrho_S$ depending on $\varrho_T$, it is known that the weighted least squares method from finite dimensional function spaces $V_m$, $\dim(V_m) = m<\infty$ has the same error as the best approximation in $V_m$ up to a multiplicative constant when given exact samples with logarithmic oversampling. If the source measure $\varrho_S$ and the target measure $\varrho_T$ differ we are in the domain adaptation setting, a subfield of transfer learning. We model the resulting deterioration of the error in our bounds. Further, for noisy samples, our bounds describe the bias-variance trade off depending on the dimension $m$ of the approximation space $V_m$. All results hold with high probability. For demonstration, we consider functions defined on the $d$-dimensional cube given in unifom random samples. We analyze polynomials, the half-period cosine, and a bounded orthonormal basis of the non-periodic Sobolev space $H_{\mathrm{mix}}^2$. Overcoming numerical issues of this $H_{\text{mix}}^2$ basis, this gives a novel stable approximation method with quadratic error decay. Numerical experiments indicate the applicability of our results.