{"title":"采用广义交叉验证方法选择良好的脊参数","authors":"G. Golub, M. Heath, G. Wahba","doi":"10.1080/00401706.1979.10489751","DOIUrl":null,"url":null,"abstract":"Consider the ridge estimate (λ) for β in the model unknown, (λ) = (X T X + nλI)−1 X T y. We study the method of generalized cross-validation (GCV) for choosing a good value for λ from the data. The estimate is the minimizer of V(λ) given by where A(λ) = X(X T X + nλI)−1 X T . This estimate is a rotation-invariant version of Allen's PRESS, or ordinary cross-validation. This estimate behaves like a risk improvement estimator, but does not require an estimate of σ2, so can be used when n − p is small, or even if p ≥ 2 n in certain cases. The GCV method can also be used in subset selection and singular value truncation methods for regression, and even to choose from among mixtures of these methods.","PeriodicalId":250823,"journal":{"name":"Milestones in Matrix Computation","volume":"163 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1979-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3015","resultStr":"{\"title\":\"Generalized cross-validation as a method for choosing a good ridge parameter\",\"authors\":\"G. Golub, M. Heath, G. Wahba\",\"doi\":\"10.1080/00401706.1979.10489751\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Consider the ridge estimate (λ) for β in the model unknown, (λ) = (X T X + nλI)−1 X T y. We study the method of generalized cross-validation (GCV) for choosing a good value for λ from the data. The estimate is the minimizer of V(λ) given by where A(λ) = X(X T X + nλI)−1 X T . This estimate is a rotation-invariant version of Allen's PRESS, or ordinary cross-validation. This estimate behaves like a risk improvement estimator, but does not require an estimate of σ2, so can be used when n − p is small, or even if p ≥ 2 n in certain cases. The GCV method can also be used in subset selection and singular value truncation methods for regression, and even to choose from among mixtures of these methods.\",\"PeriodicalId\":250823,\"journal\":{\"name\":\"Milestones in Matrix Computation\",\"volume\":\"163 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"1979-05-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3015\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Milestones in Matrix Computation\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1080/00401706.1979.10489751\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Milestones in Matrix Computation","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1080/00401706.1979.10489751","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3015
摘要
考虑未知模型中β的脊估计(λ), (λ) = (X T X + nλI) - 1 X T y。我们研究了从数据中选择λ的良好值的广义交叉验证(GCV)方法。估计是V(λ)的最小值,其中A(λ) = X(X T X + nλI)−1 X T。这个估计是Allen的PRESS的旋转不变版本,或者是普通的交叉验证。这种估计的行为类似于风险改进估计,但不需要σ2的估计,因此可以在n - p很小的情况下使用,甚至在某些情况下p≥2n。GCV方法还可以用于回归的子集选择和奇异值截断方法,甚至可以在这些方法的混合中进行选择。
Generalized cross-validation as a method for choosing a good ridge parameter
Consider the ridge estimate (λ) for β in the model unknown, (λ) = (X T X + nλI)−1 X T y. We study the method of generalized cross-validation (GCV) for choosing a good value for λ from the data. The estimate is the minimizer of V(λ) given by where A(λ) = X(X T X + nλI)−1 X T . This estimate is a rotation-invariant version of Allen's PRESS, or ordinary cross-validation. This estimate behaves like a risk improvement estimator, but does not require an estimate of σ2, so can be used when n − p is small, or even if p ≥ 2 n in certain cases. The GCV method can also be used in subset selection and singular value truncation methods for regression, and even to choose from among mixtures of these methods.