{"title":"Generalized cross-validation as a method for choosing a good ridge parameter","authors":"G. Golub, M. Heath, G. Wahba","doi":"10.1080/00401706.1979.10489751","DOIUrl":null,"url":null,"abstract":"Consider the ridge estimate (λ) for β in the model unknown, (λ) = (X T X + nλI)−1 X T y. We study the method of generalized cross-validation (GCV) for choosing a good value for λ from the data. The estimate is the minimizer of V(λ) given by where A(λ) = X(X T X + nλI)−1 X T . This estimate is a rotation-invariant version of Allen's PRESS, or ordinary cross-validation. This estimate behaves like a risk improvement estimator, but does not require an estimate of σ2, so can be used when n − p is small, or even if p ≥ 2 n in certain cases. The GCV method can also be used in subset selection and singular value truncation methods for regression, and even to choose from among mixtures of these methods.","PeriodicalId":250823,"journal":{"name":"Milestones in Matrix Computation","volume":"163 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1979-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3015","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Milestones in Matrix Computation","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1080/00401706.1979.10489751","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3015
Abstract
Consider the ridge estimate (λ) for β in the model unknown, (λ) = (X T X + nλI)−1 X T y. We study the method of generalized cross-validation (GCV) for choosing a good value for λ from the data. The estimate is the minimizer of V(λ) given by where A(λ) = X(X T X + nλI)−1 X T . This estimate is a rotation-invariant version of Allen's PRESS, or ordinary cross-validation. This estimate behaves like a risk improvement estimator, but does not require an estimate of σ2, so can be used when n − p is small, or even if p ≥ 2 n in certain cases. The GCV method can also be used in subset selection and singular value truncation methods for regression, and even to choose from among mixtures of these methods.
考虑未知模型中β的脊估计(λ), (λ) = (X T X + nλI) - 1 X T y。我们研究了从数据中选择λ的良好值的广义交叉验证(GCV)方法。估计是V(λ)的最小值,其中A(λ) = X(X T X + nλI)−1 X T。这个估计是Allen的PRESS的旋转不变版本,或者是普通的交叉验证。这种估计的行为类似于风险改进估计,但不需要σ2的估计,因此可以在n - p很小的情况下使用,甚至在某些情况下p≥2n。GCV方法还可以用于回归的子集选择和奇异值截断方法,甚至可以在这些方法的混合中进行选择。