{"title":"Pretest and shrinkage estimators in generalized partially linear models with application to real data","authors":"Shakhawat Hossain, Saumen Mandal, Le An Lac","doi":"10.1002/cjs.11732","DOIUrl":null,"url":null,"abstract":"<p>Semiparametric models hold promise to address many challenges to statistical inference that arise from real-world applications, but their novelty and theoretical complexity create challenges for estimation. Taking advantage of the broad applicability of semiparametric models, we propose some novel and improved methods to estimate the regression coefficients of generalized partially linear models (GPLM). This model extends the generalized linear model by adding a nonparametric component. Like in parametric models, variable selection is important in the GPLM to single out the inactive covariates for the response. Instead of deleting inactive covariates, our approach uses them as auxiliary information in the estimation procedure. We then define two models, one that includes all the covariates and another that includes the active covariates only. We then combine these two model estimators optimally to form the pretest and shrinkage estimators. Asymptotic properties are studied to derive the asymptotic biases and risks of the proposed estimators. We show that if the shrinkage dimension exceeds two, the asymptotic risks of the shrinkage estimators are strictly less than those of the full model estimators. Extensive Monte Carlo simulation studies are conducted to examine the finite-sample performance of the proposed estimation methods. We then apply our proposed methods to two real data sets. Our simulation and real data results show that the proposed estimators perform with higher accuracy and lower variability in the estimation of regression parameters for GPLM compared with competing estimation methods.</p>","PeriodicalId":0,"journal":{"name":"","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"","FirstCategoryId":"100","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/cjs.11732","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Semiparametric models hold promise to address many challenges to statistical inference that arise from real-world applications, but their novelty and theoretical complexity create challenges for estimation. Taking advantage of the broad applicability of semiparametric models, we propose some novel and improved methods to estimate the regression coefficients of generalized partially linear models (GPLM). This model extends the generalized linear model by adding a nonparametric component. Like in parametric models, variable selection is important in the GPLM to single out the inactive covariates for the response. Instead of deleting inactive covariates, our approach uses them as auxiliary information in the estimation procedure. We then define two models, one that includes all the covariates and another that includes the active covariates only. We then combine these two model estimators optimally to form the pretest and shrinkage estimators. Asymptotic properties are studied to derive the asymptotic biases and risks of the proposed estimators. We show that if the shrinkage dimension exceeds two, the asymptotic risks of the shrinkage estimators are strictly less than those of the full model estimators. Extensive Monte Carlo simulation studies are conducted to examine the finite-sample performance of the proposed estimation methods. We then apply our proposed methods to two real data sets. Our simulation and real data results show that the proposed estimators perform with higher accuracy and lower variability in the estimation of regression parameters for GPLM compared with competing estimation methods.